var/home/core/zuul-output/0000755000175000017500000000000015156602243014531 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015156615402015476 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000310402715156615261020265 0ustar corecoreikubelet.log_o[;r)Br'o -n(!9t%Cs7}g/غIs,r.k9Gfͅ ?K"mv?_eGbuu񯷑7+%f?7ݭ7֫k\% oo/q3m^]/o?8.7oW}ʋghewx/mX,ojŻ ^Tb3b#׳:}=p7뼝ca㑔`e0I1Q!&ѱ[/o^{W-{t3_U|6 x)K#/5ΌR"ggóisR)N %emOQ/Ϋ_oa0vs68/Jʢ ܚʂ9ss3+aô٥J}{37FEbп3 FKX1QRQlrTvb)E,s)Wɀ;$#LcdHM%vz_. o~I|3j dF{ "IΩ?PF~J~ ` 17ׅwڋًM)$Fiqw7Gt7L"u 0V9c  ˹dvYļU[ Z.׿/h QZ*U1|t5wKOؾ{mk b2 ܨ;RJK!b>JR*kl|+"N'C_#a7]d]sJg;;>Yp׫,w`ɚ'd$ecwŻ^~7EpQС3DCS[Yʧ?DDS aw߿)VxX帟AB}nyи0stĈCo.:wAZ{sy:7qsWctx{}n-+ZYsI{/.Ra9XcђQ0FK@aEDO2es ׇN# ZF͹b,*YVi+$<QMGhC}^}?BqG!(8l K3T[<~6]90}(*T7siv'=k 9Q2@vN ( R['>v*;o57sp$3ncx!>t®W>]tF-iܪ%GYbaRvHa}dkD̶*';ک|s_}8yj,('GrgTZ'U鋊TqOſ * /Ijo!՟8`"j}zӲ$k3jS|C7;A)͎V.r?t\WU1ojjr<~Tq> `=tJ!aݡ=h6Yݭw}?lѹ`f_" J9w4ts7NG GGG]ҡgc⌝M b/ΖlpaX5)Q VE JN`:a!KM/+9 bG+މG uIo1]ߔr TGGJ\B BR 4X\r RYGVق?<6jHSJ Jno#ˏl_}z?1:N3cl.:f 3 JJ5Z|&הԟ,Tصp&NI%`t3Vi=Ob㸵2*3d*mQ%"h+ "f "D(~~moH|E3*46$Ag4aX)Ǜƾ9U Ӆ^}ڲ7J9@ kV%g>a~W;D=;y|AAY'"葋_d$Ə{(he NSfX1982TH#D֪v3l"<, { Tms'oI&'Adp]{1DL^5"Ϧޙ`F}W5XDV7V5EE9esYYfiMOV i/ f>3VQ 7,oTW⇊AqO:rƭĘ DuZ^ To3dEN/} fI+?|Uz5SUZa{P,97óI,Q{eNFV+(hʺb ״ʻʞX6ýcsT z`q 0C?41- _n^ylSO2|'P'BOTLl-9Ja [$3BV2DC4l!TO C*Mrii1f5 JA *#jv߿Imy%u LOL8c3ilLJ!Ip,2(( *%KGj   %*e5-wFp"a~fzqu6tY,d,`!qIv꜒"T[1!I!NwL}\|}.b3oXR\(L _nJBR_v'5n]FhNU˿oۂ6C9C7sn,kje*;iΓA7,Q)-,=1A sK|ۜLɽy]ʸEO<-YEqKzϢ \{>dDLF amKGm+`VLJsC>?5rk{-3Ss`y_C}Q v,{*)ߎ% qƦat:D=uNvdߋ{Ny[$ {ɴ6hOI']dC5`t9:GO: FmlN*:g^;T^B0$B%C6Θ%|5u=kkN2{'FEc* A>{avdt)8|mg定TN7,TEXt+`F P |ɧ<Ғ8_iqE b}$B#fethBE;1"l r  B+R6Qp%;R8P󦟶Ub-L::;Ⱦ7,VW.JE:PgXoΰUv:ΰdɆΰ (ΰ0eTUgXun[g, ׽-t!X򴱞_aM:E.Qg1DllЊE҉L ehJx{̗Uɾ?si&2"C]u$.`mjmƒVe9f6NŐsLu6fe wkىKR%f"6=rw^)'Hz }x>1yFX09'A%bDb0!i(`Z;TyֻΗ|ִ0-6dAC5t[OM91c6@i]<ք6ym®Yi&s`dyMX](^!#h k:U7Uv7чd)wB5v-)s蓍\>S[l52, 5 CۈP$0Zg=+DJ%D  *NpJ֊iTv)vtT̅Rhɇ ќuގ¢6}#LpFD58LQ LvqZDOF_[2aޙ-did˥]5]5᪩QJlyIPEQZȰ<'Xa>EE衢^}p/:F?}bi0>Oh%\x(bdF"F 'u Qx`j#(g6zƯRo(lџŤnE7^k(|(bƥYr猸p$nu?ݣ RF]NHw2k혿q}lrCy u)xF$Z83Ec罋}[εUX%}< ݻln"sv&{b%^AAoۺ(I#hKD:Bߩ#蘈f=9oN*.Ѓ M#JC1?tean`3-SHq$2[ĜSjXRx?}-m6Mw'yR3q㕐)HW'X1BEb $xd(21i)//_і/Cޮm0VKz>I; >d[5Z=4>5!!T@[4 1.x XF`,?Hh]b-#3J(ͥM:<`pz?]6laz5`ZQs{>3ư_o%oU׆]YLz_s߭AF'is^_&uUm$[[5HI4QCZ5!N&D[uiXk&2Bg&Ս7_/6v_cd쿽d@eU XyX2z>g8:.⺻h()&nO5YE\1t7aSyFxPV19 ĕi%K"IcB j>Pm[E[^u I66|*f\#ߍp{8sx[o%}wS`ýͽ>^U_S1VF20:d T2$47mSl*#lzFP_3yb.63>NKnJۦ^4*rB쑓:5Ǧ٨C.1`mU]+y_:,eXX맻c5ޖSwe݊O4L)69 War)|VϟT;Cq%KK-*i ѩQٰ`DݎGu( 꿢\cXn }7Ҫa nG{Y bcWa?\34 P U!7 _* kTuwmUr%ԀjƮĀdU#^ۈӕ3ΊeBO`^}ܖj49lnAvoI "%\;OF& wctغBܮl##mϸ.6p5k0C5PdKB g:=G<$w 24 6e/!~߽f)Q UbshY5mseڠ5_m4(sgz1v&YN2姟d4"?oWNW݃yh~%DTt^W7q.@ L⃳662G,:* $: e~7[/P%F on~$dƹɥO"dޢt|BpYqc@P`ڄj҆anCѢMU sf`Yɇك]@Rɯ?ٽf? ntպ$ˣ>TDNIGW .Z#YmDvS|]F)5vSsiExţ=8#r&ᘡĩDȈ\d cRKw*#zJ9tT :<XK*ɤwoJarExfKB4t@y[6OO6qDfEz]1,ʹB֒H ֱw;SpM8hGG&ƫEJި_1N`Ac2 GP)"nD&D #-aGoz%<ѡh (jF9L`fMN]eʮ"3_q7:.rRGT;}:֪a$)gPSj0j3hLư/7:D-F۶c}87uixoxG+5EekV{:_d* |a%ĉUHSR0=>u)oQCC;^u'}8H0]+ES,n?UU{ x~ʓOy_>?/>l8MrHID2VSsMX^"NۯDc558c&'K0L /C5YDqNe~ض˸nErc֋@aw*r܀0 a {RQXV-/p:MP\<=<^越a/bz?ܓvjIg3MN4:]U]STa,@OKdĹgJ8@o2k'Hr~4Z(I8!H G8HNW%1Tќ^?G(" 뭗R==9!nKErHc1FYbQ F;v?ob-ڈFalG*rEX}HAP'Hҷ$qM9(AHx!AF 26qxCdP!NZgҽ9l*(H Žڒ;̼|%D Ɖ`Pj . ֈ,ixp`ttOKBDޙ''aLA2s0(G2E<I:xsB.ȼ*d42I:<ŋu#~us{dW<2~sQ37.&lOľu74c?MՏړ@ -N*CB=i3,qjGkUտu6k Cb8hs&sM@-=X(i7=@He%ISd$&iA|i MiʏݸT{r[j顒x.Ƞ"m@Hy_I )j|s#RGI!dTKL&4K>#stV \'xMgaSZNg8>e!^f%cYr]qs:"̊;isXa]d+"v=x7p.fZCg_Ys;pE&\U}ܫSh])qKYAـhhdEnU14&G * QIQs;rԩ.k83֖8Muqu_48dHܥlWW q>fu6+'}xu\Veelz`Zbym gp8펠ˋֆ:1IC8qٞ\vXçL ]X/r}7O}Wh,h ;RQ=]u00yiC۔I^3!?H6iUH:ô 4P$rT`%2Aq-֢׍qt=@x#~0)p# ы9'iri]ͪ/@繁qVGCڤr,DihB ,m 9 _$q3= A$IC"6g^4e`Xo(D*6"^eTh'4xpFڜe'fVQ7~'c L^ԯwIڣA.}H;Ë*׬=`^ 9]r鐃 -Dfi2|QwZk‹u^6DQ1&H凎c!n[mi3)WfsF:M"uҷs.1!뾧1%s,hQs|hx̗3%*v9(I;:'>uQ+v)vR/egBhAAdh]4H:nV$tHI98/)=mͭ ڐn}}~ק?g_6WĩDRc0]rY9'z .(jHI :{HG}HDN`h7@{jnE#[dz;n#y 9D*A$$"^)dVQ.(rO6ӟZw_Ȣaޒu'- ^_,G;U\cAAz7EtlLuoXuA}bT2H_*kIG?S(קjhg 5EF5uKkBYx-qCfqsn[?_r=V:х@mfVg,w}QJUtesYyt7Yr+"*DtO/o۷~|hw^5wE of7cꃱ.)7.u/}tPTGc 5tW> l/`I~>|灹mQ$>N |gZ ͜IH[RNOMTq~g d0/0Љ!yB.hH׽;}VLGp3I#8'xal&Ȑc$ d7?K6xAH1H#:f _tŒ^ hgiNas*@K{7tH*t쬆Ny497ͩ KVsVokwW&4*H'\ d$]Vmr달v9dB.bq:__xW|1=6 R3y^ E#LB ZaZd1,]ןkznxtK|v+`VZ3JϧC^|/{ś}r3 >6׳oƄ%VDSWn 0,qh! E-Z%ܹpU:&&fX+EǬ.ťqpNZܗÅxjsD|[,_4EqgMƒK6f/FXJRF>i XʽAQGwG%mgo 恤hˍJ_SgskwI\t`ﶘ080ƱQŀllKX@116fqo>NrU Ѣ9*|ãeeH7.z!<7zG4p9tV|̢T`˖E ;;,tTaIUle*$!>*mBA2,gJIn_kSz)JC]?X(OPJS3.}clݨ{e!MB,cB߮4af祋,1/_xq=fBRO0P'֫-kbM6Apw,GO2}MGK'#+սE^dˋf6Y bQEz}eҏnr_ ^O^W zw~Ȳ=sXअy{E|!4ӥ2 ]8â6 U`V%`!c%؎ʨTzrKh! c.}.D>)d_ 8rcu,wf2?Ǡ*_lDn}rauyFp*ɨ:UiM2r:9ct X1lmĪ o玓,R%!`hGT LYF#g<cm${|Xdu4tmtїUJ\~dc0KcMlf2?mμQ ߉J4WrSHTdp"ӹ'cJq2zPlX̯.0H!ND@UapVoGڧD5>H]f@!=߸2V%Z 0"G4ȇʩ@]>Y$ًF_Mm_Tt)ib+q&EXFu򾬳ǝ/RS>r,C2NfOjpcm{Ll9vQOT>9U;])>6JdbXԠ `Z#_+D[7IIjJɟUh ҙ"`"a ߒ"G̾H`6yiCk(OA/$ ^%K^+(Vr[RR1"u4A.1X0=7f/"(o9/L1X{]q`Ȝ/; 9a>E)XOS K9mUxBa"'4T[Jl /K/9,rlCAj_TiǘP,:4F%_0E5IE'rX-|_W8ʐ/=ӹjhO%>| :S Px„*3_y.g9| ;b`w NtZtc> ײ1KĴ{3Gl& KT1ZWX8?C]~We$9; -.D087?1a@P5B,c}jcGȱ WW/ @a#LA4.ٹ^XڋXٝ:^Izq. ٽƎDn6ٹBc5Lt;3#i3RAٽ9| cbpcTfp> 6L/_x 'ۙz7~w~);qU9GDT! 6]c_:VlnEUdn6UˇKU;V`JUݵޙEO[)ܶCy*8¢/[cչjx&? ՃJȚ9!j[~[' "ssTV2i sLq>z@JM->=@NỲ\쀜*/) ̞r21.y? bO]3?C!yw3ޯL_Su>o>&lrw&i"< :]_<<7U_~z5є/rfn͝MLmc 6&)e+n7cyy{_~궼07R7wPuqpqo{ߟ+[w_uOq?u-|?WS_tOq?Eu-L_p?Cz .e ϿO*3 `Ђ6a-`kIf-s,RL-R`1eL~dپ&+IhYRczr?㐟,v~,b6)up)3K,RLW"Qd9JgT\1f3@Kh% a4x,kA k ^d kYj5Ah𚄓vXZhX1xҖ51Y +Id ZZ\C| fD>hB֡#-$+Jpሟ,Cg:6 3 xH "}C[`ӨOAFn5ʬLHϰ:N@VcyBI#Dr. "h hg ۃm-qu>V&൘ G7qi#^tҒ[JI!{q*lrD܇Gk@;oI<5xZ4xM"؇'k!>V|lk'{d+ :sXӄc)?W`*|\v aVT0"tMًcΒVz]T.C$cEp._0M`AlF̤@U' u,—rw=3}resLV&ԙy=Ejl1#XX۾;R;+[$4pjfљ lݍ3)`xvcZRT\%fNV Q)nsX }plMa~;Wi+f{v%Ζ/K 8WPll{f_WJ|8(A ä>nl"jF;/-R9~ {^'##AA:s`uih F% [U۴"qkjXS~+(f?TT)*qy+QR"tJ8۷)'3J1>pnVGITq3J&J0CQ v&P_񾅶X/)T/ϧ+GJzApU]<:Yn\~%&58IS)`0効<9ViCbw!bX%E+o*ƾtNU*v-zߞϢ +4 {e6J697@28MZXc Ub+A_Aܲ'SoO1ۀS`*f'r[8ݝYvjҩJ;}]|Bޙǖߔ 3\ a-`slԵ怕e7ːزoW|A\Qu&'9~ l|`pΕ [Q =r#vQu0 M.1%]vRat'IIc(Irw~Z"+A<sX4*X FVGA<^^7 vq&EwQű:؁6y\QbR9GuB/S5^fa;N(hz)}_vq@nu@$_DVH|08W12e_ʿd{xlzUܝlNDU j>zƖݗ&!jC`@ qэ-V Rt2m%K6dX)"]lj齔{oY:8VmS!:Wh#O0} :OVGL.xllT_oqqqLec2p;Ndck[ Rh6T#0H Q}ppS@ώ@#gƖ8sѹ e^ CZLu+."T#yrHhlكʼE-X'I^=bKߙԘ1"+< gb`[c1髰?(o$[eR6uOœ-m~)-&>883\6y 8V -qrG]~.3jsqY~ sjZ+9[rAJsT=~#02ݬf¸9Xe>sY~ ae9} x* zjC.5Wg󵸊y!1U:pU!ƔCm-7^w]斻~[hW$k sE0ڊSq:+EKٕ|dvvjjy6 æ/ML-yz,ZlQ^oAn-})xǺǍ--qcl:WLg ӁvJ[ǧc~Of+8qpçco#rCtKӫce0!Y-+cxMK-H_2:Uu*corD~@N`#m~R:ߙ歼!IZ5>H;0ޤ:\Tq]_\_>e˲\oUQ\Wߋ47WwߋKpwSSۘF,nC.\UߋoVEuY]^VW0R=<ު˜˻ x}[ێ'|;c^ M7 >5\-> m-8NJ\ALd!>_:h/NAC;?_ξqĎ6xMY(=ͯl~l8V0٨T zL{Ac:&$ ^CpH*DW\r2aR|=(L X1|wrO_g ux1^^V2޲jMi^b``Q#dBxV#NBk1;DAV$"*1]Y~ d->'I`cZ{F_~ %6 ``g&clw`$ =V,9V1b&D*vWף+QweTFY:F:`RQ/F.:He2ޕ[: r/n :XD!TQET]ic=1ցR8+y݌z<^]:Ίۓk<_>uT?DȷvDHڣ`b֩@F;2WXǵ_H~c 7{7PWa 4GB:\]:*Qc~t',F)L? | I5N%xe@\WOzg;qv5V Ds.)ʶ]Rz ]RzqMTy/M.[|eȵ2W}˒dv~YRY3Ba;?Tw;ox9x/K*duPcngm4偡恥[v!iM3iߗøQ{@Q<+Znh9v^?\9JL_e& !RB1f8A8L?| ޑIB+=wdR5E&4Mc_2uW3^8ܒPhxAcrK>Koj/EI>9PýP b)U~@~wϖ*K%T&ݦ!%!l:TͼP ͸P սͩ2{mTjfN*+`A1`כU濡Yn[nR('HrmQ_Ǵ z$~`kji/|_C设F kXb||"?d+Kr-Rt2B-dnHw8yUӋ%-QXͬP,L ;γ xy8Df~5:b?YiCo )-TSkYFEGE^*m['qvWiWQ/N_t7OF̧ٙ|! ptcJC3+q,R$@F؝ m ylY;3H9>f!Qu7}. L صS=7{Qb>iݞ^U`{Vۙ! h˰tm}[D`Z>]Gb+6f:>?Ԗh4vꊧu4*=GI!i#Q1:|'GǏ*j ^?#{|eI4kxum0G~ yymLk1ɯӝc(.,H> ϡ"ΫB$c0)d uws](4.uA0Gi+:?Oy!6y4oͣ$(Ѝ.2]F#Bюl`cS99H9F>\4M 'OQfI:/DY?9YDݘC>?%3Lߊ[^qTć>U`&ʇѿsg>(ШhNLSEzdJմ%-Ȱ-FeQ:'q!T'>~úVE"g8qդ>9kϲ\uYjH˻6gt[~|~\ȲvL?|30R4rޤU4Q 9muѰdsw+A/CHp5hOk}J~5+X\Cq lG: mbs?Wj>XyQ,[xgJҽW歬2G͕ G,L)~N``> ZCDrHٗyʬ9J#Ĝ\¯y L>ӈnɁg~G?= 4BݓPi%<e}x"0![=fslkK+fm!:Oą8ʂ&&JJˁf^sysx!M\ƞdM-$ZZ#>ص dAH1qռ« %C(«sPʂ=E99CP?o<$|ִ~!䔹BI!n<ؐCOo ]iKt(ɉ@3>h_sP V](!ꐈo෯e2<GÓ$}ѻ0^V4x`\%FbЏ:2Msч_k~ӌ@Ssr`?ϕ1Uۊ_kZMu KhwÐۺ10CWu5Nj!жtݷm4ɕUPܵ71@iȕZ@XZr.jVCWVo[g뉓ߕ&[~Ju! ֤m:7ffP&ו`!%K㛍HJd])pQCAۇ踚w]k4t;p@׹d,_kS40S:WӼ&fj`GOe)+u>~:2IPսghP >zzxt-!>ຽpݎ3ҢN`19VW}X%۪hDD3|=0dL椕$_9]檁N#Etp^;LpWhrƷq8#2 V5 ,tf4 k-w@*]h`iв`wKa OCvqc,OgS\#kZoK!39m/߃!&XYՂ^7#!# 5Hi 6`PS"mP7U5/hdnTa*.S29="T)Hgq0"UBy{{r)nЪ.,l~32˪@sz7ᖣuZ/Ͷ? (G-}e>20@}fӥ4+}𬹲m Ҹv7C ŧvYNޖw7QnۻX/s@ۭU}?G.B>nfS@Pp0gm W xu37@ξ42~uqpS~tJs!}.@Є<= \\~*Gy!.]Np ] eM"jRa.i=90k+%JB:pk( C芺}U:1r-mK+_ C0X:>sC@h4|I TkFvچ[-H S ky8?g=,} <;nyD@±$pqhyhL~f@I9{Ҵh/M&3;xI`lOUL" 4f}go3,,[DQu95?,՘9)rt9)p!N_gxr0@H\,="ƁuN*L$tBkL蟴#x$*w[j4Z\?{6B/I=)rl`6_E Krm$;i.~3$ӎ:Rk-ǫ[iaNGI1}6u>,oː܏k]BǶ4V-VMp%1y{"d$2N:`qETgѴ MǘnVׄ%F8\Ud\A,#xJ& C瘺{.5%`zNJ OH?x6Y!" PWpC gbTM*'^|YU7gNtkx{U퇷~3/=`n u`눛`j(p7` خNO7W:C5DV-TbvN2#$ûsp ;ݻ, W3h<F D‚+AG4{an|7F5P4\F[Y!/n%T`w'i`c3' `0K:5 lű@P\WqZ@> s(@Y@vB*Jf.-{ g#djHLB~*q/`pNJ1ND# J^h^W'ft&('d " XrjӍ4E#$/r rI!9 HR,.#ŦCDNq @\[Iw `s=п\rɌԲpjȉkD %}c'[r 9H|!tհ:1 S3@#]1F\RkNÁtM4[jMi!:uG rrN-\[Z?P~',1jZTԮS{(*عSzٝ)#^^\'U |jGL+:VoH\ VԶyƙw! _kG^|IGr8ӈop36B1kg0dW"0Y0PgCeny"7n9^G_v*<7M; - ͬ$ni(`?:>[ZܦtqJ| O綆^2iҮ{<׉1y)LwmRjnkaD`7Yp[݉/7m 5E#G$ 3hlh-FpG#Uxvp6:^*E^80bQ=E\j$+I>٫$PJڧ% `\ C_bq%*XXbr y]ya(iTH>$딤_Rl t!^ ܿ(|3>R~f0C@gy&L! ]+rq~hQ~uF1, !AB`3 K2A(c$y.JiS ֥&{J؞RY I4 2,11tKV7eo0 Mӱ{>\TuH[fSj\I̧j,l, +@I^ V4pkT d J1'&:x=ehT`BU#,۵܉.XbZ!,'4\bHM!UHzE@<t.hԋp؞6౥aژo~0{JfGFY.O4o@RK`%X]UsHNrXQ?׳ٴ }f ,{3t97ʾ:{_-]W՗hDfؼǨvy5Dа/8 4y=sUoW^.\%O #-PmqŪ;L(LI":}:0=N3TÛ1*.U!)(|Di糥?azZ\\w:D[^d:^b;Lߊx2r7$r*67-jRjʂ7E4@dЄ Ks&Pqgg=UHwHGO*%qx?F^8a}3/BJ>- *:l<d:8P7\\U5 Y]ד~9qA-N]X`t- 8T}r냪V*u6r-WV* j)6;ǫlتlX:(fy9QFC ҞE_yrxr1$yq*< G|p1-b$/1t&O8;tCSq:΋;>ǠwbAo{F/w8k>&<Z)p0e(^:oѢ\ }*=c{èQA=xpqWhSpd1NYǎ3d3;pq3 ^ M+a fkɼ$/HNX/,<6 ̀t<,,3Hͫ-rո~ӰK܃k70Z&tkġ~*fV[¥BmKiV`2փ@,q7q`70 0Խ=DًlƇ4񆖗)me(yL(6XmhM!Q 6ptF6 Bk{hrM`#GUG&Db_V4N VX{+8DuLl?Hb(4 K9 z)ոdpQER6A躡Kl( M&{,7C6\gQ4puUC m40t4[[zh+m-Йa4 :цҳkFW: q&̮G;9|rc(i̥]NU#נ& wDiYnY]zB }T= *ūg/&hH,v1~N.S`~/UB2l BʞH(ۂPkPk B GDB- ޞPyO$ނPgPg B uGDB-u u ݞPyO$݂PoPo B GDB- ߞPyO$_C.!wou}N 5"b~౪P{`5 )Fʮp9/Ac|& de^1Cm63_xч{)q Iz#x%I9>`64@V8BrZGoǜ5=K{9tCY? k4#i)yM_hጀ -ZY5AaP4σa$9)(!NjV6^gHX(*E?-:) .ͧ?qE&yRd9::h}%5>ghnr|uUr4vEŹ{YO:8+ ܄Cxif{ `ׇ3CEd?<%$R}"4;< ''tGUSXi*Oʂɇ+n\@דa^)C1+3y"ܬNjsX Ρ-1NqZD#}`T-ۥ Ů ``kOp,ǽ!RaY׬PQ[w߁y~Mh4kI hz.ƐCEp1ˁgg4EXmOA#wX>dR=L)Rm}*q* G^P&RI1Iz%(kzXegS!a1)hD֨:KX^BzjbV<ʓ,)e V6^\4qY3bWY zVIȳÔYո1lnp#挂"Ad*Iotmja%St]]'x2gidr.FNd)Q6~/=ê W*C=;"F~x| E QZ[hʙIb#1G ] 2:KeSnY"fsʰqf"#quxuPp_IX`(f hLHʻI,dH9 ASWfFx)-a' #V`NdCt0x_oZߣ^ J4)`J-&"]ϵ r1,A}&Br0H7 5Kƒ۵-%dH#1 WL*?Q*g?["Nz Ґ ,D?nXrtFrsCJ)}p>Xb_7}ҟ*4Eܳ}x{*n7 ňX.:bǾk=w_lأ1/$>aN@]Ix e7KqqQc]2x#?Hw:V"Ů)m< ` i+#Ѿ+%٬N1޻fIC}Cַ: !VmH'0Hi]-wiis󢔱[*wb.K9 1>Q6LN',#|,F_(Q~< Cw=`Y`l(aWekS;Q5=o'x'fSd tuĤ:֐}ƕOWu 1^*j%ou4y54 ߲dӣ jWD1P3nJ\ޗt_ğÀ).D>e8Gx*5R 6b Ԏhv8iߦJB^619*CDJ݀v\"'6mZCU[kK PNVOpkC"W$#%p/[P&9** (x$ no5? 2bPIEƵ#զh/m9D}sໞrM#-Wi?w\. cO/\}7{(螊gR>jQhYnM7Fݺ)1"p>,8Ms~9"Q`C.6H @Gi:>{5d!ceILga_4:D ܱxiƄTm#/(2S#YHda߷ d*͆4o$ QkRk0?/hv}Q kdLBc;)a>VOa]4XwБ)Wĺ{\-{]B%=h'Í=w9-C豽1N-NST;+)aFsYw|[18f U#C ˎ:l7skkC8yXÂGH+MnWqG/R9@f%! ӧ$/"8ؖJnO3ӯ?z_ڿZd|}5F,<>M,c[>]HwK$Jgf认pl#=_g8!eX_.3>@ e>64x|g%{0#_jOxqz/Wa7C8H-{WG~n/xHsלNʮ4Q1QB5hQ[$t%8hjA_h]嗑OE4Ѓyˍ3p4ڊhйtNDk֕=Fɪ۟BQm(Ƃ{dv)4 UX ~*,R,+Yp n*.]B CPjD %pX:7stmHIOtJg8P9,9ܥrYxI#&,WkzqG;K'Gp{w}Â"7G[/k9W&B6BB*  !),E?[MLqD )AI+4(`?1REPh bGfzfYMjDF(N+ܷ @Bg0 Ed+P̼aYsa*\ XjW \<R,oG 57QK@SIuŃUV I02 2vuAIw~X1\ a;,h4^1W0$kLӮg9tZBi)lLRd65l ?vNŮZ6R; L E?R,We#$|OW8.pM&1|㢃,| KeЭX[zpoߡ0>0%_6b &~\$Ī&v2ֲ@ qZn!I)UT`B\˂d:*A&ˆ8HW1g9S}*7#M6Vcv>)U^]cKdòAt9YIzbGS%qhݚs z1Z/*ٲAdOx9.5vJjJN=u"[ogB tӠtTb!_s\!57_WUB[ruA3^Û2Fdidcܦdi c8❻(r$ue_Xش-puqzzq)qw Eg\Kcmrar\x<$&,>_:qrL:6QU's$bJcm}z!E5R\.疯7>쌙|s}pv44q1['Fuщ ?u¾sR}BkAÍԓd-͈Jkwm= rocZ 8'^` 6Ow{vu*hCBь1b௓øAz):S,A[A^ݥh΂ ݹS'UF-DD׾,*2tlȕl* Nwd8.BL@%D>4GS=\!wQ-W7"ޥBжYP-)juƺ* 1H,{r䴾5?5ήT-j^R_YtRg ?~J6C |ZZśSC&*eKr$kBKT(fauu7i9 h8Pj%bLe'vm,DEz˂K༖F4F1Rk#vS,k(c}ߔh]C3F`XD97fv9nLlJu=)=ۮjً":{ xXp=qnf!F)|!kb.94AC%@9CGKn<5b7Y}\I>U`qr!Pɀx+3l7sdr={"n^XFOulbum1L8y: [3,ؖ;; Vȵ~g~dՙlߏ~+LO{|'H-'fĂc%hӡH"YYܲ^8{ҙJkoI6kt"B9B;Aӈ9p%Z}۟E𜐟:;vŞΕ]RgsKXo\Tb(-&4H}1O,Ŀ2d~ub;%t<%j'\QPj&Ϝ,:O74[vbvQbrV yQit LZ^o '뀞+s踣pAm/n}WxvoYpsK :)D p -ji1gh\ǧaUBdg(rF*34ǂ !Cvϛ]_1|hN28iɩ O|:8ۑM'=T/d.W(/Jx~[N8y$p#jQR,<$9N^HL 6R,AF5BNf33c`Lk)ΒviٓStZɽ{?Y3 Qb չ_ s׉ВG-&j(x#D E&R1pk3@*(m\Yɓ1,vP E5l[9'vK0v RHIDpV0aZ]i]oZcK6X'/kd ƒ|JI_-Za^s9;q/=:e]N!r3hs> YfG:l]i(LLttxgIai[CZ7Ͻz.^*8L۲8Vo)с Ql DfW>lHx&F$'s .\){O`8ɬn$u cdE gEIlyWd3M9j6g{6?8u\'gx^S4?,8N֥G05&,8aj)^(}3i1e'RZA^^J!kkR!KJMVgǴCq̂|0~[jF$:K^g㴭*L kJcsl@M#- Ecj6=yhi:ӎߺ)Y<<wO wQ>c1Mq|dPO&6}Tǫf gvVPDVX X:!-_ow,%y=yn4Qڸܔjph/KiQ"1~^|#p%0ߘ,؏I y-))fg;I@V,L;W" N>7k6F}F̏?_dV&J77Y]d1cZ$} xSs%-eNfMgJ&g#PC, x Jbb n׻2 *e':EƓҹi(q$ e*6  ?|SOZT&TㆥOΆZ'iqʊy9!8\%E yhډO- 9tjr*-_^ak΄ҡ~&X:F]F/+!qđ8U7/,3>n)Z;<*ت5H*:E`=蜹Fv.Nă"&:t |J^~>R{32g}J/ࢸr#6 ݄2兺(+Ei`pm7kP原$zi/J>iEOɵ>[/k9WV&B6BB*PE~m- };ʞ׹yQ 8ef?m *e%[neiPS٤xzf()$5ȑvל>UUMݫ(*ڬR{LkՔ^j¬d᧮Q @8(ZXUZ LfkaG'4, JUp3avHfO jN_">3j ,bGW7z1n8ic8@C}W]"n_G/RӓQ! Bq޿E'@O[TΪW @̌]k"&21%aVôynBafq0,D?O:.SryĘv+zF׷ܖ1f*& !g w=>P“Q|skV]MNo_ Bg60O,ѨYt}K~>M7py?|>?\ ?i% ҂+4H.g*t' n^pGK 漎y[]"a04~&Whe <@ '7l %-d(tv.+B 7{$,Ε4@C>$>ܻ |2qD3jf]"6#Xz!n>㫬:{k" zBJȴ c!Nw0nbCKu<uxlj8)T !y/LqW}Y?tenRhһ@糟R<\\;.Q%nI1}k< CYb,O(.N@Qqq9}p0E\Н@Y4M|. 7thtwQCPBTy,(ê|TR;aC77e &C~9|4AZ~*k;z?a20^e90xT'3ޞr$lM ELa5{́Uh2[l%qC_W\3%_3҈FLmI#/={dKa--Addo }p+WV<;X$^҄{e/5c`h[xy+ N58p?Km P?@lm"YMZf/;MZ4sP}z]nJ@u!qoX/|I"!؆!}ˮMkXMl_Ul?9EVe˘.L jͨun2G}`fh?f6/!H}VG(K~&YJƹ&b(ޅ$& y/ht-$mQ3ZpTSkd͕nkճǑcTc"=EY]1i`k=:pEQ}nv2g}?x!B˅%h9B:%<VV1ʐ* $"Jy ċB1r\hż;mq:4KcdڧQg\Y~i IwxˀXo-#jM,2τ#N(RXct1rq9aW5/XZ2!JeQ~2r3%s\9j}g+( , feL!A^hZX$+[FӎYFXwmaRՎ{̍ēhFzsGs{4NWAs!^B๗6/Ay:a`74}=@ՓI5`a˸8HZL)yNP:%!*?2G~5Zૡ;,qw bB#] ۆT`ݚB)fanQʼn:Cם!&o *Dؓ-ax +aM8Qb 1 ;R' ;R'dR'[R)EgX]1T"kb L,He93bkWJgYqK)?kqh3g*LJ.>}++tL6/S&@mNv5}pQ!UFj$5W|1tjMq~h<>Ӫ^[hnjUya(ޞ, ̾nDO$XER` 3EAB`AأN$B_͛ƈ}i!xC5E!,p{$.%m 9DpT<` [%$)'A+FGDऀ ȁиnLJF[gx~i"Wj#J5kL@\[|Lr[eg>.F"q9R1"0Ns`,6%9솘)D[U0wQlw`m;!l)-FhK~TPZiNq!7"Q\9i-QXXZA=̂ECLKy(X$0f2&KuC"zl!bo1.=Nbp <|#K5墈67އF{IӪ!wܶ  LsYdQ0Cԅ5!ElK)⊙1 PT*Ef= +ZK[e EY(ɹ,-2 m/ 0 i)l80$/`u&*n"q J+SFZ]=?Yo FP6{%UǨ-?F)QRu-bBd0|ݹr$V )eGm ShF.`cP)315.h=4kIڹ ZS奸uA뉫j !vQ *;K0Tf44cD&q&Dh ddȯD=>Qc[ޒƊmB)*nKq瘵۔ o\jt]( KCbA^kk#ql8FJ ZR$2S+ #thDQPļ#^ADpqԡTJ5RS.Ef-@sqx4\NrP-4R,@`@\Z㏨\h%-͆"l"BBFm %'6R$XXN jԁ $01 5QIj7|1jʼnHKC 2DX 3FX-Jj 9i3P[&@T)$VAyٻ޶$Wxv`m|ld0_"[>Ό7j)˖,JM:$DQ_WuWU72Z(Iq.!ĕ=%i>fDca$Urblh6V[X-dw6+[1nhVL2E:4i0'ݙ^-'N1Ũ}$=iX0R*{ݝsiv(xvϗs b]Pcm!1\R #sE"&1D+'J|B@* WKz%I w!4&$O) O\R6%I c<oHjdBFJ0%LPXBy2qM5d*sJS0Br3Gp|řDq#g)i`JQm$1 \E튱,flPlTO*ֶ5{pw,7L`LR!鎟ŠX6@wW*<p NZ aZ0(܄*$V+ޭ`9ll(H ͈(خ\wl6Xe;5,l8{Zl$4oDZ9ݑCd86EU4՞h y(H > 8dϊzF3z'ީs_ExvSyEm|E1 e?]_dO];̾t ;QגGoI}P{ڽT3N`|3bmc%ԍFv#[ڀhʨۏ~Z#kaZ nK}cFA_g)ü<Ve_4ܤʩD֛;d-Qj΁֎mZu +tz*Gwm1OWN'ǾbUnZ!4;xO.wB8X t&y2 nv'WGYզ'/Qv/nܕQhCPN-<~c[]+97LY]S*1YUjv~wϦũut9/@iq CUOUwEaV̬ryojpu3[*GIXab7`X\YTWTz(4\^[~āK«)rk/_ ὿V)"sO,Q34)w>/7k,U'pskVSɼ2XnZu|35.*"7T9 .9N] V'R)L 3:` =}H\=%-,_1LղLP9Yn T [K*M8U2{*Ica&+p3P0]t$` [.<r99sCZUCЊV{eo4TȂG˂_JxԡdC"jI+ iq 닃+ Pd<%z>YAmj`>>} Z<`3ƒh%pL {I8y58xLxQ2IeP}B뿢0LM`giOQ;XGΌCmoBC!K7x=r`vJ~ҚJ{'C%A-ygAØ5~BĘ$<10ӡQpeNX[N#iwfZֲ@w`sbvOo=0fZ{~qPT \S{c=t4^1h-Xzzq{go ?C3gʳQ Dq hz3ƴREfvR1)L LjCǙluq& . $BˉjIo`AX3gzvbLahGqQg5QtYm겶nhǚ뽕Z fYS9> g}ދɬ"RoGcE(c ~`ӑLj={ ZYw,T;&Pp ԏEsH Z3<.P3ݗou*’D~X 5e?\b?ᛅwWWe1m\a7r*zK%E5O'Ell,{UOAv~>$)QoGٖM!QVՎm,(OԊm[l6sh.޺4{dZ v<\v<};㑞ܝnAC& W!ZG7e`*l#?`w+Gk  B "Hw$ܕT2!e}ki/Z߯Z:YWV u)TFdܴ&q{ғ^TznXϵpI" }1>dc8Yևֺ +fmbnj2?ց[: ڰX8Va Ҳotc8Yևֺ1"Tt8ӻE85rL 瑺fDRj[ɘlBfppmX~}Z}QnJ?-lGq٪ҥ 0_nP^I]wFbqGc2'c%X%Y8&,s16Yl3ؤ k_1 eF;*h]vufܪn{"w"|o3>"9.ymG97؏9 :7i!SKBJQSKk)f)|ش< %V<ߪA2vo(oTXK܋y7bn#MXǥ y%6NeFIjcWtb, <1r.ƚ$"z;5 P&x7k` ג(dRMhu,!ǰ!lC@(W'i I[0Y]]qt׹*RGE9!\N=AdYaAq*}s 9`Y=AC Խmz?jnIaFo?Z7ZP*0T#a2Vadр y$9zGX]e0rs ge5:/2>sS S2db!va'<+XЮWZǤpI(VZBm#bʴxXE)g21JxHOr D Ap9a'&gY_`~(^ܜO[gWu|>=X9%t4YZ'KLA,kΎ֓P䠭4k[BpiɅDV+y [J:,38 uI2,Rd ±qNr3x.,ZF4o9Ws6en-ǯȮ~*4*CTa00$x J֕FYQoZj[['+{vSDzo:P8>BJjV_*ȉnqs\S~D=ԩS7;'KUk&m/1ʆC0lp\f+&U"u(ܖxB,˯vˮ/ WYeֳnU&Ҕ$.yGEW8 7ʸ{}=,I%Zդ!J%qg$b;Hn a?UdUF$@ |1`}$XQ$ S.3=H걇C/IS÷l;:Imߖ$,8WkIMJeq"H9r-(؛6R#?|^4Ylؤ#4iF+SAjc &+DryTw>MwOWr~'?x7˩9H6_L3Ɣ Q{)u̮&5|Vp<F:TBRÐjM6(L2e eúL>؜A7/ؖA퇓' ț>S(jv,r~T`-/JToÓY5Oy npJ(pOU[M^p.qu<4P/AU[M^ ~ ңbTKk7?ŀā6ɔ$`l/Gi "¦ Cul܆ 3T:|bQmc I,H֠j{:&cni` eWmoj{`eTFut/U>|:m/!reaI2]ڏj{=8K´Q$RBȡރlӶc H~͠Tzӡ6{E`ʑ) V[]^vqT?DHu7FT<[uzEGeؠR6AÔRLpۺM{^4|Я:4 L"S3g$Tħ*vg-n]{ˏ=tN:-|l>qzN:iK_Umv%hbD['*{g[:6z}D m/>ޥ+INs='?}J8C`p`YAJ y BC rz{~pxi4!lG*]|xv7UGiOB,&B!CAP¶ ۮ>'V#>햻'M:;*664t,A 7к.UsbC>ZC9.ZrkRn?UZAƄxF3*Aʡ~lTUV) .z߸nn96mӫ>|/kim#v<[gі{b6W}jkFG}#ރ p[3;;hpQL8|o3i+4!8" 8`rp5\j JnJƶjTM̎ҲesǥNn&wASug2nh޾Mi[SNU醭8bTO;f 9L,PM)TošW&n:V^`]~O@tsH65z.=rR!=\=FB6z(l|Pxt&uc1Y R (мfM+0kC_'FȶS#6]SkT! +I&|bquʌ߼ՏݗT߼<6ߥf|lޏ]N<~I=g~;t2&8'K59òo^=ino53~kM-twNX=sך !,6}3{xȡ3m] /!C4M% 1k@ R΀$@GPw9Zeײt|<}~;L ֪ B` YmÙmbRCb@om38BM5Txu`^֟_/__89\o_yJsIdޤW>P~lxj6=GYgZ[SЬtbjFΈWxyH9?ClL  tOK?#H9'5ֵ\|mn,OScb6B$>F[|db8BWLpDrjqy}~r Ɓhr VBt={C)ZvS{c=N5aɃdr &ia]dqMtCB7\d3myF5b4N SA[$5 RT?<"ܠ4drȳs#]mWT,XHܔ!vY R !jalzڙ s[C3b=,RKy,y%]hRʒyY2Aʒ4.`n g9dFbXʒ 4^*K|׷i˦^{Of` Ey$z~ɴ7 M3}iHL0Al@#@!&>uy[]5?J}QO6>| g5`Y/qe/v/ˋ^[trĕ/|im./ױ[=څrǘ%PY!;/E Ў%_ʞ.0MݾimC9xcGƈ'9*#H9 >t0Mϕ7˳}Il?.a4F Zހ&h?w [|ǧOMm2);u>F32 ] \BR0k>wyxڅ`[8c#I92YsfAŬ`F70ch p&#tCעpYko[΍cCp"(k:H1rƛmhUȡyi3B;3T&c[ VUCJ7C^?AyK2R ,Fd/ t5{'{4kfb1رc*H9 deUXD6e2l&!ؤ`,Db T<꼩x)2׭ @# V,̃d)b 1 Mm ֟ C"dΉGq zju:"ƶ&i ǸIuc8za&)EbNz~#: ,ak"D08W -phjypxmCiԬ9qkF?an0S}{;ri  )W⠇pM)VI:u3`IZ[n智ɴ8Aև 3X5c|ꎌ C ]a=FXWwVw``GBŸ˒dbdl5ْ 7 5y|f՞g͙ABNBL. R 4W'rVQ[By]x5 :/y R >7U6VWb١a{pEh.AL F̖<ŲEx]@&-+PSUT5j*&Ю ]w5Gnv` SAVח:5~H7/elBT*lXU4^ny/T[M+fˊ[΁D4 ~dZg4Q @,1Ӷkuجy9;0c7XɆBC&d YAG':TW2 yYe.DiA9aJd2q z!fn#B.>.`].^k.LƏ)5߱Y R 8fL k4Oʴ}L򌭊6xY~g:H9Q tx&w\] 5(xT[eٜ1/H1`+-+(^)fJvƃetr@AVP@1p}-Z=ݯCϯVP&)-1ewBP5T ݵMg7sm~<!ѼgA9& 0וqOִ{;;bPsBR{s:i<0g&AGuP]P ,o׫Z&M84sVe)6k }Z2FzGdﯷJLA|W Tg+H9&fcp |hM? Hj> i+y\ RWF+GgVlJyCAG*:k3AX<:s:m^?m n4'k(gR;{Ă#_sTm  1l20"oџ`0iYS&d{OsHY { CB3#e##:3TF)gy,5뽠8풱4v9V?#H1R=u?n[Ao'ceov>yZ3??{泽!Mg tj]=a$ljqymi__/WyUV1_[=yw~{{uL | ƒ?}UIĥI[cW^^$}$/>>~x\$mh^/n/{{Y 1A a(͋n?5as׃p$}rvk}>o(;џߔ3wqH_ ݃4!#{ a;d[K<8AȞӭ4H K=bꫯXlR"LjT=x# 56O n_7_О fR/qg {._+i>[\[\5l(2eP,Pڭۭkl`\Z' CFPyo[E/J¹lbae֨?Ɠxrb{#T=t!['W0Ʈ{'WH3|C51 tvtY;YS0@ {>Ѽӏ X!M*̗/l[,ش{iEYCk%mh]wBZ ffS8P~J5[y6 $ZI#L-[Ɩ0>s˜^A#bMV$[!R)6 $6g[^ʺ2|閊GPI+OWrmm9Wv[|gs&d>^ΗVS9:f' ݳ@S[°Q|B`@+ &VOeC}m%}^*s+kצynh+7:(@jrnj-my7c+$킷 ,$Rg4_}.9N&{Urv~{V-CI'[hjn&3?*h-s٠%h-.ӚyZt~ hYuA9*J1DR)R_`ч}/$j=gZ)wA4=4N|nggQiG=UD f}kQ~Qo;[1V|Xr;p& Gg6G*bT=Ig cXg >Ӿs6ab|eZAMukʐMj0(#Q*i*)omTIVI5O 63q"0,f6i76Aq|*i[%mmtio$ ,7pX'G7ƈ_:PAqFX/EAIqه]8O^Qo4 af:]6om9ug a_w9ݎg:Ugk5eŴ0p0)L*KzpapֿF`gKw~IREg%>wǸf Q3v5CuzC'-m: uH,r3jas\m@wsUn۴\> \Xt{}5YU3JYh*bF,q/MO9h_tem1AxG}PdGR$«| jɇf1y)Yd Yw=[ȿnCEwtMKԦՃ'aKCe7ɾ oLtM`e5^IM::kQ"NW2)^L,]hru_hjwdLF~^: 1I`rn]ŽAۼ{_5:w=Fa}(PG4BSkL75,4&jљzc8d_0A.q–If'o$&S|VS| ǭ56ɥyohMѴbyoVc4fmkX^cGZk1G 4/si&0SBHZ&BrIuj)ηV.{E7F(GkmZ~uEZ3I%eJ J!$jh_oHm2j<~LsoIv},睝{R09>onƼɻn^>)4Wbˡ]/GY~=%6`X=t_D#6OixX?;ir[r2olq}_7E+Zz*-:'ĵgV+\6EDj*ZWjXW&QZ+&Z^f-2h,Ac1r,Ųu1biWvkX!Giјל$/j8="Fb6X.Yu$Kc$ͤU\_5Q^ m6O ӴR2L*8@iXGiC,[&Sp)8iTr&uѶPY0dhH ˯ݕ98 a/"m!ƙF*8̬k}9U/0G 4;: n$V4&2t1k3=B W|b`bkPĘ7̈즛vx=gfFg﷔Ͼ^>c=3 ge0EM3!1s`3[`VIU7i(oͽW0 ѵAU7_+YV3po(?N6Z]ݒ:_$./ލ]tL_ІK=\oCwy:]Oq0j!ܳ-Hۅ!J|߯ˏr7_?x@&cL*Fk6,hb7uS\OFsTO08W<.75@6X(D^IuŬ[e>Nj[f;Y53N̙пBG12/ N?5q_b0ħ!~:OGN!v?GIu^0s&M2<8I+*P\QdS.ӓ+F_"!cт>֤ҽ 7NWM9E70Pi:_g X'R+) &u WPH.$AŖV7ЖDN-4 W$]Qg%Q TmfdlWLjQrC:$ ]JlƢc$">; k(aFVb2CLa1`x<R!"eF ԕYY(]0SZpkhj~pu5eKe)JPJ"S 0)58tEBJk0-sw!ְ}"%ƒ1Ea͠B7,CF'VDc,H4z5]Aà-!T !K͑4b`0hQ* 2 ܅4Nh>+ djUB>@=Z'i+77.y+ֶ@I[8LQUW< R>n&TXB་HD(6)5T WZZ8F7Z >%%zKe :㨋\888:?Cw5c|0 'ndʻR4P3b( )8QȤՔ"h`I;?QEEϽF:DEe(Z`=%\8ЅV"\),uIi'd! ё]AR Z"n[ L/ \])@ʻV|4kX`t4[V\ ~j9*Q nx8d pATBR%f2 N})@i÷J­.aሙ)$ xCS[*uЂ,Bፈ*y4#+!<+Niwk 0(6 `FP"zp:T>vB1w v[H_ Ok, "<$  Xy .D ޥ+1(A:ZL7_e*b|A%9Rn@ [Fϣ `B]K P T yR#1$ag& (Ka9dA RPUR !=#[gv9>,ZR/l,1 1_yHn,{HQ v ,"##NdD squL;rT`*ʦ)ciT66JAdj|BTRXa:qB1cy}3&^8?i`Ղ;VP 3jYSWw!T8MuQ 3 2^ BX88,̴**ZmuA"LU Lc2JG{u(Q cʃcF<@ E&rZVȼ"a|R(T)pt~?Xe1FǍSPHm0_  ě@q+2q/=U=**`ɉ᫘w%,ܶ))e%)8 dx~ƒ?Fn_8y'SaA[w5$`QD>Zb =N x@6DuԆNLPK |Y]!jUW5t3E`Bf An!U Qq t(CNpl *N6G}`6Q9H#kFV8-5*/8t((Т xd.h]`m\Ad&YbPM e@P1"+Y4̹+ c;gQʐBPuDYl9i?VHjY ̢jTRpTf4FWT*-=+^^G-w (afs`@E?&%T0Ft.YSjhLb1WwjzZ~Eiv5].L@ΚQ&nj!f=0p`-Xif Iг<5#e``Pq}(˯4 |pPaRD,TPs1#JW^y5h+:)R(]&d*DPTaH>"Kk=x$Q7Q1)2'ӛ +Ip "@[)K!u/V2Zة T% EiQT5zֳT`-`\c#d&jâHYa=+NF"5R&~FO 5'~T&/3byPdҧ D2)I1e ﷬ykP*|Z#@KUqZFe ǔp((-t8E (+B{~)#"[Jx~Sj2.iĒPb.0qH*p+@7 .* 9m!$bU[.u b] 120+&)"T$1D %)Oq'_{qŽޛ܏[3]-L Ɂe:oVsZɇK/B&=d94N>2DIt)gt2`&i2&i2&i2&i2&i2&i2&i2&i2&i2&i2&i2&i2&i2&i2&i2&i2&i2&i2&i2&i2'7&u>t8-TFڽxPZd:DѴtLtLtLtLtLtLtLtLtLtLtLtLtLtLtLtLtLtLtLtLtLtLtLt$LIRG#/^F4QMd:Md:Md:Md:Md:Md:Md:Md:Md:Md:Md:Md:Md:Md:Md:Md:Md:Md:Md:Md:Md:Md:Md:/Ey߻Ѿw+59aNt]{vI5Wx`qՋ6˺__I7^].&)lw3l˩#K X4aH*eIa5/ֲ  (~a{ 5o,>t?_Sv AvsWO">c?49 iӞ\G{q};9̶[G(~1*IXC<υ~<&= ucw*a>mj񾼹wtͯoƬ+S_LW ESr %PBS,# s'[*D׼M4?|GݝnFs?@\b:_N~/'oauI1?BzZ/.j(FӔ[}U.}~TUqԳ,Й9'uB/~A>(0#INYBV詣`=Jo X6|$`΍ŲRKVi?5Z+͗h- XM_XcI8h4PXΘeZYeFVX XI局UX|q6`X&X׮FYRڊa #.HrUzHJeu^/ˑ5vG:GgaV5`IM'XqƜPN,wRYRZݫJZ/Qɛ. =F=I쟛jέ:͌c8qUYk=1ɦ.Il,`8ѧ`6}_>)_0[_~xvsa7 i.8& @b悃Zgh.8 igz62 FZr`i=:/eiG@6h=3i^ lE X1n$`7Bg;5Xx7*IS5^'h_([oWc \JvP'Xu,H :.ȏdb k$`56;kq^cck V#6#˙+^#+mg63mbY9G?XCl%t븐z,n.9c\ 1+;~Դ1=FU;ځ[e5  0ȷc^pou=cBYW13^6X!v$`PFUB;b'`nMvvv>;-;.g5_ 'Ft7+?m;zf&w&woUX~fW5tzFz19]auoV#hc_?^2.]Mrgx1?X*]E Xm\NM vW_f7-ѝ}o'7arpt` 30: ~]tZ~XLD6.2зLbItSd}}8($8=Y.''LN?_g ^JXc Wt`œ2yg{0_uZ7Fޙ֒xo,7v|yוPO6~}m@7j8ɛoޒ"` N@` ݠ*\4]_9/a}_&G|amwg'']؞lfoi{ݟorYwJ~Nl@1+FP>ls̲gLA8x|9@|9 lyA4?Lpcqm%KAJ]-S&z:j-j\&zv{R{J?ts6m=suw&V0e5٫je-ENM<0-^gr^wE0!.YL&6ĘERyB%sU-&~Ǝ=k:n\i~LUU `tgk,VF,rV%':rS {=;h3z\B)PᘋR2FIƜ0+Umdo+5Pc2x -LRiquu3Kjrت̪,.̵`RXpU(U1'ZrKAFc2c+oŬ9A$jͱ;x]%Xq36[C=#+qr=Ӓ26ӡrQ0F*ZܻAЕNkUWB›Dw0!1>k`1 Vgke 0R=c[oH7>#VgRtDz'|W׳Ƿ̜u!qĝGxyD1hbT>"WGHIP^fU▕X9P՚#αTmFj \<xrV?cVׂ乮n-<+^8:@28:A=/ΣqLLjH) RYKIEvQJ]L.B&f =j/X'nBKȈ24 ;ff`藥&?ע%=c{J%WUnON'nYU:732{dXj+!ʍ1K"eu/Z5Q=V(TTnd*Ut Ѹ j&2޲I)gx{\: a[`QQ@>p*p-SW rfh ݬaD%Jŀ!VMėy$\ cSnX"Bk)6O6xBRYAg\Zj)4.Nt+TRRmE^3C8F@H)O.0CQ7f-a)-% Ԇ,ZgD@-]i%%Ղ[y)>I1Y6Eb=f +WH%AƊՐ(@&:xd6͐ Lfk=)d|=AAI**8ijXUpktUGI/y^F50!Юm.6S57._s*F+1d a#BA$4RD&XInۧUzgfEL*Sn)M'֔C5<2oKa ֎\"1dW&KsJ 33X[v< ֨<ބy+XCeȂgJ- [ϐdDC2GQF UPpv魢w@<@7/:?8A4 9'SaFTge] %IO5 .%T0Kէ:]+IuL( ')Avm@ $D/];29xZoVM8Md3u& |_nݢńTE~ǬI@qbp#RTPʝa9A!'Ddǀn0|o츰;-x ے(햷 wi> ka%a15WT4T0>o9@~W2zܞQ##Ue =;)<ޢY7 8/,$tB\X^ג"4DiEPyu*g=VLta8Ѽy/>"$/)2{[}Qx3(nCe6 Ws2YЩ ֏U>^׀Xź3v 5Yr貚A;> V}{rv3ΣN檢q>e-NX}+xme,"z4%~:"J,tW?J'0#tKLES vКf&b9*չcAXx kTg]I`;iMD0jZ̓_x=2%cg7 F#tfM 3% ds Pz ZA"7HM}R4DMET qD519l\j% MGRC:b8k4Q9FaMH Pf9vM@ J)GoEU"IAPf( H2y+6LJH0 R Mi<y-oNߐ/1\ەI4T D8 qu3h=*58)In> >; ,fWm*5fݟFH!#d0%Bʛ~fz6L@YiFhQBÁvâD6P6%TP͡2xyX@D;uQX ;gT*d- H(:*m4 y-`͂zGy^ægc B߀uEXR)P trՠpN 9%֛eDy*aha/1'ZwY1@*:>JLW=g@UlU?{ҵ?3hɐ++9Z'kNZ&yAVZw[?x3$)B! 6R FxpaVC#- H E=ʨs!}~)Q2Mj|DhV1j\ =#GAXm(T qLF g. JCixo1KLRR:D,0IJ):&fl3EC8N.H,(`JX-*\2伩]Gt!wRu)XoGv[6跗mi+nZ7w `$aO$T{/'ܙA*Z?, Ey"I-%䷩_=`R6حyk)w~ny]t'/'g^n6N?61? >fsukT 3o+\|OR9MZ=|}?umosl:Һ$:h964gcӁ/ߦt,6ߏMBæ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:/Ӧ K_M{Wt:\rKt/ݦQ;l:#æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:æ3l:/զ4Ԓl:(` :m:rMh1N6a6a6a6a6a6a6a6a6a6a6a6a6a6a6a6a6a6a6a6a6a6a6ay!6/#G^pwk}wvU[6ͳ{nmW7]wu^^_=[e3ƶߩ״`La!`V  R"vwg`Á`ébI/,Yz)`58ZJ0_ȊuB0dk1)qf!`6}sGY|A` dV K5CډwKZ-t߂_Xv!`A |XJnFYJ^9Y K`@]XE-,{cUKپqB樅u>xp{]O ȲFVi!`)0B Ȭp,y5~!`Mp-%-e9$Zԅ掬n@|tZXRDGV ՊeE|<{8XQtlnp8XKYس^XYJxe<-lPg@ 挒#W,/,RvBR!a~!`1R,Џ<s XsVщ`ܑ7N>֐G -% mEa)`֊{c֞yk4bj-n!`Y> `vbe!`F5 YjKYJRb6`y[w Xw iKG襀ը_G*u8X ĬUb_*طuAdo̊"--xoߵ٬Wug$e}>=M^ h% a-2xwm$]Q[d> mP~QIJh /'$ܣ27y}U;>"wcsmݏoW.t߂gK9㿨YRIɘu=65=ys#պNmJ͸*v2{[|vu}3W=`]Ϻ{w5|R/rslnx?2Ǔuk+QPmO)0ǿ-O<|Mv|4WJV=_~諩xm}}OaG.N?enOH+W/ⲿc 청߬ e`C6Y{{0bme)`I'"6\ @-,tC ZWмg_.B޿w7W9O*Ip?@=r<? WoPnf-Ts?~]K~f)wmWe⋎x틞[v~3^3^IկߞsuIJ\M~_ۿ b?O׸wuNܻ4?1m:WJ]Zڮf}},h%ޯӱB╕:ۡ~Ul_b:"ݧ~{-giyB {-#-ocv.mvW ?#=RS=dP1B"*XH#P>>0k7 *|tmQYOg#?g3~$~^.ps<}BZ|92k[%DӛܑM*RYޛ&U'k&0a%Z"e5[Tߒnq*hXJPTKov{\)&!_Ya{{RV(Jm\L\Xk!iN~⦷C#'m.E\m TY ۚG%rԒ{*L6 $M;okEj|j-%N%c2TҸ\#nbK1p &^r-s,O 伾x F5KKl]3H9)z{N`R:pJR-I#c>,fLcV͠6)rʺK%KVՇSʺl3 hB0KQכvV%Sq7eI9w*piôL?ьhѡtԔRM>RX\xŔ+&p,yrh.fkkkI|.XD!{=Xoۚ7&xs\\^L:ƕ:$DyB )Sme'EJqY\ֆ$*S55ks(%9RlޝF5kZy}o4l.2 î;TbeX((E?b{B5o%@Y¬=MY %$Vٻ7$Wa?-yؙ1cYȫ$(Mu&R*vwlGefdċ|RYK IEvQJ*H[|&f =G{'n4.&*68ЇJlB:J} 37 mVL8Z3l6oB@tdPh] [mS,Xˈwhl:^MiebՈ inp k ts\1(l)y pȤ``gZ9 @qQU퇍`-"-7/ PR2AAKHSѝQ3Fwiw,]Qy{X)R6 i;U Rwٕ2wº3%eL$f^,.zFjEQ_U0VrRD áyVf *a P NQn0I&!kE~\JΘ ljinnv׋v2bRy((c pQIbY"j9!u ӷN}+MNV6uy<UbiL2^ \X8q?,}p̴**Z=낄CKT7{D6L<&#q &`E:. UtqTA"&$iY!YJvvP*1zOAl@ؖhd 嶍J# N`8Rr"(3ZcǫzDI8KJI #~ЃA2C #RFdoCAis,0 EZdK'R% f.7B:hϢ; q,F%k@foj6p*grM[oEX@W0390 "}&%T0FC:)&k1Ҙ;d&3RO࣐X1 8c$ 󛎗4Z6l!& _[-:O1`.ń`ZĔ!d$>D ):(r -( _cy](0]hwEBީ쌅G8Øa=EǠQܸX7G7@}1ƍ4(9RE#G>!9}/loxj2F7psHzYŲ^t\M:_$.{Ū,JrYP~޻~1os?"B98YN=wf3:67r^n+6f Foaizz3^M~Zxқ| aqg'kX\lEOS@\Ͼֻбei;}Pt܎I %sxPR/y6j()6<{5NjoQ `e52B1` ~VѨ8 j(U5ԷHKdUCU5TUCU5TUCU5TUCU5TUCU5TUCU5TUCU5TUCU5TUCU5TUCU5TUCU5TUCU5TUCU5TUCU5TUCU5TUCU5TUCU5TUCU5TUCU5TUCU5TUCU5TUCU5TUCU5TUC}j(WI ڌG f<֩gJ_kC}j(˥U UPU UPU UPU UPU UPU UPU UPU UPU UPU UPU UPU UPU UPU UPU UPU UPU UPU UPU UPU UPU UPU UPU UPU -hgoX/s:pKm$P`|Ƥ|LF{^p 6L'٨!^u'X)n vx7Mii \c-[cf$`iuV813 Rb  lϱtnRRb$`%-,)Hj8=vH†K:'59 XO8+>9` Xf$`ڍg%CK*g# XC;|$`u9fG{nӺ.~G%?,Bs;K1H*X,jo w#kR cYc1}p+V 3qHJƕK*ƭ3# 6IHZ# =W?~,-K;+iXDiܯCzSVQݶ/֭3!y8we:k==ptN?e~ (i7`jUOz5޾vAp%Twj;| n?qrwu/\'vWkӺ{?JşpW=3rW z5wEp "^=wwWw-+PnD;7w+B+swW@luWߎ 9`cƓ ӣaWRHswWo])FW"~wVBb4dngJ0yp?-f,:t@9jCBWik8^LVϿ/kՊg׾ڼ_~=L"4at&7 48LݕLg_ļy+_096kBL0k`ƎWwg7|mۏ﮼Nj1+]}xtTMJu~k^_< zV.fMwmY/g=iًtWw:=JfeIc\<_m n1 q3`%[|~3/_hv>+^Lz:'d9.>vF?IaȰ?} Ϲvi>L͏GtG{cUc{P~cd|"dvz~]o#tuǞcE_{A,Y߲FJi|mCDBP1SdQG{c5g4]>Mn$=i{0'mQt |Sjoi]D2Ph36fԣZxa1Fcö>[SsE4JxUAO5z:y0HףE1m+aY\aF '?Y(@.k~C˚RXe}אz"/v_|f9/ 2-/5>r3e gQyӠ6e=8O҈I_}ٖ;Ltq=ͽN'L Ԗ~m8 ֆz L=^5ȲGeYk~a{8Rl!-\11l]?=sׅ];bĩcWkHOBt ڪGmALi߾a̟S GIRA|$ލUm.m߻ٹJl뽐?5vbn8A: rH@>UH跞 H6i:%b1[DV{5s_p;߽_^=wu&SzK˔<-SGi:V ~f9wpt'~b'4Sذ'Z7?S6IY3^YV}xU:^u='ՇmCN%w{EKG{f-Mc]Χ7h=Kk*:}JWL{Jٹm[ƙ 9 ҮtҶzжp6ҾF9K.^t1~>S]L󮲔TL54D4fn}͓OUYyVޕo\g _ӿgJi_Ӻ78\t<np /fYA_7\ھ?Iݛy}Q݇&7ywPak&Ҟ~=(K ףC/OҒu;?w+F-02< <1oe,V -Ea.˜Tj^Gns!kO栴 2_z]s )0o>ťn} ٵ@pw&-.f;GdH]̻iCXg\귩*y:|9C_oC1ޗvcK$ܾ=r#7:b}xcc^v3 8Ml ^*%fXl!)@%D޹X_jt'-7 xkBnV2??F|ʿ/wH‡vɏŒpQFloy~+a9G~tzu w^7oF*UݬVR@&T_ 49?-HX#H{>m j հ)O4DML}i{i7/]n"MCp]7N&]^ʩI9+p8y|~haCǏM:)) B7A& tnM# B*/./8$A"2yJ׆8ea*M9!<{w Ah ޯVY*ZV:%a>&%V5\J1H&B|IiLԑ:x'<Ȁ;ƯMޡy&t&oN{ξ+.opniG (8ouLO8#pfoԞAy񨧳 'P9kWGf>w{uƚ[2࿪bX @?C5聜;]Qx+ ѾG#ooJ 07 ?908hsg~:A~^>_#oAG~.bs`8QeLab~b\bl7j3H}uCc^(ɫUw.7ܢ 3e~K `Le\LWLWioɔb¸YVT5`VJ0'U(Roֽu9i4Odڷ(4"8%3M`;Ҧ\bh>}5&C~/M101MB wO#{3ς#rɃZKa+ɍTL5z\op%mBm-B"w&{7"1ل1q"~(:ˊJ:/M\^[̀򻑢7pB Fr\B/~q̓3gGPZa͒/>=M'EQaN. 8cukQZ_U_~OF_¹Ⱦ5ԓ'4s.8'HEN bFj?nNfj hUHTr T9 ul5k,jj.wp7tBRb䄪-1dƁ%EζIޫ|k~~/]%$(jjϳ2*٩{x={=[/66Q<1R,a{"'"SqD];Vd%(k\}3jg۽fTrsRYr,ޢ'F'% '#L[bc4K콞ۗ%!~V=Oss6ψ48`HJ c;NZؒArx_,w]컎^;) &yMf\lWHtŌkޓei}^y'G !S,_[Χy0A RKwDttjäpOM#6A(Py^ƣ(-&!IIk5#A2J,s"0wKZ̸iq}Ik7q3^("40PiEP@}njAn:hP≠(Fqa#Qu)kYפH jޗb8:-!()'$ (\dE-y8L@TP}ɬgJ\><@[(oـ8o)AĨ|> AW^q3yQ}{E-q>=~Ԍ@ :x?QWH;{8oQorH8rz'q[ OU I|*YITL9$hޗ#г_GATE{}=~k?(/o{|-7C5s-.;@\r7'_E/%H t='*dR$R9NHHwlĠ)ޓAÙøO[2 /8׮XQ ։H4qIF ovgl!=C /i5//J[oj4ޒ`Au֛Rڣ1Vx_2]Y`]큨h biý7S̰Έ FVK6HCv(xt[SҢs Xn#1T+, 4¢x4_gy]o-(1 +ּ ۻ!䉫)! Wy=Dn(4lE*ԅA CuiiM, fm Ht0})>truX͹Z ܔOp5%-wKo|s 4hd$>Cá FꚄo>Q9&N3v#(;E "(v`)4FaL%-NLfp6 /up}{` \F_p`^ڣ} ZxB=%k7Ȓb*c]DD]yMUs QrČk~(𞤅}NYJl}\Yb9uJr'Gw%} ޗbp4_ǫqd<M]",3IiF ȾUl"޻3g̀{1d 3/xٔ%\X0D ZLAN 0RmJ?rOƀPbOqZ|$*|̎Z}5e2R j`cIaXο~䜣*ޝ, DŽt-%a/Y$ГNWm\h7\N1 Mɪ=MV.8+yu١h Or|te]pG3ޗ|8VYN?Ʌ "/@R0 %%Bd_39ձJɢ W$hYV/O>0ό!ڮU?{u"16`I:Qm"Ȏ^1 Q|[Q^-xJ=hNPn{榇gՌ:8$"FE,s9( M9v,RP}InrBN {RwHǻr"M4V3$R,tEnƳY@KW/qys n:߹ *^orAXX}Th5qOT?W!f? nw RxwmӸFoq i;:C4惪#zCXi,UOfN*W-% zB} z ya4MŰNC !t-bIc;|0KLH% Yրcc:NAO%ysMP2ɴ/6=b]2J'" w-76pYN"< O+z9ΞlqclJOa2J _+S062QBT3gQ6Gٽ,;@(JwypfZAձ%0W_$˰ Ɨݞ)!Ǿ=ōL$K02Rlp3$R҈Ki$tt*l(Ihmg9Lgg{ԥ&|e_sXv5P|6d̷l1d7P ZL(č3n\6*a?t1@h4yy?-l]e7nCVgw:2][VGZS`&aI1 .a^bE[~Xfh o)e7z?Al-0u>xv-(n~|e]mfFŚ(#G|?d Am^s,br~?^3DB5m+ngҌ!KÛb+ LGůFI2~Dχ5d}8Sܙ~U),Z= P:f:@,ňP >~%)yy<&KF?Jf Ȃ>*|=3P^M?vϬ, SNPe0eɛ()n_ `8S_8-’np ڌh0A-AL B`!^: |${w$}g>#ߖ]FwM' }0) 7+9(ic8 RP:b %I!c8ԃ_Ke=y/`%wX*zxia=gIÙ 9rL#Ubfn] yXiPő$Vsd=ղ2:|1K:fK,3g./EcSgWE33ǥ#&Q(B_iE[I3c Ejmk}ɻ8aKcAoJLlJ7xn˓8'7\ijcp@)3f2mAyJ@q*K1L5 aln3nLЇ ~c$ңDW7Ngu,_^tr4&(ѣ޾hqYJb#%*$H&7y wˠ1(8iiAB5pVhX ˤ莜vKvL.y?sE~ iQ>ܥaTjK\J/_Vv':SIJ3*ms98Hr#SuY//(֓v7f$aAO"FkEloqGcj+Ȃj+̴akw~›y.6|5κu& 75yS81H>| * MԾv}GԵ ]8.k/Y>C_sή̏ Iڪӭvv+'pww0FܕUˤYV̦lm'y GX1Jx4QbXH1L;*؁q'̗cފz̀UOa$~Ix9U  Ni gEP0 HϷu~7m[bD Lܣ\dթvoك!MII)ôEP$"((!|wQG HـcrVEl?;"iU18W;)AOf. ki g:屓 ]bH.9.u99tDW`/Gt ְHBBp|P>KZq1A;l'٣i43hh g2퀂 ܤQN^?+lP\yGE_3tHxJq,c"_^Ȩp6a׮D6?CcAU޾jQRQ[oc|Kp|cvjLSK[m:90nH1 R1H rB#eeϲk(/W\Gˎ=Cx} eıKRTDp#tUԪ)*YYZ#F1h3e"cJMYx1>xEISNcbyX{ms bv#O"s^$D~ GQچV hR-a"ND,I=D}yd!{GġĹgҴ4xy4RQHG 2% mj1q6N8NR RlL҉}vQ]*-'R N=)ثu 6` ŠFQ)RZ * 3gp`*_kQ< {lYR(1 dҜ1GE[nC9;qKLtK1s 4a8M*h›t| mۨ:J`ߚY]uWS!ZjiL"R vDi@&1N optLINGspN ۠GQKzn)q<9/:u|A+$tI0{9b=mQN,U\d#;GlW&騼L?7#qimQ,٬|/Fwz/-sUu>Nq#=$uw;4?hlnF0$Kؔ 177]lu f毣ew/87{sv{48D<%Fp$-SFk ǰ,$;Ia sgVko*ju~a} 6?ϲ 4qՏ h'z'uτ9vͼP`{˂Ǟx2 o7؄^c`4oӚa8?򞵷ɿ"v8`pY`?`_n&;Dm'Yb\vlv. UbUX/(\EIBɁ&.TP04{?_.QCe, c8<3GW`$,!䚓<[Nq14ՎKt0X9kLU>/}A)V"{%0%FRhaF˥")$Ϊdt@a%k^׼իpl!^Eelqg8:<)iT"/?S'*&5 Ob;p Qg,0r|ƁvFB[qH!B %O3Vp7e!8e)L* }]gkkёe"ߖ(`ldّ1FdMs.{BB40""J.W#89 ]\.;Q(똚*&>\qO(cTޱ!rV;GaK1$DjZ#%T(y5=,{l 8 g0´nr/Xn#1T+, h(Cy;1G?m5"'lw;ᣤ!䉫D`}y YemOp(20G=sN۩ Z0B$X׽s(, XKV];#q0x;4J8. :|%W$6JFT= ޏ1#n9>mGӖ aS1Qd8K]p}0b1}^ \g:I pQ9 {Trln'0iJ&Q]5ysSQ6'8eZr1ʗy+Q}crlR',[ߜ{1 ̄r}rb`pqb uTԱԥRFpt$D m>n3b|%N,UA1yĨʱix΢ITsˈSYC/%wOBP %N׼0]낮>fFlMWI?x`O9T!\ù4tѼk5#A2ZRՓz {`ϲǏd4?|rM,W`*j/J:~Z@C;V0bC|;N9Ivc,Ӿ>ֿAXd:J򉷝/̉'NnkHiMj|ԟ5Z:\d"J:Yp[wkcrgԙyOjSK ڲsj4Z>P(L`V4T)%ꎤV@ЎwN=UӼZ~4<Bw@KٝMCf,MjZ7[,eWY\ݖuHww+/&=c𹎍pqyqzNez;LK^> 9OYo*σ0q;^Y-nUIYezV_LrEwj?^~y_On#pp*ĉ8$l2{~*FGUji}2;qڊ~ K |N[/O?-~lBG?ž~/sϦS<xI0Gx3[yӾ iazfo~?}ǟl>zާ ;hr7m@S-jK@G_/ffVml'2ƋY1~kvD ny{_КHL܂ VW+*཰.QY-B2oUp.jjjS6lߧ%/ڒ s,znã?Kcńi[O7l6͟d;S8D>i2[T*' ecrWdu,o6R,Wǟ|um2+/_P^_lL)ظfʔ`٤m9YT@9 6Wg3] WŻIKכ۞ep|zKmy&0U\^]^2nrFK?УFWy x٤'A_e)%;TgTmPe*/蠋#\ۖP4%gubax*>1Z5+yS Qy 0Y "bV4pqww-f1XYC`<>wӃWaenȋjeis>ѦVLp}zKf[桀ϲ%RPZWuwŏ1ZYk[}syy3W[%oyD[pFe A$TxČwt˼wRHT0QΓ%3kh]^ãV&ˁDŽ`Ca:9DqxQEQHwDx#X%I &/| CԖ2v U-F4zcE#ctTI}fv~Vrzݐ4*}Fj9jX2,Lg'9(Kw3\0$?1(=3pIbƀ^]rsHt!H, Ao!0Xב1FRTew_"Y}ťP*cmŢY5@9it p"A*3RbuSN_3If ƅm%>Bx&c (yֵ)q` ]ra2.rK V*}Z<D)JHu-a1χभ8w(Wf[I"_ 0V;2AE_QT/+|Sx>qՔ:[LXK1VXUR("ץ'[MzOo`Q88EcTcr+[YŮ.U&c X Nn?a ỵb yk1#[g>Vj1vڴmX4N7#ռ{c#f'/ax,'KZ4D\D]޶59_NJOy>!"YFr}eyݕޟ@ ٜpԹ]QgлOϚ˟}T)WbHg_t:dv7^vSm-`SVRfbYZŒ4g7el}Y9wjpLbc91n<-tYC4KQW~Lt%-t@AeSzѐxd6 P]ug ੩dkmhHD\2TCFvl4w> CoD"g"1X8L(Xc(jXdVU]U:[IZuMzOr+3F=0U;AǘCkgU5,*~[,q^YD"+1*Nv-3k+@ 0X#c X\+ }Cwvrk|6Fn8jEtdZܸ"C [$Yqz/=HJ5~߯حnhQ@ե*v,>WsqT8EUY6K?.|6eo >n?E`Iyy) h:3A 5CjbDاX4 Y}I'Pպz,Ub~ۇ tp`:(~n37ˤDV;robԉ8F=:x>NM $[2'JV8AB^Fw^bcFļ"G<Oےީ5ʢ)Z&'JV8E% f %C@Fpg9˝L T[%B>#A#xij3(nGN޳Vʝby"Z|U/MAFNy<$ʇ\xz^9RK-&KCiCq "B ݙ6qa=ŬkWO_zR@ΥS|̰K n#Wb!B,g2Iީ@|[9$ Órާ ԃ9oo)" zy9Ei&hLЪ_ nMS=JwtOߋތx^ r2>8!|u.ҜDXfbȺ>{R} t;B!'yC!IU޿pOyAhֳע4r׆\o5(M9-ޔaE@%=ow <@qft .ɫQ(_孀)Mݫ?߽{d"SurbI漣0")+P6C AV2P"}.~֜לmʟ&z? 8pͭYQv mkiٴ4c뢖oLƟER1p-I2hׯ&d%_o}Yo jd]LՏc [Ӆ'~73Y 졁w0Bn]+wa\ތS7 m3/p]Ⰱxjbjli# *QE7Faeo.Oj{o<"IA#M?McO;O~RE-?SD&ր|5F0/cqc(29UXYE4zgՄ@h2'Y-LFs+}1o~l>F[g=ALMǖU~?Z-elb|~>?ƶת囷^Dׄׄj}poVǠzǁa)5.\eqkĿ`#.0K"y3{+%Vx4vb|69Wm]W#fe+j'ճ=Bflg< kB`'h k c}=vi1<7c/Ǵ" @ eDBKFH[XAHvlZ~o"roJswUu[s0ϊ Ϫ$W|U[ ݣ]P%w/Hc3gp@EMKXg? v"qҰnr"T?֒eW/%h{N?*|:NW5m/?0QIPn%ҜA?^KwQS8v24%\v41|p `2)ɲh01Rwi-z[,(6 .Ǣ*^ÈHQĂg\aLf/KT y)%]f "cvؘ[~\ u0DǺ/C(w~k-Y`B>38( Nڪc)IG:ZX\12'Rz,UOW#;{L]5V(ZMyG2tU,tTބQN]yt8!qs|{?O/zUS 'H; $5bBa>AISQ7euxB\YjI,>܄e"pIF"C"&%@ڤ;Mi!Qɡ rBªB(;U/drQ:ўrTkf5f5xgZ:5; : 1M`6?9d-l(8J@(! ?J\ *)g A`N{!½[PubU%jQ67+kNcSZR4VAk{¢ Bt\mdaMx\bnm:`_M'[û-u3Nh |N :Ќh?lg:zŐB|~;B g$АX`!L)6V8'\:t|wج՟v S&L"މDz(Qـ`UIVOčkEZR=%Srg"bufoҠ.'c\ɬ8,aܤk.x簹р@eLt$k;p~2 nVĻHpP=*>_$߿YXkZW`?~/nFe\7?>yT]jL%eda<'Wti}yu]n_T0ELc; V C϶_>M[ ܪ-`3"ޏS7u=UĨf~zSVC5OG zS(tQV~~3gesp< V֙ WGGfˢmju;r5כ1%32;`O="un$p\I!>IOaw*JxPK: )!t\XMwvB)w`Sυ88; u7KxfXQ*nEceCdF(zv_|\Yd.C/פHI|+91x .OM2/PBIy|pצūR>m>u ,^7^yXvBYn5)ʠn{==З;; ^t &&Vey% JX3f4 ՞>焳]=,%Nq*ysnIZ!йv\3l"&w*FI@3A*l3uKc S,q=,r?xOJ_u~*YZ3=r4G1IhGLxQ~ τ_}3މ]T:]Yݛ/Fݪ覘]3mxyXzl0SeeTʉfR2vho ɛPoSэtR$mCbg+=T*n$!RӾN{]gtPo  5f7$- vxK޹Cto 4tǜXĥ [UAH՗>(_|qEɁf.}H2mV[fF} XIb/l$jds٬:L CE[eZD=ܤJhGUҎo"^>7?JZ]("DUႛKl?s1&sltQAFRWRCYX;n{\RIZL5z,NK@#оNY;&[%\BsݾX+}V)z1s(Ӳz 3Lkbqy콨 -LmZ#Ł;}uHEB =]4d~1C&}o7{(ROzNVsv-KhcC 55?.Еr1_~ zB=vzW:Usg7_)g瓾Q6Gzydw_e_ {~ta2ofTyf"ýʅ\͠cǥ2e*AaI,Pp?7;'j ?]\ma,gY_-PkzP{qYYUm |xu:d*322.p bm0QҔC YoR\7w %L3~NMl?YӾ"Ȯf{>t㠓i"M[4Pc@ lh|`@yN2~L|-v@Hҷ\#mrF\4<1#;?X]ĥ^}#rV:4L{dt?TAdt+ q)$e')b^S"Hf$B0)G>_1>qpQv}l=IFZf`f= 08:(w,Rmɉ"  18h!VWm+o2dyzxTGُ}q aŰ5pA 7^:%&()\bl:YʀXe@n:i27l 0bЀb]ZcR"pLBtL i@ã28v{'!ln{\})ZWYgu2Cl1S N1%9N!&k oUEH(s9^vc7 1].ˍ 9l+"yyy Ę@%Cqi$X`Gep x*97CTXL92([ leR 'V^RX &4mxT/s8ֶϓyz%Q8](sd'6  GT+cP `5}Yx}Bb>ה_g/\Y^zޫ4~#ֱ$w" Cg H h|e=R  mCJF' ֣*+p>ZMGR!xA9a#^QqtQ }HuQ# dE<ѭ cgg-2bgQK`2hh0t &9^SϘ*loäEt'C+!gF>] Prnl{xT3ApVQ yw+'+GepLɗ!IEǙxt/ã.8E<\c[`l'P֎jsZ! JJr|秹R H#jl־SզY/̀+c7ηQlW֋I5zxTB.joVVh]FֲhxYAy[di)f] E܂R,O9cQFaȆGepX{%w9Ŀ1]kb*2"2 *ac0aLGep۪+vt!+ӯ*=HXZl`G%ȃ9HQB6QqQٯG.~BIԶ1W}egϺ\^"_\YW=<*S{Ռ~#L7R !nQG%cաwã28'1QI/r RcG@)0XAq^9<*t9_[֦@)/kMP - ɱ:X~$gQrYyc@d UΗ>>_V\T%7_OB"Q*\'8Θ#F2mxTGg팛V.?46ysdZ ֢9L/%guʔh,3%jc< .lHWMv0 v5h'D 7>oJͅ(ޤY[Po5 @Fk0worW9"Za~>W6Ѳʭ|Vѽ&VAd*_u-735Ŵ٪FofYijG7w|]gQ25!w!z-R7h2%Lؼ%7G2W4N7.i%{EfK ](ꐋڽ3SVDd֩U棃5ȵڠQHhǽjH˶A>OE'M)SpbVS`f!w 35VFpjfoYk"HiW{%3XG #1!1) Dgކ0I`^Nf+yF#[NY̴h;#7Әg;Y5[$JϢ#bVE8d$U ^e }g ԙ{cӤtuG $ GC]]nF LZe =ߚ{0XRGMcRFZ*(1@d{47h3a9Է5o?aەS'l'@N bwb,f ܢRzF˗7NooŽ᳝ɬ܏{poJn95T8b 1 axi^7htnY$$!:6CgQ1\QN/3]VAPgkř_~]$8?!-T%:%{MKhg74XbXC)qC(֣u)-a31qʵ)8r#N70NR!?0!HtF.Fr.<Ş8b@9S%f ݾO0?}iǧM xLfQx,~" Qol-mv:=:)48n/\}btqg-Kv~xwqr'o܇~=7ήg}?2nfwl>4il>[7lMrz6x=~}ogqgtm}o4Z%|+zIwXEﯳ@T.ud(|vV ݽ"艏skXy?@)wOZd4@m1|(NB.6a7tY GnRDć`SF.V޿'T`g ^%B/.~H{y pRؤ}VOv2^vH,Oj*Hוo(dVxȢ249wh,H //w 70u)ã28~Ӑu=3]_qg%EGcR9&ω׆us9[@VdV;{Fձyw}8}G1 K+41NA󈘴v Ȥ킙0:`XcvH!j۔ZW";8x8XLű MNt3{=K2 ,g]zxTGt0fh:T ;ʪ2Z: )\`(!^r, hqx,4^,kqhozs/PտaMBkӵSx toei7]ܠ͔D2T,=|ku'V!MA Vt^ޡg+ߑф,Gj6;ߏ? ] ]"⿙tI͗ Ⱥ@Arh8Cel )mݭcܐh6Jb:V51՗6kmH / ~? lp&$YC0ɑLWMQ%KVK$eˣ"Տ_UWWrG=SF.Dl2+G0zxcNlg6橜' c,M 1)s{v>7+@JkBWiJ~".vSe食6t;mh.;@b 0.Fg`.m"o)kl%yZ&MOEb@ B`xIK|윮r^abW&hin HYFuIH`iE@!Xn[C-k՛{|GFY /_o~y9n->aںT NppsϑIi w+NN:+ۣ333WU fa4{"+{'9gGMQf̳w f]h|7Ÿ"teQq=(>HI5/,d=Qhvpv^>OϢ\"3}վ\ YoUt8e|*_G`i=`R{+Q\w#*n 㭰jA-ӨevQ{>?J= ^.[>AX碴^h?xeKotq6$zL,bo6~6dToT>G).e>wgm6|$U?g!!R̨IeZpĹAʥ&B"*-^GKe*ʲb5*^x> [Ҕn]?lixf7}ʒzvsh7"_fï&Tއ"$*p9 bJR+"9`^ %_3Oi ۻ87 w[}V,qV'c_FaW#:8֝Tmrb)V~&8ߖBgN51J|A[L^jʈhA ##2&"}p?v^e>3w3No'uRd.ꬵ9c]03wh5Z`(3*h2:ڀmT 5 C *cΜk_VX=;[߄ӿ~ ux5|wozЖ+.R;+U'Ou.UL2OwYK#)ve4`6ZP2*qR9},/Ke/K1&Uw6/r ܜɌKEz&x;1Yn}I8wq.d8_Wavۘi!$a?7]Em WpU2ewLJ.S Lzwcl<}vmO[Rhvlڞ jd"4xV@kYzr\νGW_Ԏr; {&^x ֯v;7?D^勅 *Ɓ1@fQ#ռnUۊy,j񴘟}wfto)0a.SKA-hp]4~,HNZ^b)j|`WQ3 .*v?1 >[ Uh u}{.N|>숗7b>93T[fF+f/ؿɅ6V(̥9G'$M=!4}o9fFɺWyQKPָ>E;SBbq6F.o# S¯xh%R]Yl3!g㏜Kh;8hR1M;hB]TPR҂ <d^sXevTsYxX0_g뮻m(V"n0ahإ_߆Q1&L X(b.MhTm\knes1?60w5>en?fU{ciDuG lf?ѱii äd{A0'YsEk]F_oҔ}ҍT RHμIt4QQeM(&8DR3JuT:a-u3Y}N'\ QX4)22%(-Hr8ecĠH YE÷T(z %$Wԅ(D5X:0_[`0cN4}޾ z, Uuݞ8^r;Ad{ @9Vstz=B'ŦFb; 3_ ,{DHh)-˘+c6*&cy嘂"ўmГL*y&mO3:A4LD5P˿+8+]4Z_]4F]箼KuR{hV3{9g T9w)UN:c@w)U]3UN(ز.Jou @S4!j7 Mf1CΏg= @}n^\HK҅t!y!F$FwXFɣ2BVJTt`LDdH dswێZqwΏnXpwW4s}68 WsitH d_ 7UDr]|.!ߥKwKkȢ,xŬ ‚rjfH5h%0Fq,@Xv`|;X``BgWf X`UfQf0;@2;vAq]0O#|[3"I. v9s5R&'mmVhߍNœt$9$l;[P0!#kdM\ ˡWa2 #mHWh-5AQ+ a(qŞX_y]'ZN"ICD2ؘHC#SV:"lL LL b! Sn@\:ڒhI>։ڊh Sю4ҤiN֐:F e@HQBJ`Y&ɭ,eK i*G\ӈ^r&e`bA 3<3Hy'yFSl~َs#MwX'$L0EI%L֖XR?p2CHӨ0X5VT`K"Vx9P9)$I}Ennrk (LR꧵ кzsK`K` \2k@ӦTJ)M0RYyZVy l*'@.?W-vbL^Gq3UxSKdrH8+`ik{[@7xXlMϵuW7[Pׂn)Nux@&5_Y K bM@ChX mpi=h?Er6fn>\Vy+_?Z!Y2R W)eW:ƨk#F}[m+)_&U8L0 ՔʢhÈCxɐO2`Y|TsNX#~K|P\U}9n|A),mv+@/+W/5Xi'_{N+O|=?-g7v&Yګ eĶQkx*i;)%Y̵̞m{0ov:QaBbıf?po&*eݚЋWo+F\F#unQ|~nqЙn/DN F5Po#J#Ց:ݧF{RupQ GyX$Zݪ.E3g=ȅ\8"9Um*$೎@5~JNtG4F󫆛\nrz[(@^*c\j&:"LƙP'Y7dO sѺP9fdI#uwn.*#ౣ$J K zQgo&ɥo^$_l/Ri+=Za_g4 |7O¤H a}6Ku ,x!b;j|Lt~=.cL?S -Sϲ3'.#j,hcFHpYlRx$F{- ƚH0 L?y}$$"H:$W PAc 3 Dk v@Sk%I1{}$m׵ I)&]WP7,3 +lp7ybpSdrozIP88EG-^x 3Zn4?c->]r5 ׶gk.oYW:bWn|_IE5;*ގx]'Ӵ.NкT#9'R[C:'Ő{4t:yJ>ic-frzi2HMӸa9y<WWl~>Ei1Ofo}1c!V{ח{Bq|Jԧ1",DPz2""&ZH0GH8H||9^ A.˱ Ki+~a&⽧!l8q#_6z?h,vL0lI`g-KnIv?[EâD[-7Hi\USUWJ@anB ≍b16qOSJ^]J 5lG(vMWwnT Nml{\f㥌,2|箬3_ܤY<3KZxydz'fuycGimebcf"Umu:VkOyYjgU+TY9<̗9vu֝nJBtN}=}U3W88xbP Ngmt7 ip]N'e9tO^O:{{Đؽ=[fguHѓgiM~8)zr/ H_)\U]{H!CKiS<Kyj@T{E-Hu/V-ũQr^܅/N-_=:tF>|; 'U y K46yW*_+&1OWp{maI"d ] c_VӚwE@5mCϠ6؞nx"8f_X3*Q[LwpFy m*Z+  TCzT5pKi0]C:XQMSۃy7rLӦ:Ge뮷RضiH='Vqf9_CƱb ik<&H Eר]eD`.0TZӒaaR"0KwW:E{US; L("Mx#$Vp).}Ԗx931tak/%8u=Ȼo7y۪lqeYٗf͓hPh)((}U4$aD%d@tY$˘"R:"  ?QYXBg,Lq.9hɘ|@Ndn9'RMF.#%d1]Xؽ-1dcZad (QHLd$fC:łKs &9b6-Ҏii#EkЮ^8x~NhId4j!_NeǓXv+f`rHxʠ e6*V;RXMLrʝ;oB!gh՞l:FY^1]!F{1 (:2g}^|YMjcHl;X25|x=ʐ(yl4./aN"AJZ0la%Yv صDE[] r 8M(rneZGEdQ0A[R.keH!i kSzsUEǓ}>í}Ͷ|aqSv)q誯ŗpl7j2.>vb2 J(>*eI7Lߤl [rI}jc%mR47N`ZzeL,v'ĂAJݽ 4;k]9_eEMܦ}?LڀӘG"h@QAsglrD`p'IF$ n鹢O'ylz^N͐LL\#%:_3켜-&۴pKE9" ABETXLʐu^º(YQfEo Sv{V<|S;4p0v'^)t&Y`zCoŕ/&YcsNoBM^!Izzҫ6/\Y2s.Rix`,ڨze6(SX8S>@9]ȴjξ}s!1*scOU.]swc}fӡ?誣!ckeqsVA-m=3{9A 8>eZIR~>ŷuSW#8b,#j`q;XԂkŰ"p=K[ n:,&D-Xm8 Hadr)|COOS %E&|#*3& ] {a] 4]-.oRB }^"]3Q"+s#ARy'1ZidqLJ00Nlћ@ TuA LU/,Cz!zz_4RHTpVT@ T@ T@ T@ K@T@ T@ T U/PU/PU/P>y@ @ T@ T@ T@ HQ$P^^^'a,I@ T@ T@ T@ T@ kk& za]"^^$&H%.@) K'/ q»LbYp($\bL[t`:},N .L+lY/fMC7L~ WVm]Բ 0* ﺻK!w6[,]qmag',1E4^JJx2%J7$Yb΂wDðq` d@7|;K8]D\(ib%;]v j|0mQ.O.K 20007V` vŭFF끩AF.NzPH6~Bvmq٦FH7|r6=.gfRFesW֙/nҌΎ, PUc|}O"zƎX-Ԥqḏxt Vu t諚:ӆ6Pzfppa-[Ρ@|ҩ3t$ {T4 LS4zS4KթײR ^a-+ =rW\QBqoU2weJRr*]Bw%]%ҽqWY\tURA认+b7*K닻J2O]%)B^Rbݧ$07*+zR,㮴悒,쏻ޠ,֧rUoU&~CTh%ĞpWc_=4%&WWqVuR7n'7˼*jaJu]dD[W UhCX|fVJlǛsREv2͆9sMN6ˢY*LþaTBY M%/4ӨXIR߿m!suJJx6b%{^"7ł9 ˭.":F z)nY.uq?F-4Q/U:o^>z `i=&R_Ki)EВ)E 0ËglMcܔ  Ƙc|z=+/3wfެSyJw׆իhxw΄rZч=>LsU+sCtߕ\ͻ9+QNPL5 )Fu's=xome<)?af  8|(;,6eIIT M,U(F6eUo?O2ޝlpٔ2 xYJrH ?~pOLɑ` L&#gˀ̾}3aҚӸ)^18tȏB+o*kso=f+Lf+5-PFyv۴)}.%}^d&Q$XP!$>5={YNr$Œ햓_o3(c'6<4.7p\ Nz@[Sɼp捛wv҈7Ӂ2@w{<:hZ.qͽQ/=Ԕuaƚcppmyo33,>GfJ˥VxG]^@6|!H*R[Nˮ;8ᓏ{D٫ +:i2ZwM]33vx3(ai5Ѝ7:Np_{6tkyzaBD^> wƎtёg"I2%\S9ථS5]C^?JR)LQ5ROK9k~H T&_$QYyngS4b)Zk#PZ \m2}iN хmLAQrB3!L)Ca,;h.^I+19V^\e>P˰߿kEx6X2w +.2 #_DžNSIcG V$RK0՝:!\բVv_WU<@8ƴKs42 *W, dzqb @b1 1eE`"B B!hDLwJIfVH8I&)hq~]Io#;+B/CM҇`YVI`JKʴbP.;$#/'T,jY"TNT(שM6-Xϋu[t(M mQF `0z4 K]zOzwcTJ!872r&*) 0y.aSzOcZFM ũ|]'[d'r\9 obsn`;}:ylL偃.ޘۃ*b~1q:x>[enjyk|->"h18NHh!K,`7#T[ΚY>`˸&zݜǑce(6 5E ~~ RO<0ei͉t"8%e춓O&rݔ8C켹!$[e+A@SSv V~+ԮE\uxN9!1BA^ qSV onha](Cծ7p˼̿^l6fVL$=|4;ȎFO?rTl `%-slT^r'yP-bm&rJ;au@T;- ){E"FXjXձcFP$aAq"QFI{+f>GWбUӪUMSnZzPdcZjvvxfh~ެ 0=azТ?6o> -Tl8kޞ|T$l$MF}gloQ'QS;'ueS}+pHXjģe1j$*-E;G~*{d;܈[/d!z hKIYW҃L,!@GV(:RmqnN;8{ j܊^(-?歓J9*Upq^s&R]~<7ч_pLyE$4 B>M&Qt1Hk$)9N1Pa-yU"ʻDoפw.i.8,r>^Om c6~|[7ʆ cjn&rnh] eCέڳ!75'̆Lm;KOߖpn˜-7G/R]!tN / -GeZ1wAl"&$~Θ ,xpBi3F|lWnXluyBj>t^.Uq i)Ӑ,Bf0,xkR704trcbl=~?oӘ@Xa'Hӡ˄8"[ T緱n}-9ȫx,do:p2ϓm%N}ro$Z3-vYGQAqSPvT InB1!eLXό'!JʘĉL< <Ş8b>QJ*D;nCh=Nju Dh6W}/]qHy9.)w*JG]o> :ugϟ/t`5Ա %gsH<&0.DXPq*FRiEoX[)c/,_&׿-2Y띭}@ۂf9!?l\ %vhXb#VG#6:EKTx,(AȽ6'zNW܇_ V/)Sc>Iqpn:sWOx_.Q̥Rͩԋv:p#KhB6dH5;>C9C~`av~´^|b^bJx1'?v"U(C|XQT[خQr% PRj, [C"s&AS"'47H2KFrp}խ7H>O,GF oy@3_ׯ.wmaV .4˻ ϞGBK܋bʊ 5 䢶I.El@( 6HT*c8&0PH 'a.cmݴޮ+]9i?fzH\*le(p}=fZ?k!5xLzm]+: :b+Io(En\FJ h-Lמ"rY4.`*O*zօ͵m50#/uޟE˝Vd Dg3G6':Lk\f&W5(+A>A\sR3U3!z=.j=Ȝʏk>է;ۤaho*r.4}?];j0ho: n0?Va%Xuro;^7M)>')[dK];LMެUx<$T/[5*̆-zS*7зz^dWA3_o3Q>;R>ܾ̪%z(o-<ܣU/ϡ G3bVżbxUC՗ +VU3+zcī0]WN^J%4l7j[e^z]XT+wMVng Z/&e k 6]r<_o.]80gIW(Fnв`MZέtP/A=\o3[LeXb^Gcߪ7ϝPic΍(N*z/Zj Mû/.4uPԬ86N ~?Nw89lVM 9fu$qPJƆC%gs3K$u+0b9ĭؐP66O9T`{"|I u 嬭ob80>=@Y濎FӫKӿn}Vu!e賲0UUY* Sea,L0UC=W3 X @n.a3[K V2Yy>!'j?ip2ϋ 8E|4РHDf[쐳^OqSЏN>KHr N'PƄpe2,8'ɹ{ZDS*0K:':}̓zvq=P_] o_ T3ʧz({;ϗPXkǙAJ 1)L~^S8Pg豾R܉Hϵ!oVWLLp!YOY0,qPES=ysήBklHPLړCپ}@~˥POIrKk+m`waJׁЊ߱ NXg21= 'k.c.jyu/-7sMLPxP"p@F ԊcdIL$\bߓ]j5]D4ʌ:vvIzŠ崦}55mMM[SԴ55mMM[7cz،L*Cb*/p99<˂0!JBHg0@=o 1EZ|K _} VZxc#sĢB\P>wHhL2c[PC +>&&/C97 uV@u5^=YN p_!I̡}#E-cM5^  LZPZ* fJ`ڄoJb\Xx l0&PYqPBxMxi'.BBX9%5{M Q;PBxCHi'A(CԸ 97p4L|1-,<)f#+@yWNKy\.5ހ#'8hGVlmJ|TL[[PC )W="$$ V"MCgn:,-־K %g"(qq#1 jUwӞa"$`v Aºm ߥ MiINYr Řp8C[PC %f'!J!smOf*-P8ilC EײRǼ31"0n pk3S/GiC %ɿ"0Q,0s]`pp31HN&Ŗde[vc&NMzUE+ʃW$+0wi(!oukQiPMRXBΊi2eJ#X ׆9_1uC O|}{Ë)>YŦ}KqEM~IbgJcCØLeυKweBҐrDistI*̭3ER:$Uwuk+\)P!EY"TI—N4Abf l_ͯ0rv~6?ug,v+LM<{],=ml[Tw9=!3Y}e囊1OYf?O@r~[ۘb<)&KoLD" ɽfbS@,P#~-6&jOHkDAiŅZE%(TSZm}!q/g)xB,tF}q k)uPXM`\&Uh傰ml\k"<,H}س?B F,5b_M\@hA|僵ŸzK5/*)DDQxC 4!9{Dj/׊rkKbixLs]]6X۠XY3(SRI5OU%,򩀐%* C-{ -ƴK_vZGZ7B= 'ӓG-Wƶ>)G@b`֭ؗPDj~')%C 7)LįM1)c$dŧ,*rkPD3  'X֍::xH'n*|:͋kmw񤳴Ƃm7Z9lb5,'f-d& V&Ђxes"KXUĽ@2% }ChBsW>Pi_q✑Ly # ߇@h@9[ mp0+ 'Ֆ{{+@hAQ?B %6uRQKZy%Ϭ9uЂxmM<GhzҺLg[% kO=2?B )jN|e4,&T L1@hA&Ҙxυ-p-ɕW-eX)&t4'^zqf΂!%S]Z1%PĈ#~Tc ( TAZ蓼ńhz!Zoom*!C Un0+ڤS=BU91jL(9W'0B l^EβW@`X !ZC!Z|^⽎B5@ 1+BHNZO-'\Ak K&*T1@hA<ԘU56 rρ"=N-7ε!{%AB`!!%K$R!Zoѵ@kuPw֥I?Yb1v, h,D"X_\>8]m[O>Ntc=dk|dg;1~h}Ad6_JYHa?n}ߡI~=%Ƞ"(Apr!RE%8dʅM#;wz=9 oTߥvylk bg=`gqgƟ}>=9.O6Wܬvg4Wk?}*W-0[lQn^EaL L.?tSw@+K>FBQA([M5b_Oj+uE#+wcTD_pQ;}OS8{vl{?(wH(`>`ވG !W',XP n:\ GmLA/mWF-܍-Q=5̲X}_s.hFis7Y~:0 k}ӯL_"v\<0JRG;&hPm.ј4iպ(sbusjVV杶ҖHNFz66~<\7P^O'6N썾<|O\'E rϹΦ(UhiD#y/KU<όΰĄX(g{ݩaIx5K\{^JrԺbն͉ۺM/^W7"v]*'D:l2 #vz5`i RKJ{k@hPgB&$R2jZͩ]ӵF(sWXsYss5)O:2?Md:N^|gsW-7rR"'럦I}f]KZ}ͿvPೇmҩvtD/M> [mz쒡¯B0]4Ȕa3Q_ʆiwjݺdg߾1mC TLQ-\k ;5\Fꍖ.?9^ߕ?flA /lq$ o9S/%rsA|+uP쭋v茷WGs|.\|Շ;fOۻp633K1PBy9!Fk%HrM\G.ArM33p$wP.VEWʵO$InTZt pg6C_DwZ}%u^V·_{zT5M5Y)TjfMTJQB$h} ]-!=$t xc Z(]HL=oGTߗ|=Oo:*@V'URɪ.B0Hp\S ƶ7BQVIXh| V ]Re^({}^!P ٕPWcFddz&WK))( N\m=k^VMi֥r*]}5S;UK6VOxuM2kR=-7 CAJ xuZE𮽷$kj&;<]Vכ}͋};7ߠPF[~u2Z H r/R]"N3옠~]Yn5û_ɟow2_|aG9_k"{/?nDR]L)EZZer$]Ygs}Gϼ{ AJs& |]z==vcHk5ɗ&}]E%tZ]~QݖdͯF 6 y;٤lR &?ʤ],|hRHMhlg:]n,8!n|_,.Qd'ǻ>z@yD"lѨ+BWs~,b+KWWKFuՕYi,-G*h~UݨޠrR*8&uUp4ꪢXa JtuU;WpdMϮՃ|fu lsa^wPW8.V<"uEG]G[a@-+RZ^klo?̹+;oXH޿*gDX,4, IcpAb-_vJ9ٟ{zS*"gEWD.[P*fc`&/_o B'>#S-uRpC :zctp7FBe=.DSZ*:$gؠE*jXPDXB_ v,pymw26q Tk93W?>;Ьٵ/zڶsX1|`pWOJpГ?/NN+=y әKF{PIk YA2^gѯ{~M&:^w%)TJhL1 (yYkc!CыӋhG$^N7¹\}Öh"QUun}mGG=%"_]9hL]'XL`R4c>[L0 lZ닏hϓ'5c  bT`(ǔDd02gR*cr{ȍ[Hx4^[K*!%,ErIvW Eq 4.[%!Lӏ_7V|Ԯ=Xk@* v,Z]rT>zg"FL2`>B XH^-]D$L9x>ȼ Hւ(XX.]T&8LͷP~,h\'ˮ6Ov|#ev3T5=S>TnjDkw$Η$ RfiRGIZ`)";Mx-&Dhm :gK\بU1"e΢$\g9FF񻳕÷hb%2zڵ~E;R ܊{~8^HXA$t77 V.H?E, aǥ24(;;DGSSsMWG|6Zme紤 ,-F=%.O-@٪;=29۱g艆~#c\)׆ػ~Ll=ϵ{>:y$)Iu ,>z O& ȋ +A2B;59ncgp^o%za2`n0$8D!%0g܊) 0qW뀉! ?9Xy]r["}*[b#g;z6#ڐ VwN_9Gg9JCn?@8@z~8vHHIʣJY&xݟ&X3HڛIJO#A!CVQOpPp'tֹs* {+\tZ`\ .sJqUF+["hD)Q$ wH{Ds")^X"#D7 g(ZZӦ_o#)oNz.(&<t=01q*#LP}ha=rPb0C߼D$ :bHQn5hw;G~[P2o}:[ (vh_T @A} &'Wd> >7:ҢZ,IOnK)Y@ 4i@8ݥC XJCK:6* tTP c+- qұA AHVV@X @ 4b﵌EOLF3`b䴼5ox8܏G&K'~{I*88.ZC'߷wC~: U[ߟ3ng՜vߵ&r#ied y<'5Ld\9K Fit<Go>mpe҇1ÃݼՄi$J#Qͭݴf~}BEFn|/yŷknaOij|NjQb k. dE<7C׸V7-%ϭ͏3QD lGerոXfSٳ/)IMf77B@r3CELM>ڽV^v˝@']v=^ϠoyŸQ~UXkqzLf~z.mrZoU`tY'7ܜ ='>'ih/%<i 1l;Y;Ẃ <]C,Ziӡfr~E 3KO;<w'k$N0'+u' ruud╅103^~jX*av)FԋLjiܷ \d'-L׃OS:-wJZn%iVG.moJKi:X>ޒttTtod&N]d6lP)1}o1IȜܷL]Lb) u&:׮޹ ͆MӚ.rf(jڝ8Nx N F߁\>?~#a c(.6 #c_DžfDT@8XhE#[=?Usdr%&02c[WUxIRϒnU9M'N.Њ=Onf`ў϶@}X ),Te A0OAsߥw jAj:R9)qB" 4PF)VJ;V%86R=a~P+Ïo_KM\6EQ1uX*F $hmwIYNxoNIvJ*cYzw 7,yb%\o;8 `,49BĹl2h)'ZiHK)E ᝟6JAy) }Ulz ,+{ZyϺ0:8 TGT(%yG3 xÑ6#.JB0Tn]}ug>.1poBXlH"=0ז]`Za¦-5TfA30ޓ샕ޭuu.\y9V GV+9g肳F(aLB8)3FBR)@)ĕgXjD$.h$Cq$72W roʕg^"'Ulr'uO![욭O$[xre_^*مh-Vm;6Ks]1+@c,hG J ZA&yf-msUoO|[[0L$*~x/dտ,>^/ryޜmӴ.{e6io2Ge9XʵwjW?wKӃ0$\U?=~2WOoN<:\-skݲ,5R ~o7Cc&;O4뽟~MFgǞ m,bLlrM| 7)ӟB˓'=&KT֌x /@ "pX#˭FL -OK ٚHq p'q7eџk_H^Of._̸V(`2͐3n4y{o^|qZm7Ͱ׈ԯ)u [C65sQ3CƢ٪Nn販>L0]ϴD)}f#Zb]7v{YJm/ecb`b.ߤ)iɍ^R?*4%W\Łӵ,)_:'fz7{,%vtlIT[z) m~,hLwxdTUnwݼobI:kI_p  6 7lX:7j1]bSTs<:$T/|0m(ceuךG%V@XZp[NJtj{yo6JBHy)<8kE39‘[&)'6 qFgG"r١ K|=ħL]{L|Dgw+x~ݟGq/fm}4gARa`07 E $pZ2[d -srɩK`e=c-6C >t:xE \JHC*hòtq6>rzV.zKJ Q1/CIc1 +9KIžʂqWN#k .K/깴N69;_楽t^9~;ӪU:@MmdZtIFA <$0}Ic$" 4&aY]={6rU'sjWdUZz2O[im˝J[%u4+hVl(M?ڸnد.jt/@ִPϦff6c7i<[wjZ˗1pӻbl@u0vL6nqh8s<^|t4M}sv'{@V40Zo;s$}nIVhsoh>ѡAAY`1*$>`D)2 :ԵloMa۝BPXGxGLp ~'T` 0*eQTMO*G1)O냉K̀`-b(ʘuaX'ָ|)]UWI&ŢU=ydo2mkC=>ԸH+Ml'ZKg#5t&g_HAAJ}; V;`RDLwemJUKc8Wg<\OaU%ER?y(ZlJ j3'&!A?t` VYEЄ 1QAsgbrT`r'IJ6585]uz)Pj̆PV 5{lɭ^岫QoSSuղL4SVg5?|7<59{׸)fXWσWT Y7'O`o)'Dz')8k2CfPviw)BbUKKձ8+ YYǾʟQl7s "DyHS! $ъ3_Ϡ}li}Ѹ_?+`+z9VN7ԿͧqZƍxĽl6?rF3_>nt%]rAԇ|Ro}> V>B:i{]]>YOd u vai$Zɻyj7M%sX|y.NVfk-zS"HL8A=jJ nhK%wF8ZBo_d_Ag==-,s=Ӈ]GZqwkDٛ~AM^:3yINȟqU%=滋hfU̪S=˺jh^ݟYRX)6o1^%C[ji<\ Yh-en"h9uUn^Wb[ G Y|!fG] ,TjnܰGodچԲ֔ԋփ3Yj3ZVb5MZf`OOݺϝPikB@'}}}On፧Oњ@2ᔷ㤀oTdg b1DN yq!8P!)VZY~5f_ZOꑦ}dU!\V-IGT`d"& %v[ʫv}Y`n, p 3OG/|kS,B0})fRFt/2&lt҈T%ր=u! \HN) 2) r"I c}Ƀ4qU:vC:XHH5Yq($a^YNAAɂr.Y IQ렄uUM6]  ^JtWřD㯺d15wjKG[EvFI< _ P 8VHy^AUmX;'RY&\wp)3`4 j>$p+И2E_I_$vθvޓNP['IV?d_>ڹ)mSj{ 1,pcҕYN2϶鹄9?u^rN Db K[}S勑#O/+hQp.PƆfU:`|zTKpQvPƸqLq(w;bVJH/*dtʉbwB{B|9o^|6K7 Vs&LrVHqihӬluotD ,0i0Ib * g.b=/%Ky'k)!|DN ' 1`>_k*eQ8X3JշXprKr'߳Z>(i+>xZA#=Z:6 ʶGIZ`)";Mx-&Dhm ;LlHlT R3Iæn@2v07V?]7{lO+4Bl6ŊGmWTL%O| ?Y7/$XqzHei^+BB"FRpWzc/z:%=esd2 xЊ>kɟq%6,o0LMƂ$Cp(\L[2`&*9oa`\䒇I 8NlZ#8@?39n IPL_Q,WS/$+ jvBz!߱^H~7lhYA>KP( xA iezx-W jq:2aJBD WL7sX)r XTآjOTNh%ɤ̧$#*h2 poId#48xba8=,պ|yƳnca:ޱ>pJy\;Mb by}.Z94k9gpc-?ԙI> دB3?LԿ煉PWE?Ra ?oƞLR_֦L؜QQş:i4Z amӯnԄ+Xq1U8k/Y}; fGYF (WL7; 1N}"#W,.ɤ}b}a*%(=C ]]:7W'M ۝HaUb}uLZ$wϭ~++Vgy1tʸO<8h`fQ8`0THI4C Dg=,c2W:Kkl)J"(+T ;͈RϸsXј`5Y,·*x]W8m%όmH.ܿr%6Bm, w{}RIŌyϩ5JhpqNG:}ms8~<O0DMn=7=oe:y3Mt{8s9߫zMIM3pbQ7#Ss=uIyXش9%cBh\fKPVp\L@ R 梍Zߔ/RJ|$Sn4i ȿ7FC^5jbP{ .>壿XoZA7_2le(D},;d6+O:mUuð|4jrovSٿmz-UDZ˪TH!"{!bO|f=h/%<i 1ЛWWnOO~/cG!wQg,Vjq-Qf_p`PL^ҧ9;w\Ug`tps9C\tvRGq )C\Fߪ7҉t r^|b\j,PucV-uP$1f:,)jg5'Dǜ=p%X'ޢ*C®|2}>{2g^2XW~=7q{(ˋmP9Q[u2ejiS[ԵSY+#o#+Ĕ6J/J u%IM"c[Up]bF.KBBr~ꃋ_Q?9gT⚦o pLZt|SooM/y1p^2eE@Y?;Oҥ':{.C cp!::>*ͭEtHyz0:R/s1 Vр7'L/@eA87)W%+ߎbyn.\łgA-/4Mh!V'ύU[#BOpTتj8!ɐ[\HAۑ&P^q/8 A>䮪fhr{V:jV\I*"1gp5+p4^3f){хfsu uӅ[OݿM)[K°&'71>׃\й`mp>>]CJk`{Q@&h6LC0̂>bQ;`M'q@^R )&H&@ Ǭ" A%\ =Xlk;!YiDfm8QGN0"tC>@ -hHkHkH&Ҏa[EGpdVEE Yj1{[ )]wF+BI:R՝u1w%9i!1DRY֦v-eDž?ёeƴĴ8-4xܷ Tq:+B0ܢQ7:c8i/\Ӿd_ɔ ұ0Nf@oS NJ rZc4R2AV[ZyRIT(.,qjfzj7u;!W6i S\!BA[\8 rnrufX: ۮwFÓٝ\=Fήo1sz57.dÅP2űX75g?twxd.껶Ǔpyj&|na6Gzϭ{%JY?)xKjD}Jt9{ؠ% ogK2j *CdwJi,\X-HsL*/]x_{"-r(ͻ餽x~?z鱄h{C̜.kiim沵rO͗ԫz<2\!d6`6` uC\3(v voۊ =|4-wHZna i,^IXUdYnhD;&b`4 ?oY^,Тnlؖt6F_{fN/ܲHYՒ(mۛ5qiTx2M< fxܩKu;iki$.js<PNEێ[i+agLEYqTcL%,tfhr{0Hʒ R=)`@eb^D"  &r.#5c5qk,e/x.ԕunp9ټ7ekI|Mz0^> :7 'k(S2|Sq\Lz/ hc ^C4FchY0X'[Xl9jSD56.xT"K !0ĂI`tUdh1QCF@pIwSJ"U`MG-&fSI-IZ%E-tH+h ]H+|["_tq#بϞh [+gZw hxtbx}zT m'ulLB2M~ЍMֶzZJ8͑`De,G;\ *C <}!eR>&e4VV "9c$a:[Լ!VgÜw ~"[dWY&%^Z6Bwbn?O٨_\}qn=ԙȭ5K N3 FsE2og A+ `:Yε$a*8t97FtjG<p<8sG#*2qbܯh%P^eRur~/M*\zao6=>\Az¸U4 Q+sz;/M,q/ ؋9Aӻ!q7k 3{v~ig:I&!Ocm9m Sa/OJEhy]uGYjz~f~u(~z-﷗ו f"F1%\DIڎV9VMI+4^IA)jU3a&3izUMG?M5~nu=75s?mgG)\2wX`a0Dփ3`l/xێG^-NuzKmMGᰀ< lm[`\ՠY]w(wI(tvgOFk>=5۴'d93:;LЁr>6Hr'v9ǜN YTh5"-a4hgVƩ'18]f̖jbgST8\hW\źӖ*m8*D v2ײ̭}h}dmxre,j68x8insfG̪+ab5hd*ZA,2Z3觫 by]aӇ_7y$!qZ`|ȟwO?nӜU=B BǧOI^ vG|6=:1aXUSBdڽ,B{;0mPߞ?ۃ~ms;Pcnxlu&}"?m@ tNl:^t $5:@Ngv.Y bdy] ;eete&nr9) t9R}NdHQ$KN{yHAa㷄9rlݨ*_3&/fֳ'w9WI;`!L>uN_R9uZA-G7L*9.ϡw63gg~"j}捑azw|O'Je8qn3p̲^:_Cu-%-ɪiĔYEC+tf6gH-O P><}ghb@g:+]wA+ߡL"zyAv%BFc^Ja5J+%E̐$1qZs~ش]^:=ϝYoߛ׿L۬}k)dsBPERHl9'fE%ٮ~}nTOKR`pZzoLMmΐ563 B7ޏ[tIb852UZΑ.ZN7C{$|-u)Obav^ ndHRt=ѳ!GJ?WJJu&J:yE rY`jVqUZk.Ar$~Z8+OkwL[! FڌM3>>[ դǏ( #ceǹ|q|_q3W]},pUKOnUjYK Ю 9+:4Q.l0.Rz"J5LQISZk&^ZNB-E*+5}Ӫ@) g}W9 Y }yoZsdb4O~moR{ vťvvU:m~sT+̿n'_+sCݣQh :Ldp  "lkIS avRs^Lך 2-K"LD٨J7Yjzge񳫵۾rwL0!#q6Gmbw26]4PtDk]PMOZ_}]w[9] _[nK)_Gxꝉc={HPϫ:?ya㉹iaWm:"t*`轚E{7wk!lHڈ$l $M(B, *2nZr +&О%A{[{݌5&l{Zxo. ~*jov9qi{EOzʟVq\KS휐˟V)ʟ~FH-Zs\\UqjJBJp5 LઊԵUVӥUhW$%"bvUUZJJ 0++J"kw5pUҞJpGᚂ*0+>j3R\*:}wxsml͌bvUY/ӏx\DIK[J_&49֭~ĽQ qXml~7Lk!`zg pyq{ąVN[k9.m})V['VK?|lZ'qvebO}EyR^1% G)O̶~ŴXe*_n ى>gOxM:c&FDտ_>\ lk-YTF^b^dj3+ìe4c4_>/[Ս&)\ U2 d yU"k`F4Nqzm TZ$Wk,}Z$4 e!BR1iUka$Hkoli`dN5$ASE$VQw+']"be!F/"H @NOjZe[JۢNiyjvNK9~0ӥR)(TR$ULṲL&YB&Y4O]:$BB̏()f1ύQ¢D"[Wb^ %z "h g}=S-D% I;2lBZfmRC(d#`?@hm4[>bDW!3(+asa2D80a8}_~c޾ea(U4D`ky`t!5MV1E +V% J>\8'2e-vVDdJhɼACζot%(KZǑQLKag] J$ XC\۠ɘ2I@r"AJ/F y+#!(QfA)mQ_4%`^3b E}V0u h:RȢ؅(giIR]Rc 2h(2rOXk0R(oLKR̢\fF'3s "̵N-svs?Fe& BNYN%_2he,1yVm"eC\P֦hТMNa!vJmuE]y'-՚XLBgVWŚ*g#A@d@9 Ä.b>. VGJ $ˮ> eсb`&1rfٙ +eiu[FK̄v9m$ 2fj89ÀuXq0[c90-m ol#xvu?KHbtq$Ŋ̀9J`(5|*dփL_gJ{:_gϪ}10t|#SBR^>QIOd]dIr:LMVJ (ȽUv i*3jƀ(.`ʈq"*X5J ƸE]% NK\6"IX^,.zVՊ!1*` x9TX]R5("Xkp`c#. 1pi/s 7Db6y ܗѵX#De[5u0IaBA*(S`b @pȃ D:\'iT؀b-JmF΃7,5ՉԢ pd*j]~amTgP'0,Z(&Ȁ rw72Zdp>="nP\- c:¡gQ! ҉TI.{B:b:K4YdJ ׀, yUuzˬE,ʶWX@KMw afs'D|+䤓F !C:kh:/:Hc$hY-ԵnB$0z`2o6:W W(;5,V{y. j0 3|r7e7xDM !u>fDsC < В>=H(R(."PPHyDBJ$5?N( z=jd}F M e῅+Ap +Eӛ)طyM',+F.TuB%BJUEGUTk CGz%Ϡ5(x L.1#dІE:zV19EKkLT?R;9֠Qrq[Q˃>d UMIM}ܵ5#[rmnO uMg TrGYE;(F PX?̌pHᖕH-t0EWQX)#<%Cq<4l9mِCL*+:_EA] 12 7[U/Nh^'CB% "!JWRv떚Q{`vHTv#DoaL10@5aWG1'(nN[W l[)ƍ4(9RE#/u~z O@F(&TnrNI\tzydv/­{q_幇yB<*5OTu 󀪔J9b!D< x@"D< x@"D< x@"D< x@"D< x@"D< 7_V3 {:/>Via9 !F< x@"D< x@"D< x@"D< x@"D< x@"D< x@"s!D:x@K7f@ z< t"D< x@"D< x@"D< x@"D< x@"D< x@"D< =W|0Oz:<Uxzqr>%~4[ps>X3RϽ/|F]h\%ٓy8dE`@b_bēx2nXA}1~3Yi%-n`ƽ9>we)/j9#w%ANdjz7'O2KHW1!Fp<ۥoo=: g+?6xk0Vb{7j (A{Z-SFZpE^b7ූoY ]# K6箳+lᎯ]r+d*l4i]R0^qiﶣhs+\;w}!s! w `dv5Xꓫ$w^d|pkz~28^ dɦjKxa{Ymv߽x~u >b5}c[0_?gݦav ,|:| VMGhF#9=6 Zaћx|uV{2^Nbvt1>ob~W&ս5ZUɦuϢvnVzV蹁8۪@E~ldTj i2jhfEϷkGg0wgzvz#:loy>keftw==_t}bztBm^U$2o}ZR RnѤ=Rky TX¢~E _rikq /C7۷5^t5!Uϝx~ic _4['ɼڎQWŔ&QTwj}ku#Zl#Zp}rZkbOus`fź*)2+gz_M杻2|]b~p6T]ߴl2Gi7C\vS<uwN u?́һ#5okw{L -KT3ɇ\9Ivhd;, ZfMcWlXd_(. Cd7tj`QcaCB* G@ 9PU^5>XQ#/ܵ8yN-՟>n~29= []m=Q{Zg+3W Ya ў :HG7M>|&5<# ubtBk5s[3P53ށ 'F]dGL}L;m b[)I/oyrh DSNcu<,q 8BcuWэ+i]Wi$8Il8CAp60.9Å& =Y- |O-&fmǒC_6>ڠj.NG(\Ol̺J6mW]J041zg ƻBov݉tWplm7zg $e\AWh;-7l[IftRSRWYLg/;S:Xɍl^dERA/Uz,gzȑ_rÄ-@p6 gfOkYH2~ݒ%+%[&bWWXePp@W_A_4 ~:T( (eDXcb~z[i}emaSmGS.+ǩ5cߌmx0q lw/rܻ.öjcpF Օվp;=3) `$nSbSqc)6(0eIj-hW)YJ6S28 3k {JSk{T#gg˺:'GFo|/>/qE{L—Ų&caU:wߗ|?h^/$hxy3/7(;W76P\Me;'W%$FQGߟz{n]F%_W^/-&O%`g;v;<xb?A^hNBPp5Jx!76maA^xoyP si|׃woA;hO7h 4hm;(ض9UYQ^j|NH-N(' 9Ċdd.\0-Kmzt(P3#5AX"RTD oSƽT ²fz*Z+`d@(s .K3T-E`%$\'@PjDfШ*')& 3³"j++|UOΈLDbq.l;XFΞ9'[ͧ! WT4n^̓m7l{]HmW]1<ȬJ bHЊ ϓ֦Ei6!YYVQ2.d;eLYKIVg\b>MNi͹ J8dgǓx7$9\{}zyvM;7:Eb^ $_1,hX!}h/x!JL6#ʺ{T{#İ[Q'YxZܓD,kr9%0󋼉G[wG1O1Aֽ"Pg R~(-  Od\`d҂}N-8|"{YENT:΃fQB R3\RZzkBru+uUX_=C[İa|dQXޝ|"cN4FX@p9M>gq/&5|88$$$!ָ@~=OS39׌W!i#(*qUfArBDR( IV#YiFȕ!b5r\~pv}E+sHqR\;Ts`pL?;A}<[X }P`XvdmWi^'THerpdiH3qi`O~`N3)U'1L7W3䁛oK8ԢlgWd߰7'W 2Ǿf3]GN}hxU~xUQrt]O02R˕&VW:R젷PTԷPhC@[RVDu;`5VkŢ>+*uU-:%ja6 2Jhuf"GwV8+Ҟj쩘 >~k{V[n, xz u2K"6-Yܸw]z2ӗH `nKؚ`KkҶ=hӻ6۵>Mؒ`K2_wmxyxw9{rz([nF\>k݋{; m7yo6762o_fE|Y98g gu߼|Sf)pACnCZbqfiOͱ-12g3IW5C7ի3VGfP_ӆ%Uc)[.iK /n֐q獾o}>z)w~ ;/fEwn!XgY$8ӎi:cĽYf-hoXoX[k ;&ip0@{3y{ͽR2{q>GN]Y"j}=}}G!s=ڈš;tpRl;]e/qAyǩdF*C2,D/Hec Ҝ&kxVs<'&g?DŽQOۛ6Y\_ܢFTh4'8㢝^ )uٷ"l:{Yu[B㗮׽\>H`Ԏ/rTjj{B}vV0Yx>pG ӻ ,"+H3)Zj^ ͔!<g.fR1pAWDW9tD[+rHhJz Q$U٤{nhL臔x" */l)> YE fخ(cj_ 7Ri۹j48ڠN Źe o+mn}(粲ޫf+'R A*k:Ώb˻ac\l|2g" " D(YktR!.;(O/Gta/IkAk_Dtйsa >~EF߇e5R T)x4 GS.E\8V`L2P)V-Vy) If#o_19X@o4uR`赦&Ԯ4Z,>LzsV['3V?Y/g|RU;W|g`[E9\ CI d3Ý528M]ɨ1f祝ȃ[Y0'RNl=Ȉ&gp)dV&咆dFnɸDV[,T,@S ƶ'}TsZigiO:`UO;ȑLjalg.&ՃW/ d)XNC)e8;JVޟG9DoVd PudY7fm1Fdbㄊvq95F1.*('Ya{'*=qcؠjʯSgx&_mThRT*s#{-u Q3o杣s]Ti [iO /߻Txefɫ˛YYlS8.iB`8'*; UlPX؇h($dOL۔Xt>AL6E)íNR~~BJƅ.RYJ6S2e8 3k {JSkU#gٺ:GFxl|/>O}#cI`{s$uc\&`Ɣ`yn&vP5*+Ǐ{,NWWq15 {bx{4 yȻc&W;#sjIʬr1j.BLu^~lM̥w^rqXۅf3s6I<~Jk.%>^]I#o<KWJؘ(\h2FP=H3M_}d%ETȀ(s .K3T-E,$\'@PjDfШ')& 3³De>*'gDb&"8K=?T#g4=AHs𡙅W+k^㘷4BEfK9ZM܃ȬJܚޕ#"˶16`mc`wEguKr`JtU: lg2IȥIkS,  x x*{d ۵$dNWx&Z JG*V;ӯ8WR83pk)C6t vEw8Α pԲeGoBusIqi?1*iBf5,(m ޕ[j=*#QHUY Z ,禎hʲ_o}.`~pTM6j[GLA[N{7yxQC((u>+SΘ'4H\tܺK#Q.:1崵΃QA Z DH R{dq6;% `M-qdؠP"6!%~GCٸ9HHi%8ib4: m!E')4ָx#" O{LJ!'A>9@h11`1T.XBD ܃RDҥωtiI'TYd&Z%X(seXM-(_3_ ghn<5owQś1MhU%gʇp.C,h*:_}^Y ԧ)MInJ v?R{"'"kudwӨ.Vw~bwުiSPNj6vp\,{F3])Gxz=VB ZX:'aoM&fqGz ov7dikﹾ+o?eCu7.sPt.NJe^V!kj¤OlB#+ԧowo&<q%ZvX@k~5vYݩH@Y# 8u17N;):>#KbL*[Ce2SmTdA9pYEe!qJ"I;VpYEz}wR :WE`"q%]"4WEJ{᪴/G!UטID5bj?g a<da.M0 'd~l_S-] V@,&HDž@MF(T&C:zZyPHKPXalsht6u ={Ƅ"8 VtRpP3J=ߢAz&̀m0.T[[v Vۺ/OW6`J{cHp锆{7`6VU7jQCKtse) Ȳ2GO $RJα]Vk^洒x6퐓 kQ*&hrΠNӪ/cZu1rN8M3M1Ss9 Vˑ>:ݦv(گ 3~E\%W5ԣz>IN]>8{Olo:\ ~!s־'e/o~<uxݝ㟞"kC dfe ]k9IjL&Fv=9Ң6$UBDV-&'$3-GךPv-q6/t~۶U-4ݖ}ݖV.eѓ#a#zCLxJ\qP9v|=5X%SZ4VN:9n-ʯMlT{E]iS7FFBl6rLe @ ڰ!{pƃ"WEF4:ǘg5ʒ JkW"B49Ӵ 9b \&ٌYQf< 2b{;Ɵ6vuƗՓ?yC./o>_^;sMeTQk> ¹d8w^jX^VElid&HsQ0(D&)MIH]z\A746?1o\gı$~4g ׁbj$tn%65V^-(.RۮΫ!U$J&a{F;L3ȘFCKSvcmG¨sVx[_&K 1fpNۼaC! 2՛47o.mƑe]`Z}q`w^4,ӫ<؞4CikČRL̙U`|H¦ģ c)p4ꤴ/BJ傴.E:kk3p̐J{^Bf2RMͧ%QZ<3CsfǛ]ӌh?t,6|p]]f`Ko=9rKyhmE@<4>T;> G\@[=-}p}:o+pSEAF 0!)i_4"5c*EYv0*41aDKU*\(w 32saXUi,N`5Y7Dhc*am?{ Wb`40y\Co58ҵy#V/:8d4rE[A* ?*{d9OKIVg\}6AZP:ZT9 ټoonyEήG:?L3m'li-)&{? 5b\$5\F{?LH4B8kLaAi{뱷{DGeD0: Jc>{1CO ,Ǖh3jJCojCO?@?I'16GO1D1=n1w½ЊVU+ЊNZ` ;&Q /Ƙ4= Xݵyn]~*_tw Y'/J ?^9+޷-LW%O א%A"%IHBK iϵ HpD -_,>sꛕ>  ,̪:Te)ڎ#yfBP|+}R; =wmmHytx)ٞy[`vwbVdGr:oQ[Kʒlێ-Q49U@!f&;-U}?RK+`"tIɢvYt!`'CJ" u&Z2*#^J"d!Q?Q !YNTM0JYauV?_]8 \Y|7 l~a2gU5-9POHrcGGGBG͜046#\P)U<>3@~i^%A;b=^,pX头%hPָ4"2VQe)D UHbdV>Ri0&E3m0Jdui  g8˯~YY/vQ<ϝ4T)h2E|9[0gKq*<MN.uBAxHHHKywʃrWB@%aP&2!(`<8ѩ:Ԇ #k,7*"5Q2(U>SA1vAK{YIX0t A3lMQ^e9E:&RRpE$A#<뢲7 gπE=Wp\<| * ӑafgm\{eKqz uZ:i݉n]^'Qoskaz{|;޵پ9<Yrbƻ;_yP3mo]kg_IfwtHs-|* +W^驹eֶ,lۼqҏC^"}UֹQ;䬳gtH7 e <V D10V#3vFCw19YQȍ_FÇgfE({e/\U&wEӊ:А]ߝwyr0./)Y˿O;[Z;tzS4_I) 䡁goO)0cCnCQjZbvݬ87A~-!w|2av+>-:ϥ]R^讎rw/tf ݁7Rˮ ʠ :iePfkcgj̑12q ,E ʈTJE$+298)3iM7E8pIH#h&\i^!o !Q(u!H٠#Y##lDuáCaW sdQB!!uZԚsBBLD3 : HũT˧2>(T(kt\dzI]O׶N$:^F|j 6 ٨B _oetVVh9 the]g7׋ǛֵP A˛=siVg\%q;o?ץ\^,ֲ:-LcQ9/=3$ gG.?#V?wu|1p'5Ol}A\(bZ0!`*CTX)O"T"$.}ɉt(v _DΡVQ{/%  u̎mCHVoS0LRV}tc=#_Յ2J3\t2/n]_E<?]\o gM~l6_o,uTi}]k߸}n>FCeܹ_+g6'+ټvvrvb#E9J T0(up2xp@%K/;v<K| X(>:RVL[I,|pʤM,L.gٚːs uL1 *"%I,FyEe7ΞEA:s|FvmC4TZDǛo[Sp`>;qz EѓtgуAn<"n v^O< R1e$Rm=[7C/4j|ޮv>qXe+-χ[[=k|G JсQW]Ssy˦;_{o@^M93>G\xW[GZ;CqY&Yx7ITӠP?cT\0\}!8x$:gŃ$Vg yԎVͱIŪfV6O|*AY)6@)̑( '_~96 4S2^X_JIH,ճɋ?a+Bvr(yrD) jwTZ)ὕteKe Irvb*9??R*xK#üzOxmK(J8|Nlѩ"w)Yga1x xkW3\M^No;c'Pyv~s$~cwBLlͨLHq(*HCp^AWYy 0;_J 骔J٩Jy(2)} eR!Ց U҉3*ʟM:.ݨ\ĩk k~嚍SBÁtG(7_?1j@d ЊXwR.ѺA>iꞙ1gۏWf كtF6۫"}{r<60тY0qv6dfc,+QdҘ҈6N9 'aH(ʶ36JFQeSHU12*cfqJ=HM O# g ^B䡒2f6'͘iN$cfs!r;!K*tW[1) tBX+Mz&Wjl' @àuPJ; "=TʟoN">xCL0$&F^XrfॵթA ᙠ(/kQiU|0:A<SC( rfE^p2K2G1Ȏ JUvgH5d[ M>b9HiC@Y ǚeDIQ:!hl8-I 츦/N;[۷ʧ9ŜanYYZeh,x~@c]jGRH S~1DqdGŸb=vxxE%M͜F2K yoS3(hR HfIIc+.(I eSQ1:R:Kr͆{<+pmK4V&Q/?yZϓ<b b*oZO9/3-Ku0|HWCjx5’!K(BEa(Sd]t'&kzI rl3iC tgZ۸җk32WíJV%/I{k 9cD")MǶi5fF@1gz4‚ӔyƧWiS֭-^@*aF],s}_] dsI7z¼z'zTq*$q۸0=9EJnӢ Jтd k,)F!tuʫ." RS! HnȈbN0\~3O+!Q=uU9incҹ7}yD Af:&Gz{0m!}I٩3^䢫OLɫ~+[ \w6Hr=+݇$OWC r0XwΧ_w~FQ8Jv~3u<}ǩ /aiXᥚ- ۊmJt,Հ`.%r֣S[zxqYՕ9|\ %^JʳWW@%m RWk$ӧGWO"F]=\q❰Q{lnؤDgp]z,u};sD-箮+0Rj=zݮ+H3o A#.5І"@*kWߍwG:ʹ<}w T6XӯʚPL,LBc$Njk;͚\#Ȼ`IbחVNZ.B,]ȍ|Ђ z]۵ 1iտx*q},Bv jZ/X~|V-|Jagם|v8|ҐE-t'6?n׺̕\91¸R?VO=^ rsrpݨɗ6cmG`ۜ@~My&VkH[d7//M׿+a_M'bJL5Nl|sR M/Wfͻu5Nq].%B>>@35 *;v:>D43(NՁ#ѵ܁[\SC-$zQ ~R沆;AhW>{ ƋVz!}­M+pm{073 ttvCց ǚ6QyQO7r[ɰ?+!K@JEiEB&u",o7$9t5\xv_e=|/VDjE!Z;~Uvt;Te[/W(4>ύO.28fk6<)?Eeq|Ys 0Rh o׏yVt*XQ+l1u0_@b ʓA:, qY,9*8REQE`ZFF\X9Mi+:U ʟ!SգP1i~w IFl4V++ؘ^ݙ:.w{U@п~px4RgCǫ`(R'NQ)B/s0']ò ܇9+ @҆mBsJ!N [y$kpC=tKhG='IRa0?c`d͏p{#WˉfIh ~yE,p~dƷlǘ -ahܙeDITO%6, U`#)E#_ϕo|JHu  L⋉LR|JֆxNurf.0'C&aGmmN}V}YZ>|r9c2vV`W{mY:q5_uͷM˱?'לe]]_oCCui4)6tz;wn4h / Q /X0DBs$֊S/Z8u"H-t%xiscS#ȔVNASlR=R`Oq掗!D!e!x0k5f,`ZFL&Z'%"sl4?/o2Vn讇J}8==t -=j6^zޠɆZMƓpWA ?쎮gVy׃fmΞ5&r+iT}j+&]^?5\Los!gh]@ru1m{xG6۳vlBZ3nm}g zxscn~K9|GdzG62iY j]ЫR6ϙ__yl?C vHzsk}有#*k4$0(@Yu"c`XR8g5`Uq0J9Y@;ȱv.iPY нԀ 2N g)#7d@I/HUQծ4JEۢUIn;jZҼ_tr|%2HxJ(VQ<ַgi5Ky,rɩC" y5F ǀ G㼐`ZHb,VIbc º -6lS8$0'ܞ]& iNd&C'5z<ŝp5B*z&щ5gd!rN s-- iZ5K>qu9FE8^46~닯d|dL$ d=hGHrGiQ~&jQIW&Q$qyg7Kl2Ng5Vx258ȁZ|/\eJcvón7/M HEJgR >0%WGͥ;Admqb8>sM53me8x7.q kǩ4J1AFJ rNA"$A V$VxS@1Ǻ*K9\к .;/m?iZ6dU{6/.<{]u7Vt>׉s)3g[ZkZ܏w  2Rb 𧠹owخYJYK=Q fAG4B*'3b1RNPAPBRނ,t(2/"H>2s弒L`h|HkL>ϸ(>**IUhA}afAC: Z oᝡs]3Mye?So|7Z3]J%߶ͩ "Zs/2ȕR $j*"mWXd!>A5O۠Zd-2 2Vh+ (Z_-!J`J[4 1 y(jRuP3u!KJ]XA^_*S!W`zijojf.ފ JΩ"ଵ+GMB8 5F΄$R0<yCR"jx:$' q 7ProDe>Fn 7i# O g.J OrQ jݡ3fQ<jc1J`rl V(ˤ$2&*fJFHqAH1rCxD#3 V̙laeO"WbuQՈZnJH &Z u@,7pPJ 1;^ :⣔Z #! 紩V/nB/>l;vDZs*{q gk!  ͤw۸%} Xƴ/9'Diѥ ShsI694fy9Kѣb^Cl/ rcMWWofV2ozc}H{g ip|vTh#a dcSF ZDFXŀ棗ouh${;A]’Z[2քْ #ּqfYF8kYyX׀ќޖI/AV"D 4H[ll#0"}DHZ7g2|#~cͽyl-0W+*N.֞Zˢ"y c,F9 'c4#^Re47M\h&6F NpgՎseLۼC놦`Hh1r|3u%iRNa?י ̪]_q$y/-,͒NTDm2hx#KRP.sWRL8mxKN2LY{@gx?(}r*4`z,p3%#SJAi4Ro@x %"Z(@>\tvmE/^Usqa+E2%ٗϫ4hqC]R: v{YN\Cs@}(ni|wQߞ݌k}uW61zN/u;grs=3"dޏۍ@Ȃ7Ѻoq-~ _3,PC+3g1^dwF N-#u2Eo]4VCA_/00ĉ d>?r+#~DiѯO n1| =T"O՗ ߹8Cv׷N޼zݷ'O}2}W'wpF₋Hlm~Yu֫G4j|\fQ'|&ཷG-2h;~VOϓ':]hw20 b~^P ?~DEF)UFC iOJgz^\>4O0`4!3_6;PZ|뫴gzsk\tqK V")e{);.4fո35p寿`DJι+TؐۦZcejÛ77˛ ېNtnk-i.[gxgZ强*K{,2e[[͹''GQ"=qDU%,xp^Jg+IŸNzu|3w+寮r".j{ \@tr@'8"(6*GR3Pr\8scc I L<ψO9>mnD!ex9Dμ=\)+lJ 5{Ѓ7+ݛ`ےp.&+8:ɓyR?(~?y1>.toyy]&~#»6q#97ĵExףR;1I֣fqhlyI+wPfib2 g_YC|3]Au [QkwI֛k񽵵|39kkhN=jNsY`|l~d=9hjCŠpSr&e 2҄HJƸR1^ ھ̞l'h2-GLKR6 f%DzťAk  ,: 5C\'umB-AL ͇P ƗBR++"@*Ո>EmK'AQqZN׾8oKMp04>QQB4ih zv 's()'sAn -^ZdveTdLjG, 㩎F+iF4$O CY"z isg3W=o?ƺ0,ƙO:‘w\& ū[«Brq^9YV7|P^tm/ LPRraiF0 M%I᯸;hZ  ;~'5 ҌmȮDK@6WA蓳$!*@f!8eHLTr!FĦ sߢZ-ub`F΋* y4\VKʧGm\;!Q7q4V"l5;q0Iu5s">/ YM٩3p|Tq5UW+VsXr=D|Vn4LUIwƇ3^-*.+o}U%0e@_:./8,ԪcF7jFn ۜ5%jcI;nw-g˕~Q<%)?CpiaLZ&א}jH-cL%-V{6X-0#7-Gض'qJ]u8j7f\G1={bZu)eL=RWH0gdoU&}QWڙX;2 TW )+$XQ7*a_UV]WWJբ/R])-D^w{mQ|7xq`]xaJrZ VZ 1ԷYމˊG/+)eՄZ2,w)!EEf)9̉6A 2+~>ڹb JaaGEEF)c!++ R FtVpy[q9& .Vzy$IIpT Ɛ]8/mV8nqiwB9Rzy|sw#*yoZrE Nw-y}ޔorOsr[pMSRU\$1bQppQ&Zf}`:{*(ƈֵښRYE:H%dz0*&eD"E97\r>m2*la-ԕmm^mRroۂ;ʿ_t5^$??4~0}\Z-v24>k ТikYXޣ#0$`i2n)6ie&$ Aː&j.L܎Y8WU&lkw%fFbծ6;vڦ6/3ץ]JŨ VH% ϥp+VʺU`bȄ 9 ٚ,l]..vqmcHa@0I" `3>ʙHc)xR>jH|/{0aˤEw)^kFoIlpCYEE\VяV)> NRJ.!֥ӑ\>%}k/(Qڇ]$lue/N{^o丼]>U|&[J]YsyZz}yQׯWc'gC=7=2 eN"RN@0)Qࢷ<~ޒt>܄-/Kwf2ӽM<.{+PΟ}mǯF3LF yh,Ӷ RFs& Nмt.zUJl4吳iCʍ>9g3_rԉ^e6YP}*J&KT1u'9^#EʥqdׇFuvU2{@eS*eMNss19Rkd+o坣c]{ݴ#GU67YA+|78҈h f3*o @\s> DGe @d *& UBD0h{u(X[->G F9Q2Z[Lq%HH:3a,ǸVT˒پn3+`/f;|)vH}]bc8nzoѰjԳƋʹ̶I4.8n`0T5"rnB| 3׽יC9͡`lfC%fs`&",D'!sZcTTKJn>G<& &2p!k(^,6fhd&fc+tLΰ+to&43oS6?]9Lh/O.ep¡)uByE`'7Qĵ'7Qupy$%g}˸o1odSK6=a{<q l)n@i\B!Lk5螵}l$QN#ԅdOhHjr>+5LG##6'+e*-0x?s²dڽ$852!N.0W"7BX R]f=6Z5AoXTzA+=gGњ$*lB.k5DgD ʀWL|Oh4 3TZe;,tbXפ{rg&~~ҴE==|#4 +H@ne3~{-ش>Dd D㘂8˰n?( wPlGH"dzk A:G{V-hG\JBy{@ !Tar#}L3_O].#)V<OYI(\x-b3.xw]MlHw~20{7~̏-Y9}d GtP&2cC\f)nZ?SE89'Lgl_ZOg;!wW÷!-V Vbŭ>E\[31Ɨ6_oi, ǟT;s|<[f/nFzٺ9;c{Sa,8_߯ypuq9;wXǕ+zJ]ݗ%<{RĄQkh(|4-zrͫ麛37'.n]87)БraEq,g#~l*Nv5>W74~ÇC~ą>F`]W 1xr .j]ťy溪-w\(}$*.Vx߮&&[kwʦIb,-Bb_[&gq~V?J:RԚKb7 mӌ[!^ˑV ETbJt֏$m:ɸ<0 Y( J͂Q"d4`L )i%Zr)DDY@xٓC{GiUc îZF3B6}RSY!3D{׎G\ӃE`R(RoBUلXV;uGuyJuQV(G7]}[q^h b4W`+@nA&M@ L\}mQ=CZeITEi)o OeK 6ѣ6*h WIȃ B9$31~M2ED57R|2N&-9g_GD=]S>خpa} {(\׫r_:2d꣣X63n}cF!ES d>6Ģgu"{g=YGCsa89K(FjeJ}0$>Ty%t]W^ĭVy8aREni*yxP)$N I\"Vg 8_2\F~D|p??Ig'R7^ 4b{;ɲv<0ØgI9 hcJ;K8{m پ|Ef競iV,bwN3A2чȅH `d6z,+ ,rg 1p`my]y(}.۔]٪HUۦn=E],-EݎmkI3lo(_ ii_̢i]=9Q $j\f2iuq>妬 >(_WՉ7x_Ixo0}'uv_A;``|tm},+-Oi↋׷:|>%-&4Bj=kЦqʂ*P f>J{ƋƖ& l0Qha1a't:c4OR2AZea ;ːq"\}IO0O tea5q!y8ߵ9ë&M7w; uWoqeǎ"?Η ۟zt3mI9c[ VI"jݚ_wm mR$CvMHpvؗu"Kd98[˗deI/RSd5cY]^ A!ÝZWԺҺ^=MpE5y'~lRˊZV3nݼU;Uݠ祖pjo~u=oa,mIGt<+At{S{w"95?oicYm<'Cs(5E֒{0GTkԚ/j{, (|+5e%R/[)BV zUޒi 'm,(9up9ν264ivh4atRArD2XI1ZtR0>`8OA9S™^iGUm:A߯v?^^ۭdxvg|r&/o#XP6r1,iD!)grh" /ظ ăJ`s` S0(P@D V%3s1X,mg4:4BO<_`T2XsDƳՄKV&Qo<ȹqwV*1{==#K}XWFtr.ZLT39qg… ]$5F8OI}kRtHkaZ4GCQ$mˮ8HܜȖ6@.M+qHub"m 5%C@S9*eęuiBx)FVAb9gm,}|q{UA{ڧiab&Č;&a@.;񠾗3݀.3Q{\+51I~mA ͢3j7yMl2 ) l\1ɼӮ2"п:r[zK9RE 6AN&fL4Z-&艓;4_߹fK{-P9ª wH 8*p$,ҖR`cUz[^?UU`5ت/g-UGp"1m|O~S-/5;B.c ?-Z{ZL~ OdBU!XU!U|nyナt)+ RNV^ Wi΄UKGK[~pϾsg! fq`0:!g+sRGzͭ0q!` N8._E%Kjz4K;Oo+Q$1fNg FLQ;9ItOR$։).IA "3gC]}n괼 M^.Uǵehw u?v11`߀Km7KOi!UJi_rQ{Z-yG$6{0odp:R2h9v+`ZJ--wKiRIRL#:.k8sBCƞF=slR2α_9Ԗ-ՃnUvqnP紛Ϥr{.C TefZ4I;uVZ*rݾhiO 0.r͂|"*F*m0ˆb ΒOcݪ>&%VY Cܦ=pcp[x9d.sґ僃ٰJ>3`8^_G(%9  ]ZJ:M2125o7?ff,ZsMmH;x+&'D3 RIDګ!hYP[#gOŌdEqIpPƵ'ۦS6nx5Y~Ƿ:9Um̲Z?ڌuOuJɜ-^?z߇H!iն<(CƈB*2j49=CE2H%=)`@U{I1gpkVe5r#c{JkXgle,Ţ(uь_kǓy!n'7p54OAїAr#vғi\*ZOL]bA8,)3NA杗&"D #R1\Lz/ 2( h&mf6-U&[FY~4K&桠vkܱ-jQti8^fK9rzBPFK d4<;RܜA2a+%kYȤd251RwC`Ф8FǬ" dT'\x96箨[m[FDCĕQ%4FȲ,ߚ *#P$:VN. U[b%-rƴmֆ9 "D@#2"F '\ ^gk\%.VWq<2 Mwz0G932"Q g3=J8wqǁewakN+k8?*DS7]{'Y{Oڨ-]y?ΟEä.]@uB`$TJ#09vn;r{\^d ={d{\jޅrԮGKЎ\bZ˘B.%{*ɮ.ɸ :z6q$9rv>pڽ¾]oɬRh;f?& 0%m݄C_!! h{.HҚ~*/XPCpUF`e * v7+S[aYےbDyF}zɄk;Qx>SdJKf<Ep)IΉdX[!DJFJ9^L_H*\ #.ȴve w)q!X`h@zCפcZ}4hYlI5r6ܞ_ |JE^JT(释ZY*!CȒ3Tix,w_rlA$VKIcerA)zIƷ8$4U\OJ=@Fcw5/rn20F08DޒehpnH"#AHBvATy(TgZZt@*پuˬx,=hbkD6ed&Ā9'7qK-MQe&-[lQ6@?%``J DG,%阝r2uH˂rkPjCGdVflz#P~3bǥ|DNIx_qBW>5Wˎȧt}Y1N$ol̻[iTarBmك3JIbM?'if#O2ػgLIIv '(SLp%'d p&t7'+47jg`=w]*I `j ׂYohB~:+\Wjtbv~bKSw[]HfD3nj_Doߞ^]\.U iiE-`r6jEC֌ŭ״V٭*5Sh>Ӭuoڼl3z+?:~Y9{H ?̃J{;KEj_j'nAFGHMÈajfY> 4E3Xff>ѓ1U7goNգoiԦ{MaFHX}4 WS~{iN[թ&״pzĎg?ߝtθ{h U %~ B"#9s`m[Cx%C ]&|qU0CQvI]I I(17:ݮdUz{q"b??Y5UOӢ5U"Cl4 PN6"UVԩjsXy:u 'seNY7 JLU!J6n֏$Q_9nN22LB OR3oY2 )D4k-"*KqD& ';ٻFncWXz9!~q>[)WW6>ONpQ$Mr׻N忧13)JF$xƗՊ340@wgTM?Nq %0.Ql6/s#Xۋ4%^lQGcLuPEqq|*tbl--Z羽r=:(QV+WO`.8Vi a:yzrSA4$#ta][y] Ks_ OE, ,^P_-Z:_΢]:ZaFGH>X RcX3A2< Xf\b)U5vΉ_k;7Cx\W{:zpTBa[lwPHPOpg  Aφ-U [(BAJSutlo-KkՁ%v>R# E$&bCi"55* ֫V$aZۥGS*Ӷ FMu `=vڑ<5e?^+ۛapoa)"g.#փP&r4kP! x/a$uPμ ů ' U Yozx 5Hl1Ww|W?û $ ˆ156$D16O9W; jڄ!9,s4i':ݯƮ{mk"9h3P @^kjE4[J[0mz7]ُ-*J;=EO[+d+T1Y7Q=]ϛҡغЎ3q2!bD)S*&pظ] +/^KͱIP+҃2!BJ5KjT]ͽ?8% ֟c`z5?5B.g%ywf1mHza2= :C؇fAd01^=^0}#zM54Xp[7AIɹGbl)d#7z#09WXqЊk: {p攙 d<ͅHP&,hfypdB MEI$ZD9B^Y~u4PK>;!:~S,1 ۘpcOSQD Wq'/-:IW A !ǓrM ĨqD9"8Fiy~ߣY^dγc Syt3Ӎ+LN}.>Y*G]s6vF5M,Jz5w5`!Tr5V?@B*J\R}6'dֈs,6R.\?t~r>q@],䢮LE,2:Hu[I*aJ%]<ڢ~h$Z5!g >XSwu`s F"N~h1qZ&q%_a}v?X92ZfYg=t.3w$gN)AQdݑQ,z[g_Է`ΡZWE}ר+([d`rrevb!#!-Ä*դ?wфc8" >9FJ-X̄#mK{pR3*Ny*8C;EÈIFk$ $ʗ R1HTc411 s[ <pF%R 69k'\1qZtն5l g/ka9Փ>|}jCo45$hhiMŗ,ޡ,TU?63M7pGYզ[XZk!76)!u 67[K)cp7Q}a8gg~6f ܉Nhkf7 7?xshޢ;7CO_EhQ.N~>p^%3R7нޫ[ʮyk-@ EQ^U5%ϭ͏TU1# x U9Ga)\h;O'r><Ⱦvi P3AI2u\0M;Ʃyk"MC`;/q!3OT}U)mQ(ͩ)5ͨe l~ykۧ1m 8y)e,H19|Vyy`,^,BF;#H0$3cA0RXp֒ uZ$u,s~+/OS@N:79疎{`#AIѐJs\FDK!)qZ-m h2 Ykz:,i$Fy4qL!SK#.iΰJ1:,5!JYR5<@YWo;p\rKrB 汌# C0')*0 \_'#l@7z7Lևb͜m?J Egw 8i/d!Tˆ ĝr3cU; M^&%zkT.*=k*)a`c`|g`d7$1ILODC&=tQy4'r(ӯ{i2Ox~ UF8^%gɗSⁿ KA47&1 bNBJ?{fsq0TQ0 p޾L (_%?Y}- wݟ|t~tW+@ף%^_I .?a[DM'i怖:xy i:EG֗ 0x ʠ24*E)& /ˍAF QⰪu?uB&n5Ķͫ6͋Ŝ^ri  9$7$iQZ0R{n M H$n*,}"W_wD8Ϧ{iN±f`yn,B-VD,`#Egq"5(5*aE}cpOh҄Ok5i|9_Z&̛xsXz%ncm7TTЧ8 )J3w[HI% ^|$Gwue#'>^_L3o??Ͼ2s,"O[Hs`T0KnRQM[ޖF/kOhg;Zbt+n\Qݸ2q۪3NH#҆6,10@}-eiR5WYJN;s͕fz 4 W? UA jVgp QzEz nE﫟>mvJݘ\l,#FXB ,5 ]ZiρrSj@!?#rnz &GoI:0bz>+YQG..E32}`. X  :^*;8o*-wGRB|oh_Gy^O#ٔ jx' ~w}&(j}s]=|חXhܽ[m'\+5՚qYi R?WF3jZ Leq>ԱZRAfl,<2K,vo% =z\#qՋX\LJO\+ޙm_=QD`3W cvʙ,§nZv + UX1WY\E\ei 9usRJ;s͕޿:YNY-b%Efjyj]ޯC~mpqy6ŻvVR C=͟z_% y4ouZl\]X–bOͧi'BSP8 ***"9srŌdA%  3~Gv8z\de4Bmyw4/`ݸ}s`Ї;#Ə/z䛨jYR(>vQ|vR<~Hjc k6:L& |uܛf|ӛ4Gl:\Ou6_.[UՔd(J4?֯c3FЎ{HY*Ĕ5 䢶 q-bvJOp+mCIwYyBoqt`S. LE,2:HhSFn,7$!Gy>:$u@v?@)H{@֜jd<8IPezH_!ifL0<%)RMRvYU$^)UeYUdTUėeaOΒ kVDh0[TTQNLU1ZCk_|Z7ruI)c!"rhDԂxY9@:N@Z$fp.B^ڕ+&ep֫{J 5MQ󖎻wUJRu%µ%*<$TXiͧ9GEGiK+AY!Q!r0t  h+& p%LhbP} RXV"($DC_տu%o+QB%Zu..qD*> 9M΀Q3&ѣ)kDlH=I:9xFMg#t֐GiC{\ʝY&CVF!.YJ1:4`퐧Ԇ'* SA║>3$H#!W!H$0M|?Po?DI~,uzs>B?/\ }pa'՟'_dcVrbMOfrg>}$O{ a!#WV,(CS2E kE)& Nd^n0 3j4Z˅M/ݮjA~ӛR 9[4$iʱM#NWY gCn +=v ր̈́ck~^̏9t~m5_?5B>Ga]?4I-5?Ȇ}_.О:޽jd{qxǚ)>2FyWckkFҞewUN{65id&|6}>^l'&u! Jj3" GPqڻ_6I\֙{Y&!QpLgWzwڊkEr5GK͚4cGFstGٱ~iQ>ĜU˥O|=_Ru/xP=:"d5_08vТJM.Lb-b}FUdk}}()aĴ8"Ȁ 'J!qRni^L#ѩ,y'zY}&qzL+[L7Vc%!)9E 'Ug\qK E+]hWB^d{DB`/t)&s*a ӃM3!}++voSzrჵ4C kNSVw9U>EĥF+ B.<כ8MwN^}ngy8:eٿoiA^|toOyW~j: ;wsm_KZv4rYhWza=G#/..;y07;;d=tJb;ip)qKw]wۡ۟]}]mg,n2OHo[:KPűʓ6`z 2}YJ § ; ^;{P^,[sy/ڋo./9]>O/>vr:J.tQBq,m$LÔ!57: b{VzI\O\O,r)60Da-t\=Z}+. ' A !U_憕(D`1J)KD*ǣ-C+u~;رSeR3!0.s&OCq. JI΢7 :!֢RRdznl $I 1IPDR"O5Zks_ J*&nZA7õʤ_]߹/~Z7;Z.\QcX}ŕg8KO1J)T}~z g3 ;W ~.7֜$:)*A$jn37>sT̜2s86^8vȳ ID\CS0$,&Ξ[s=?F?׃% ֚ }y][ZS/F^jo{|CRܫYRB(  W$KrMcPjC?|I@\vy)7  V ZDhQq-ڽv=kBŽxy'l: 'eP$ hLN(Tk" 恀^ [atEoCiKMM d2kB{"9!HC[n)lH${+,L=YٻmeWmctHТ@O Jֺeײlό>w7k 8(Ǚ<>h,wK 43A?%|kW 5hU1 $e=4p>2zW=rYT ƟེCq]3&]_#Kc}QB9a1]hO{,0\)6-5'@&ԥ*.We|zWtsE%Jm¬:ĵװ#Xy4I1Y EwD{4Ջ7x 2LufyQ'Gi>8~/_?^}y9e䧓ӗ`Z`R B} 0iEE*[͍ؠhߤ\Un(j5%Vm.jq~}zVyI1}b5fMsT O6 a6Q:ϢBuT+7t+ 1COh +-wHJΆT<ʲH*S$EB}zwe ~޴0}&q2 )cLe?4:'?_찍:NhBȘF$pE-YqDH` LcdᝯLYs\ C秾KILr^R4!d ^BuTAA8KXBtPUl/|\HRdaAqnP}  ܥVt7_xGCP֧a 01=/wΛm fNۑǴz{1(d:U13‘$N\bܧDN{^|)A\aML$,!x*SHfC%ŌS~|0Aj7$,Oy7Vcx\d?VݤD  ?]mCLGxUsZUqcRiτ"N4Ml”4VWw9D&S+ZQnF;7#Jm;n>@nڮ lSnJWdi#V3 Q7]rrXuV)PBTx=yHZ엩p6^VeNKxP Ě .4-K̋Qkws,c2 pb$AxǨ|mol1AEﯞ?}b4M8cł+qj s2X&|B"LK!v({@ɑu'pMfewzbc&9Нc`%^|Pw IS`ٳ\ؿfn~DhGb |e1MKäԅ@# 'xbl*gI8AՔkK3oWvpP(_O[r.ֵOA ʚSaүRB~pn*(@'[=Oh07X n_s] @$y -_;e-"9dmmUD&ӆ1Yp⨡[f0.̘ _uxY 5m&@}M"o3LL!arR*,RQz kNyY%q&EC2a>rr]ꢒM6RaVWt6 1KGjqqVeQ:d*]*N|q'_gzًsO'/401 P`= CF:yEEs#6(Ǭ7)W>fʽ9|ZMA˹@Eofy^#x|y2owev5*QB.|?0̇P^F< Qq[Vҡ(Āq?E/\\1X1jIa(^f)}tt]gCu*eY$N)|}j>2?oZ> 8L`X ƽڲW_F/YvF'e4!\dL #x",8Q"^iwdPdΨޑ?P{=$Eu>GQz.qo||oyZrJjEM2f2B ae&`,ML :6Y4e 8omy4!E136.ת\XzdAi[[_H:򟧄mjV4Eh[u'ѫp:B)@Ƞ/cHRod_q4^`żW="WI2#=Z#1*k4>Čr=7g&@ $xQqOl<&,nC Ө - 2hL111o>fX?j'[=$XG7b0ABw2artkd֡R.ȳaTˆa&mb 8ֱEuFRݧv[EJ}>,15`БR@iNR{!' .$S)2<` 7*kۮ\2(׷ ֶ7W,h0YGNaNRAzQtoOM]_bhim^Fzj@{]8 ߮h5IW~U9_Ѣa.dxb?(Ogð`bz|%ARg\XJFwmvTgNۑDžR⋩z _Q>@z/?^t"*fez8_cf`6>9AYXT$ɐ;?áH]xݔ(yXg=53]_uUeloJ"34! -K̥؂KNp&)c$&BJA:Leƹ \g@YkC+6LfQy8ZYѥ[EZ{Ko6Ñ\pwL3}~`DT=i 4Ҕq!,^%Z8o&K@V4[8$Cg>J~%sy 9ޅ}C= ]# dkƤLtII[s$ƕg$VoOY#D`9J`ve,G#HƬ\XMU(@xi*Cd<}Le4NtaHhe6"f*#jx x21D?(fm7P݈2=[]|ަv~ȝ3EL\zŷ3]Oiftfzt:9εs1I@ he6rcP|́ae &v ~g=FvsyLG˖Ahk5>ETn+z!3AL[ E`-_L"Q/%kcJRZ cW'=(HAf, e`b5(*/zg[gT3xj<, K>=yIB{ЃE2sB.&m4$&>Z\by>[El1@VG<71 W#zr@msVٝJ/l<,~D?:98sэ5<| `cLlAIʩroEdACsrM~7 tP`pFsc$7 $9Y.UfA ]79(\zNDImr.)|=L"vJ̕!b5qvs M.ɐVN&ӛfq>ym\3>Op,  ǫτWu+kYYuYd1|M O14dYYˣ4;h X{ { ّYwl@v'v ,lɪȣq)d,*g! q,eb3NxoIsj>vN|CNj6~bh~Y_'R"eT?]"fgy#SfYY`kO8 o#,)kK*171! "7X8r)Z xcMa&v%A]]rսe.ڱ)Z=^iQ^? hJ y߷6}Eo}g]A+Dx6GL eY5D'jY#:8)3@YH7zI-+Kly02LQ- W޲2}0]ɕd#gBZ2W?\ֈT Y6S wVO:z23K{EdcKRaeMZwSrc%2LhY3eB=Ud.Kz8@"\1{t iㅮ|XM5Pjښ`ËGi; ⭭Yu~oxpn>Og)n-Yc7n麛wz4} 楤(sHJX>;LB<ḝRKWXXJMNv<'aӿ0鉇Ab5gӋaM.dz1FυuIREU܂ܢp'||Erh<.S?cˀԮ0#4I]V;_atwx8gYt4L)++rI=cLvZh%nݴxM/ٶ>}r//s[Q yWQ#04!{9ảg1% [F3S#f'%M뺗j.ݫNܮVzA~jmu饜8-)纎 y~|]=t'*7A={.Qj~,$56%0ucdž|}GX??{<׿w{?O#;%*,wËow pÐ&V?t9<+w~X=%7Bjt=ȄBp7$`qmc[%NKCg'cokL@u9(V釓nYЎ|,%c cx5 1'%Ny@@+IY"=.VQ~dJ;Aߝaqm#RHy[Q*9tqBChnq^ ɬכ X5}r囃\ڳ7'y)_h,{I-{-`~Ã/, L4?:A?w^bjEl}];-awMZЋn:q=ĝ0;,=iA-]#4VBȓ^+,3 ~yG6 ~հmS&ַ͘w%pH);6WË2-#y0S,B~m 0Rn}V>To}3.{1SVi =,e?($̃BI"˻q'Y3M7ZjƭXzj {A6c|o ]zx6-X?x|ר}c%lh΄ߒSO-e Vl^y,<7 9 8H. jc9aIRGLv B \Uwd'ŗyck8I l|Hk*>~QOUR[1[˂m1gJi\*oLB\fmHh H@yŽV`*Q[c3HILO 6XU^DRD3!䚔+\VnajʶP7as$OXQtyf?o`04^\Azr4 f!}ɴeV,̠̣%[[ZFaLz/eQj4x M&ߎYLlb0b}lMo{ZO:`ʑ+#7*82Kȥ VJVfk㸲e/ n= ,@c&ʇ.zL(R }O f9m [ptսSU7ĜؑC2053Ts(<Si S]CdpnԟՉqWqx(#Č #.xs γgsÑZw.J %4jUjHʴg R>Xx!i>խL*)z"NZ`b uvbF nbx#9Y<GNNj‹7E/ N;‡z;>rlF޺8gO~spn[~|~Vuܜ-;:\ n{,J֜׆L[:jᵣU6ly/iw<>Q}ƎSiunDy{0>.OAz?]ۣU.b3 5SqNX)Ҷޔ&:iI+_|-s}I#)z&y8x=;wp2gv's],YNڞY}75-||t<ʍM{K^w=s~{GH~ox\?._p|{g = Sy(M8y1 ) "l(>yޫ`o J>~E[ᵃлZleqY? ն޿<Bpr6(GGPL^M'm)!X7z`uU~pokoj/ﲊƿɛ!>WsP7}:g̈́}8~3po'Ϸ^y٫QK']˽kco ?>J|%bvIЧt; g5CCsGCL*}B˯6_{[bQHz҃SJY?rEA(oʖ~_qP'ǟV/V_oBc=yT/]ȶv%צJmݼɳ;ϫ'^V뀽Vp/Pڎk55JV(I ogDW{ͅ:\OKW|/? ] (կ:њبztAnj@i-{/tЩ6̆:Zv:Jzte0LbTtz?h:]&3+K2#YfCWkh6ꪣ5tQ >Gr䜚|U2ZOtQ3+Oge|֮:\BWyuQe9UDA͈Xk5p ]uA:]uHW]7שUGv~ؚugCW]& ]uv6kWmy3Q /t+ D ~Ovu?P:[t%+YSOf%̈:'^pUGKtQs+V!̊X1| p ]u;OW@BWρ2b{S16'(AtnxPƎƽH:궗\=سewP~!ROr#8?_]|f07;RRBp֞#|Fሺg\*x:Okj #QAVFQ<0ke$ w3p@6Lk片Y'oIsh/rd?SuÏMR8GcVɳ)Qщ|-t7=eH>S4O1cQs >6ݜtq-?~ ߶7gK^HY)-A%fQTXrℝA%)&K\*JQ(vڅRѹݏήl 0wcMN\#u`?,4zs)6Ԧ B \ I ,ۻcSB"gkɸh1 z }4[?%VDHLRk,T2&CJQ$42Pl)IQZ6"7b5rIМ"R\dr&AI+Xj"IARKA(!1MkOfHIYVbYUSʇ|3 ф`bG}sp-#L% $BA2VM :Rj*0BoeJJo-Kls2U_vHTD@U(Xy2lr6lU )c VMBmS̈́"27FL!(R>Q 0ȒQBthE2R5mTKZŀz,MȐVb)kCBmQx& %*  a|C Λ:8i+ʄIÀDmmTbKJ|dp-C f!KDh#2sD$l0E k*PYjXJKI`.P:#2gv4Tk1Ԃ2 llQ< RSm'LEjD*Y0Vr1Z̆6Q6:lI 1MQꫢH$(MS cb2 hZEo(!.1;EmTzSApd\aS`hu 1)  AdBE@,feʔPs 5%@x`?ia;BC\͡LՆ88)ԙl.Qux+9mkLFyY{6ÙRP(q y'+&"td oAc9i[+W:khZ=# 0|IGHVh Cݚ7C6Tm* ,\]TBNU~T}*yf]sZ 7Yre5[BD8X-Ʒӭ?|tpMow:.p-`I~_VoxHY((cѣQ.)$nȋI}L! Z`x W `#8Bdk5b L `ZӤU%,!F8Z LJuX  zb@ qv9;v&Ռ.8Xj5yZt2,@'(YGn8XTЙ$ fJ g py Z3;썇#WyhT1'j**DŽpJdUS(!0cRj-Vc 7B:b8k4QFk4r›Rl#׿z+EFBZEVH¢0\5p_ US!Wjq> WpۗG?O^kGs}w&$P1`0u0 5M63Q\% l\FiPQPf1jW11;~# #y$4FDx31r@yu1'M iQK!k8nHJxKT*ArWj?# Ck7ŹfҤD˱:ɨTP=`ma)ZLYi PCz:i^7îP >,7`E]("lEJ# \5$\F"١O!# UA CuECTfDm#1x \:%XW!hSQIb1*ioLF&eǕZcMZ *-Agޝ;(Z 1)B5g9y6)You~5(niݗ7c@Hm;*^Y;8mP Fx2ifaв5A3^/O4%0h%#p&StplpOĔ *5;,tH02Q پg3^G!&S,(`J@Bׄ ;jwmv B`laI<{ V~vt~+n:::Ѓ\CdC):aV_Ⲷ$-g`N߽}{-ZQm55RrK6 O;XwfO3aM?jW;іZٖ{*iPj=}@]\fuGEf i2]>eвhZ-}@>eвhZ-}@>eвhZ-}@>eвhZ-}@>ewmI_\P~0CHa\0kdHJC#8$%NvUuUOӤ"HD: t@"H`"ZGB."Ut](%'1ꀤ4t@"HD: t@"HD: t@"HD: t@"HD: t@"HD: t@"HD:c)!I;id<: |4: %P:M:cn:t@"HD: t@"HD: t@"HD: t@"HD: t@"HD: t@"Ht: SId<ht@ְ5:FLR^t@"HD: t@"HD: t@"HD: t@"HD: t@"HD: t@"Ht,:-׏nb~hwB9OpA@c-A|<%kD4%l #yEc-6WB\|pXoq4D'ڃ̗swBq9Mzoȕ;\^%];_іyU)dn]7?-^1Xnp ONOnd-]sj.뫸YQW^ވo4{H\j$ f/_gh[ݺwk0$0y%Qݻ~݅ uU!*pFh^;E6>UPK9 \U*Os{pR6 Ez&xa:t-NJU>}_~0$Imࢪk߉uc[8|mRĞ?qO kp0{;^mA>tm\,M꫃ }*>wzc7 8HMGEzF$ 󢭢f P~o[`~-Q>xwB>]rx\_ۦ5E⺌6 _X8πjZr&Su7 UeI%_Qf`hۦѰ4_CNK5w7K`qНfٓi6,ѧD 揿a 5$0d/vI៩]OҸaF>q..0fW6es0_MryQ:*:%םO? dwd?_Mˡv GX/7{!#?o,-.5dXvYpe2]lw&!ʜ%H%`\4ޥ_Wg't(2d؈Vu^a(5Sl1ㆀMBO4;]JWCW|CCU+U;ڡ}MU;["['\zYDtw},th9QDWCWBI]!`]\ˢ+D+\ QJOtut%uLDDW[BW֨tDWHWJsyDt •Еw we62]=]Yv k+kAsYig6sjwweǎPKr|H1c!&#wvTڑ9#9E,>A&F~a;)6Ԝcm4(1=oEFCWъT.lLtg-µլ[mXLZl]4tp-y+Ut(#:B/#+o<3lEvh54B]}5t%641+:|t ?\qf[}wGWP:--J]mkznR{q ]!6]+Di ҕpv=w={.ڡ5+D):Fΰg})߮t,d4DWHW3+cJnX CmdhRH#+핵6"NƳԎpu4 ut(}v]= ]o6LECWGԎhE QJKtut"+,"z2pQNthm׮d3Z&"B"+Dy Q䆦=s#C+ZCkVjokɠlAWj[sJpw+l.qnFCW`;xJpm&Bz Nth]+@ٹ.h[" * ]!Z+NW #+%71EWJHD4tpu4th]A@QҕcEsBv5:]]!]IFB?OC.րO˓ӫlr Gz֛f4L䷛J:2\Pde.sR֋ߩa9A';4b;]7\it?F^-Fbu7ўmrydrvXejsSÕe xarkr*defԕ񕫤BAeLg+T鳝|ȻO;~čWË%Q 8.'ɜ^lљY WXfDkDfD87`X]dCh=zh}r\fj]$׳ϧk?M&/vWz`/Ηռ;ݻw`YgYe3}ؕ[]I0w-1"=_;YIY=Bޝ<B%CH`$LBy," 0[X*{k @gaׁ3,,y=Πl<ϼU%gS,h66KWt>$0PΒOG9(ӸIe 0K ܅-N%p:e. g)R[ }fwYl svsG}X)+V7%rXOe w8K~(kݦ[J?^6trsݟM?iFؿ oٚ^[#;1ͷďfx_c^坭çʏH߿PGӆ @BÄTxþByY9óAZ 7LNj?{xeQ"-2RUf:a6{g)BE1Ԛ5>d] *[nļ|ooHi=PXEn *VXeC^pW SR3# +ecȔU<gCB0qB+.sǕۊ˪ "?csa78ŃݡE@6,s3ovQ" '8/zܸ]7Y iՖ5j-3n"yW .3 ҘBj! Y+.P^uƯ̆/u|Qi!/3 +T\@{{@ qM:4L ?8>M%'72_.}4*]\tOfuJZo5vz? |") VK׾9;۷KXS2)Ôf @;\Y+ ]Իʌ٪,<Hžx$#08_<с#!1"9"Ӛ_Duɖ@Ttz*yRQmeiIm $d\j{jlJ5_XM3^ ue_{_W_ ㎓$INq6vO߇{0e i|*B- &C3$dK&cex18=v<]M;^Memz#M/3cN^Us#'* d4<{nX"vJɪì1XŁ2!CĢDC&!ZQp$E?!eX$R8aKꗱ]1G/GG=Ff6HAD,HddPQBYY9@AP;ičd_ ': "?RP$#K8;_!}ԁY1uVӒE*E/nx 2 /#YVlG93"V@`3OJĂ//?G}?܃ [%??n:vܺx7<8UpFy;OwC(E?J-ޢ7&|u.M =/2uJ/+2N7ƛc3o"ʾb0O}^#@i#*fr|&t\8-<V"ȬoN oQaǖRPBG}KE+oF_]%589y͆L@rBkAs x9`"ru .zj5h;0d6ڞ(JU0iD| da)eDwlDX kD߰lz}fkmH?kyь)H=dkYIBR(Mc茈AI}bKM)5It` oZϱSCAL,?*$POW} МqW`Z!3]z6l!6`hSg`93s||{d$yŹv&'xuѽ=zg<Uڦ) 4,"Nhs4fcjΥNǺL(,IrL EzgN(3DB2XWMOWz~4~__@@dZZ[nWJuplݏxxH0gK\W<,BKjȝ2PZF(VCq}p9j"R8GJf)l9$5]4qx;-EC9Tfƚq Qd`0M6'wsC T jG!ܷcP89-׳pxT*١uˬxfYv)Фd2ʹ V&fblրI҃O*3BQ(VK.~!5dM@̕S3*;+O.H_%ȔܟRY딽YB&a~=jǥ޹0,iBW~%6WN?ʎ%ƿ{UD#n<Ŗ}st#χ3=z岼nueԊIē ~={D&vld9D^:q_MlHs?uޝOgqŵV3@ G0x,5┉XK;A[.?-Q=sjv> ә?g;֑,ff㵞ÇbG^ q{5cc|j7-W%e=7V5=xg˩}nc~>X2y_<\^V]("U|;kc>^jr830$.IucDc@>f]Y~!CL8>fbYǣr\f^t`|ɮv`?i9{ť+RobvJ+ ~'~UlӠt Fgq0"u?˧?_~O?'.짏/?@ΟhfKu.5soxhz[/&|窚ܧA\IϷ_ $[&P~$V<(Q:K_!̯gCY\ JRIJxZb C t XίvO%lHY{;'Jq=IBPj!KcH֒U!H4=QzkZwd2)'82CTqxt*85}L !*;SNDC5xQAT-EA/=9@թrg壗.1>0!2钜t}TmzYN!#-Y$'Р-I&8$r8d,8Jk2{Vrٔa.[7~g?%>Ĉu6wY7KŴlP0PO1AtoPos1̊M?y!xÑQ(FX+,4єOtd\^__*oV?kِb4 C$0uXTT.of-?BǔFś)idgw*t3:"L7hhpE>>\9Vfx|Q 05Rh݀FӸBt6}]#kVVSY e:#hOS}Ul"~+U|S(;7>ٗCjjU>m89@k ~t၈cFC|1u8,&4Bj=kЦqʂ*}e&@pCRx#R 2 & -,F Nr_YgI5S& 2Q,,rpR!8\d3˜; i|dPM=f7wm AVߔZGx:I\aIGtMvkv̂얡C'v̰V.,FѦ%%kL|B.~s\0^!ݛM gKA)j&MWz}%-' t bek")IB s2>lftL[e[K 95e^yyfm l~ykCۗ[3h2y˼\b{Cm8yWro0,٥=0A4C@r_n<8(B,ڠ Q%KXmχ.{.w^o\tJZC^۪2RKqIkp`|])#c VrXא,my[2/12;v|l̎u|cYJ;˫2a'砗pɠ?>,'TV ?8>R38H5'}x}j-~)^]ߕLqL0,jmkJRHtą&sk YPw_Y B6B!Ǥj$V}uiځj7*W4%bf}9'ɹqIeO~_J?bYR<yKw*+!$}|^)}ra$ N,1gׇe:Wo?&hQX9̹Y59x-'ɿěK߁,՝]e+ &L4?W׿;0N!GkEݐZGq?袵;Qwo+KV%IXۓ-|^XBJi暽*lqq3FLOκ&GC}C:'4xTej+O[ɰ?JKH^v*mz>GUS[ޫWo[H3(w, 's# BdW!@sRoTFo~g ?')| gUj]SܲQ)3TJ%‹t:GI6";pBze<4Ms&Yt4eEOKs6hI}AF NBjkhpQl:1u@cS:I:fM*î A>bıDJ"+֞ +Zb#Z-F r\liEt-m ]p[CWVdHW0L'05tpek њ(9]@2"ZDWXժ-thޯJIl/,ᚙ홻BF!t Qn{ȋ+-X dpw?XA =; Te 旅8<G|)娐D;oO ~+R_T Qǽ8k{Sz~-V-,+_߿J+/v+.Ttc_ G;jv.NRtRW KElqU:"yvcm;}˭2*Xz&r;fό5:cm׮j[DW-thwB姣BW =`ik Z}+DYGW/8B=e-thwBJvtJ جEt ++M[ t(ҕd69X•-th+v(uGW/ɺB=µjWLQxBBwtJ3U f 򽷮_"]m]!9WƺBJ;]!JYW/,זi ?jն-te!;]!J&:fJo P; -v(t;ڵR5tpj ]!Zt(meBWL(Y [ZCWW{OWLhBlGW/-`m+Dپu"ͮFa\ @(E!)]>vfҖObdڪ^bB/2uwX碒$2ٚ ̶^wo`٪;Ymlsos۪gy-L׳-lqFkw攖n 9Ǝ6ZB?"cl4V ڊ9BAOW>M/UҒ/?ˇEH;x2#RHeRO3\(~V2eN~<PC/<2oE*tBߡ,oa]"u/%ܛ p~UǢ++tQ]ds]%pصZ?dz&ifZ|њwXoŻ7?-r^:<9>έQ[mvJw>Rb(q!X 2eY;gTt3w*8MbЏ qDR|3Ǥ:Z\9 4! [͙Brc.'"YuS3~)KWD`8^|2Dz4) K$ʹeiŧX~oD[,yEuӳE>qpō:‹8<.0/8j,ɹ&W7f9'Ƿc/z>: ~Ȣ<7ꥐ Ft>)o?;gI"RԛlAW2Kâ4!䆫hL / <TdYPH$Yf61;?uͪ6Eb_Orud|p_m Z_aB!ɛj' =@p2]nTP(}2ӣh,{i8c.,NOBFghFqKO?L0iR $_%(xAQNϒؓ \yԃ$P׻]aQn_ߔVĤLA@M7;>dsGTP,E7=Kfra,$8auv +Zl>NRY"wn^[6YsJ=Y0Hy`Ϭ >7*X Pe&[k7kgsD0(ŷyoЛ-,_ٛbmi",)?+^|9auoח}`x9w C+L[rD,&WۈL|Ht1u#q]e^yʭçe1HM*Mj)UFoRIL&$4oRͯ)J](pOW͒Ol(QgwF"8ZH*|;j v,riD] Jy.wNE)"=aRE}і2@ry|K?蘧0h%rLLk 18+j< 8N3p8;z~uV9_ENnxdUrǗ;_ռp^WS R[?ǎ29-Wk=V 's^pbb {FsNQ][oɱ+ľ$9}.yplls5J)KRZkS=Û(IQËa {U_UBTV/C/^>l;.D{>< [$0ok0r |ࣛ6b`"N[W?"/b`*#~o%ܙJbVh <1&o9].R|ȥ}xKA2<:?ïZ\7ghPm$L`J9&=ei&^ޚo$/+!p66re_1hL*X| č3DDs`[AjV 1 ، 76ra<QICh Jˣbä3F>-/,<.sWL3S?zPZg沋/{D3<{G3bQtf>0g|Sw3Re~=UJN'Uߧu8;kdMQ" kN&&STWQ$Z<8j=!٘ H6\\Rf_2A@s@| 1!Y7VRV WWJM]wz;ꮾU0Hwڟ\G458z{MίYBsy4ݝfėʏwӛVSb ro {7rX튣Ymݴς'<$'=?FǬ2'RG U;p/dmomr~rl|uO+Pa_$0ĹKd eߓ .auD᠒lT!O_5XoP]ǧ/?Ǐ_O_(3_>×O?]s):({1 ?Ghh{<&g=[}+rsT-4h{~Ow_{ѶO>tZh2z:a$nM:~6DI֩Ԃ*JbCe[#GbH/)p*Z_P™Iaw+d.qDA0hIR P:aR|~DÅwZ .1lP6H_ D)IA:ṮkGڃFv&qz7|x|5`Cw.u_o}ĸ% $G TzA4X(f =Mo:Jȥsi=IR:4@T%1'L+Mxj5pckK|`)mx1AMw<\RLpo^]7m[s&#VG(6b vK,W+q0HaPpT!TH Ƹ1!  eܙjbGq-#Da$Z(R6If5Dy nzS3N JˣW1:P]4Wo*cPXdN  %@aY @erYP(Uȅ5Brd<3)Z5c#'M׻og9T궸jqLfZoW,KEoީn~ kzKJ) ^jc⮤[뱵[BGi7:0.4$QƒfE!d" :!d{j⑆~#c\I G~.Gٺ x >Zy&s!~PdBRfŨoV6}!G`X44_+*/$- M-=ypɥ @yi}r FD-VD L9hTTQkIh`o%F_{jkjg^K6Q'XHȃ.Py6whE4C+=YhDb 4)kR˭!ɶ-Cs֬LIPFՎJdi+1rv ug sLzZ9^iǡ~ x7>jSn>YJCu?]d9oҸl ijp(RkL!fU #攵Kĩĩ%t.0r#VX'4 B@#MiHrD}JKQ`C'5dB$ @Pɨ!U:vO8X|r> Z.r{oGEݦ3"[tttF6JeYTuUv7&KFH>Q,Dc} +h6 y8GY4*T(sWiH2L T Y㍴Qh9e<5>q, !HU A0Dl= b2\qe>keV.o:C5V;&Wg"1YƣALWτW9Ώ ={=%FmЫ#K]i-k%wJhHBXPؐ|cSL6m{HO*q 5;C-Rq5l5|4k?m+#ͶEZgBZOdUCnm)e (ђH#4#yh !(ZEKAmcsNI=SDbۏĀ֊.]RDjlDƨdbg(o*D,@'S69 "S1հC1rt[8.Leo:^Vs͐tsp8k5ф1 udoLB?ζCvn0ήMV&[(F6CQ|>J.r{𸜃ɎAhsW,bƭmyyYK)+wn\S~^ ׫k!Cw <+%>r Yo`ۧ&yg!f S{r0v(+Xnm~&\ R=E>naԀn8BY% GF1,+wXUHmk;,}&r,kZ#*i[.ENHn(Dy$&,nD=~@S=5>Fq)]B梘Kse0N\_oO̍%*p^'dimQ-HH4_"ZJ(q>J̍Ŭi5Q4BU gP Uz$$O2WI@9Qk"%B_f[`k5f ֻWz?A$op5-;v)^[=+ f!'I&ZhruskGQ(J3PRߘ/chQ#&CCz(jڲC|'!Ԅ<xawո4 [$*ꨙR2 Ajx]6n{lW <1oASaLAJNDJUH2cqئUY p{zaLIm'`~/6'dt8~A*T"- V[(09E)qL??)4F9hq$ (v R|b5DR>~~MWMoܕ.s=\廣]?]0*y=>1Ξ vJMZMZ|m2>'S{$!끢$ @ChWbc@ˍ~xF )q2%59Lnv[hx;^Lz"<b?|[ɝIҟT R䟫s14)MOë6nҗ&Oͬ[h5ykOS.f._r7`~j&"tW3|]ŭd]iɪ{Mxr_d}o^=TD:^U˓;o~xQgǓ/V|qP_;*E~k{o!L?i~GGX<|_֨ٙ8>tơno2\us/i)nYKjkNҼٓ&>;lwcmdk3{ߵWwJZ!1״}'qW<6i%]Lϛizu>\>MmW&뇻goW-];bnw][4/|`z Z7AwZA2\.r&C?K?-0|*yVsNFbz~ӮI hLV_~ޠpzzvzg'i|jq\`/W _[4){\Y:^O&F},qy1}gC/+ekyê`,c[Ē/ޠm :]^ht}#gM=> Nil[k( V 'o_ͬC>F|BZesZ}M؞Z3qڇp'ޣˠS~E}@0+FV1{t,:~^bE/5{5J|bR e:WLR/9q*%}vToo,g’1]ksԑx?K(SRPEe̕ @R*`gƊsp- w I+Lf cBsx3 fQQ; QbȲu%Q;[<Ùɑ^_-4d Z o4hFPHf$!M";P0T0 3.MmW+*%}*Ś]1F /ͥ;4`-Fcm81df$ZZxy@t+mC ui@<1&Ykw9h*rS_JZ$}.F)82RIn:@G*t ]:%ZLU:X}ck5`UF5:/&ЃF<@$9-2eF rL熍%17 b="A0ޠ{)-Эx# nAf6&MƂՙQ,TG,G+b)ImL1xY'aO6V8`m+o6,d+cCX'X.z+zK,#aVՋ hoT ÆZ&'K/p7i%Y cJ9`: k0R;*p k8n Ȁ!U.$ds\A7fhN&?W F:) ]Fd*DA<DtYZ! n) n0V*V5'U[W2Υ <9 @a~[?"5^0 `\ؒ Ep #%L <шuchep?{Lڰ b"c)X-ḙ%հ]iԬT2g hy}LpO` " ˓ey~5([R։zTR %q$z6B zxX@aܨ,%FX&@-0E z^B{~)"`pGchFY>Q"VBruDL0!";ff%Ф!F\FC]Xr준Kc$diAU,u5է+zGj\o0yTUA??/V0_(@6\SlbgpTQojK;wly 0)QqScrRL q~ޏk[JWdjG(U]-bS,q}+;{ǻa=~7- ֺn|"OlU?j_=ܪ4In%;[՝ȭ`_ VhroQnM[5U[5U[5U[5U[5U[5U[5U[5U[5U[5U[5U[5U[5U[5U[5U[5U[5U[5U[5U[5U[5U[5U[5U[5ʭ$=& ȭ`Fnk/R1jroQneor&jr&jr&jr&jr&jr&jr&jr&jr&jr&jr&jr&jr&jr&jr&jr&jr&jr&jr&jr&jr&jr&jr&jr&zPyTo *hVV^n+iroAnum[=촎dݰOak3<=AGbvg_&s8*.}zEG딒f9,.[or|;Η_1i xU2oDcN5\s^uǪ~Q]3)lw>iAv9:MK%^..NA҆=pήN7M9˗ 4m~}TW+;cƓu2-/ c<[Y70}ή~8=`=SwS'fNX%LNfc>u$mijvd.&t FW0y4jGt9 l&:n=|]dw3SpD[c!G:g &37n\W-(,зN~>\bW˜jsn[I@<$֩+%5B]`".0^>Qc;*ݙs*SjxZUßۜȽU: #{ns4_F=$(E)*aI`2)y?Nе~ui~H.fF ᙶN&HfJi@-pq6Z4 HPF:nFpxcbxA-x q$A$ % 䋔`^CB^ŐJFŐ ]w.CyiK?wD7i5F쏧S•vT}0L~fu +$=%IJ8 D _4&١HcQy,rɩC" y5F ǀ GRu WQ I EI*^GX7M=a8$chQ2r{vCEDfr($ݠS. WF(I壨%`y- Tf5"/(kYiC϶<[9f?5h>^۶dɛ'ާ_oڟ_kţc6#S0ZI~^]7oy0hSB!/E s7NǓIJ[n RG3S?e}[]C`m}(Cˇm`z,|_7mΊuAj.-īͨʇhG@4\ >J^R)+s",Co,<>o%j#+ߗˎ.+ǛTI>vo-`XOyz7 2X髵'//.Go;yE !AY> ͷcgF]'mwCCb teu5;bR;}޶j A4k9`),Vszb혅do{ HvOMNdG3 Cp⽯10JfҤID7YOsˋ!)rkb^n)%0Ÿb܁N&_拨_[KPVoy5s[Z37Q"+S#ARy0Z`he5U8j%X,Yo$x7{CϬ|?2gZ[þ- 7R%Oannv~BSwUdsTscgmI WZrV9Afxs>/!A-/-IJG^erd9Z=oɞ0LՓ۰ KGQ6 F6ԃ9K`mXFwG-ݛ &'ڷe=OH{)l6eM/KZ 5ǔZ;F7 Gi1^ȤRRKqƷg5YWͰUFl(l1uޑ* kd.bC9lb|nu619Gђ?Ή`iy@N2ROgO'#8#$$JB&6yk*t`V{+?G2S`VrN g\L p46"p&@fS#a 730$5AZX,@v8> B" 1,,VJP5954Xj fi>nrM&~&MٻJ>fVD@s3S[ D %p:rl V(ˤ$&k fJFHqAH1\- *Ff$$ 3Ss̚195c>[%.BY O\2%y`Ǚfh4~ g/\cGLŽZp4{ V AV#G ÙgK(` #U43 j-I{* %^Fґ0NUYeٯk.hlq֖,ZX;[=D, czΝ[%wÊK$SH]rl)EYaBg q]ʞ~]^sǷ}FF%Y oz٫s&VR& 1A}0A%͍h o3䕳s}d~OfA[\qa.̂^0+ .(9SE\Ŕ'FU #PHEVls:^:cTӑк'XsObտO2ϑI*1$H䭳ABBb"D[TxǢθHXG"\JlB Oc*5(K)µcނw#9NK>?pR>5,I 1{XV-n?b%!=>7Trg[6KỜ|ۨq"2I]`!:0 E$d(pZg{P?$D@4:&(f^3K`@ ,4ȣȣ]9)5}.g SD.&F 0Y9(0Uwo(85!!tHCfrI 2t?7yd̠M}8Ft+S7_.wK^%fQȞͫIW_ᬸM;ufM_hn~~]xn=%F:hxsۮ-ի] y?o"ooM\Qb< vۓCyc֙]Ys7+Szd#>XD&rv]T8-()qoc9$% uyr 4F@;e #F1ipɼissodSM}՘MCH_3>Yq6*O TѺJć|߼;zuWOG{ף^;D|pן@F`\%D&&p1Xj߮Ѵnihn*EӄjivsX}7ӭ0OC?q]8INs Vq?$>OlX槅N|\UŕtUT 1ʝB(7-%Xc"[;.f; Ewn(Q챈H8S@),8ψƏBP1' `0bhg(C-bwx2.Fxa<$ V,`bCXRlw΃ʲOBV/.껑3Mw}٩rӕN'oa&Hi $hUȤa =G">c`)Y09sH$LX,#Vx=hK. V0@PHx3jg>t89+8pBx}.Je{i?Tou~7`!֗s?l 3/(Fa[tPBwP՜Az/yU)0scv<g㜬.'0Ǐ;ڻc[+V'é`#9Е]2g- !r&̵('J"0? >sٙ#7-#D8XN w&rbVH8IxD9A}zB 0l!,=F{3*"]t6QlvY5BÒ8qYyw d!pYp ?.2)0$Hm'Վ2c`@RDDp+ịUh4BkM̨3hQ9b0XJBK~GG9WJh4Mnwv}ޏ1G\mS,β'RH,c釓ń$A4ǤQ\*>(49.!zj=rEP5S(mJsZg㒥dO-c;_h|$t~IB?1IQ${~u'uα >:yOSaL6>ȹr?Nc Jc1Fd6([8Soʅݵ 3gP^hftsʀ a,ا)($\bL[X0B1TpY4a#A2sVk/;0[,} *2j"@šá5P8sN!7G!x4ޓ?Fء{BGZ2DykkL.Y>D85Y *ʸ*A.\ wL:a)r%LN+EڈK2%'D{]GCJ9ō0Zr;4I(F[ @DmD,@g.,=7ȝӤ)ga{Y`Ƨn#.6.^ b:z?xU"IyO|!|-^00i/` lV 6L]< y%±w aY!7- { ϵbA2ky+5  ʅVؤ&[X`;#"!_8$/.\ V\=sqwȘβF`:"\1䰰Rv0X$IJ`S&y*L{ %kbi1g(a i͋xh{9L7B Օ{e/~nܞsJ }g{/ԤJn;?Ɩ(AH)hW%23@f ^" IjAsJyI^Fqla֯E{{_-Q- pQRklRfA/bg)VQPY69miϽ˻A^RȃULzٸ؃)N>(M ur]jB 6-Y/?wu (11"w|I>}v}SnuYR=¤؏l v!^j(^c)FYG mxMY'Pp$Q]]rJ!cзwPlu.;/2>ݖ9UlR \G^! !M`^YFua'!rƝ`= oLG& q$AcJ]qfEJ| ;Kk,mh$j i8~K׆Ҕ~wǛjڤB'%WS 6PovxǂkXL(=%IJQ _4Gd|ΡxՋ<7~fb컷A}_V=~y:e g޻ry߾ g/Ï&?_VOA5^<;M@ٗpх}B0L_rN3<콛 wEV43ާV?澼05֏eh=̪ЏG}PΜuAl1qL‡h@Yefa%(h'MP|އIysVa <+;^a%-?IM4m6h`Rgl4ѠW?{볬?Φ_j2_#[hyT=^SP,= ך|AR:WșA(n~wl& Gf7kxTU]Vz r ȝ y]-kZV yky'2^U|Tq MN6ggCղ.B ;&Țd՘!jL5&tQ쇁2z E.B̸3GEYË$rdU.FR}I!8{w@/9 Ȏ>{"&*{M<"aDVQ$|N^,1]CKIë^n],/ FDQHPٕڟy0K>KbJj}j1CrFXn6~IW7bi o76ĥ̃34uJ _G(ƥj FH6۳Q e(m+@sJXA7 y$/Gd f [^ diiA $.gR\)rb4{]CH 9TʔdS2MKg >饋On(E_UX 9sxX=^L|zC]Ӕq0e~R^ha7$:6 lKyu{iMѫ%v~ҩlmUχd4=\=5UkbɆ:/]OkX.T1I r 2 PUj3{C4뢬AjLmRUՁz?jڼ^no"nQB@T{tY!b1 6Ӓ+K lViu6|+6*hK-}gd>5!&!{r+.J.AFJ M )(@$(XӊDzJ'n(1үx-zv1_%1ؾ_ƿMʏqთU8fT ]5(Y)0lgi||d w_} ĵ &uD9hkcԃƺJ֜?cḾmFځV2W7 ^:h_3]y60,JȒqXXa oQ>|%4\ϲ˿A. orn> OhXn5]on|?ԷQ6NA$ӿ k!ڂYmYvH랐 ֑^?/|9mNi`P5WLb8S; EƺLj5hi?449]ihdJGQؠ)6))b@JK)Ft&!D!e!x X1c2b=6M +t"גV^JcS52[R{UyYE<Gsɛc.Vnxjnïyz~8WS}jD>]])S|a]vMFFR"JDٳZRAy(Q9G. 6b[oha43L./uZ4]K 'lL1v:T"wjs4y0ܼ,R R{RnF/, eR:@gT G<_øwgvO?`(J9AsНFtH% jG*zU;؟@~]C4r҈dRLe0F$ *rHH$K-P@ց*Lm-{P$G36RF ҔD)XecҠB0Ѱt. FMG(M5DCѲ+S$ĂLPsGcEm֝.Rmɶdj,.X^{_S|abŘHIe^@! 4Q΁-ZF\\Dk!{ 4ْsQ8:ȺU6=\rfJY]),+4Rc(5fel֝-c;6mmil ` wE8xI.i6n]>OWn2DӆbtZd$GpF!g04f1De= mvTaA B2䬌Mlg[ '^f̮n֝-v<5/Vtlkmcm= ػ$^iR%+)icV'FӐRYY`JlMaƆ DfIe[Ia6/"1&Bc{ج;Þ/}aL>EtE,j0ecqYa5Y-XEY-\0,ZD"scҨ/JڙcrHIZ1acج;p8Cd8]٧fR]\ob`xK̔=)LFp"D:Q)2kU  /t=\D=܁ kE^w9 ޏqQ{bVnEʄ01 Q| g )Q7=׭(_ 7Ht:1Yt䒙DjF:aUƃRRFz P pL6lz 0v2` loRx̃zaS6Ŝd=)"''FĜORFXSȶ%!/>@ #Spf5PdT`*I>Ci֝5V .<8XEE(osW %w9̓:$u1HDKD3`*B'.2=j JLMIbc>Z &o!EBm1hD 10 Sm.:dƪCR>PEh4z**ͤKbB鬚u~f_ :(Ӯ>iO?Y|&scU]qMc ]q:t:/4v(:A5b 6w[ Ŷ>"MzR;/x>6:J2YR1ʓ!Y##cDV)PCB#29aq26NSvm D@d'Ut>a .L&d/rkߥfyJ"V Ͼ=9I/=X)6.n?nMQz7͆r=%/|P]ƨd-m:ˆBY(D_A)G${2cH:_rZ#0mӕ^4ZJBbbr$z;+e@dy^tk(#Ko.RV%[Y"KU((>爆Ś59aѩ,lɡ+YY Y,?4I?E6ke,炰f]L>XWI*Й0i&00(R[E;F~cB __r!T 0? CL.AEvpқ٘Fgq~=jǥXpq?r/Oyw3ۉuGf5cӈ bK2!}UeDz[`-~並{=3{[dMyitL)pp.a6AN4?LglOhŽ%$eHBP TsaiuOI:'GUh2ݗҼ_B< }w|O,~}fvzfR #u7ۆfs~h%54 JژOmFoc~Ņw8Vy> Lx>_FYq&wLP -I-Mf̺Šir(:u%??-'zz̓qwAnjuXDw0:R,>=H/lʨrt6?Žjqחd~{NΏYϿrûC?<_%w~wˏGE|S$< x\3_4?~}eh9ń?]rR}Z{B{Vkk@Jti>-ӟztj2+>IpYO{\O|+^f7fիt{Cl6 POƝvB|'GzLn~17o()!DF*2"Z (aRF,TZt,0!1 @j JO㰴y]>qaSkk+!䗆u$,SI!4b T-QN).%LlnBӔE*;{:UtcՉ\50=Ejg8LƴHZ/-;W}_y׾;Q&]k$]d 1> Zr'(Q^nqCz eү#H) UF ^>c&RcQ١gU-.J=F{EbݵM1~<YwKŴ| ?@gzF ꭒ윜ϪM0yax>n78: "v1Dj0DDD?5iQуY0@bVZdJe0"&O$s`ƀ5EAc ޒqo7 ڧ(- lҝ,3n-}C:bzSѬ4@AM ;"E&$ (1( S |2*1C+l$t>D2,kA[R cu4,)4EA$c9)g%dEEFrT\-u!غPi܏X^G.+lph'-^$MOH?gD.:z&t$ *Ʋ]|gJEeP3!u\]H]tE2K8hȗ5|Y&i)8+)9GSX h[3-'55eVEa̗>KW!%cst 7 Jcƀ:G3w9AP6bҪ;s *f(xnxǭݾt.4??n*6{JxH)[K3g«jyۤ[ -qPggk "ss| KzY:I9ulE㏼:,:Y  ^j/7 Ռn՗@Fl=;$ h";FlCdF̹ep쇲!{X69?Cu}w^m:85OXS AYѪm-v(4qVM}tAR ҦJ\0*rF=29KFL5+DiAA (CjѬ;+ u-(sՎ5LYNl㵇!n-O_~MQs> ܵ%w@9XqY|z4` Ɋ7G5*<*(xIпsU1w5t]`PK 7 $psUpdћ"዁!^d6s- #rj~fNb?gեQ۵^/{ӱ|nzr]3]~H"hyw4ٻ޶n%Wj8 9,moq~HZvҴdˊ%K-9y-Gsgΐ3L<|wOp=ma@cY-nT|1od7RpZ|hQJ|Zl4,hd)% ;Mq]& Lgd H.s֍Z7BTLѲtstְm ܥnzEӚ'woUgtwPYlWٽ,3wFMe0K5R(L|Mٔ9P.Ϭ訦rJ-_)Q(3|`:跕p2dvqrmS"]Va"xZN/MkɎ3. 0)% HGLRSch0:`tot<@O ÎeAjtGg=dHwC`QVb1J|}2hjOh'Yګp H?os>uJoIؿO7` @J'Ly: TcR:ĴpFZ4Λ@JyM9ȝmדcNF[-O7~Lx9itdќLbVˇe9h| _Q;~ڵv+T & ^6*E l Q}chj`R)hX"d刨y\D }@( QG.5AzmSh׺^4PW8ВKƒ2R>uȁYd* (Y)f,@]#} 9J1$[DO1GKf$0fL {Yj1qj1d0 85a ElɧSBf. u>s[iÂXa,iCFsmM\DjbHIGȌ$n4Li\޻[0i}.Η ZI59'Ϥmfǯٞs?`cc!6;cc:Ϫ]Eo>R񾎡"$@BAFdaCb{ZOi3=YK3?UppdPI7*y8al4(hMNQODTű^ Ɩ,8%YA(.8ASr՚8k RlXǦS2: `bP\>ǕJ\d;36{=@=ض_<>kfmO=|VpryoWP~\wMv{{?Ζ /}~q^3,[~: VIalO _.W&/H xYitC/9l͘L*i lfgWW^^.\oå{6?x.69[-v1zHKvE껜|^H3}3C k{Z%.l1A*Yٚzg Kh~rP"{.JEۢK>Y{n=LT.ͤO,hyy1E,듊0ioQ{mM*Pty=0 xJQ.#x4"9.%&kV^FEC× >1p?\W1;{ێ$y=?Ʒ)6OwK:ÜT`zXΐgi'ÿt5|}i1|/^sBpBf~:Ä)kĸ뫖y<~fLl4> mLʗv}}^ B71pߨD㕐5@"l4|{}Yg'}s$x{Bow$~*߿T ]*A_j/UIPi̯'!@i Fn[O[I DD' ժBm$)4+_Tߺ=KS:`f] iё0Z8(k@(hSR&/FٟȘMt:М iNΒ>ha鿇O~ ^8R|30?atCƆWM S'S^u f2EOOeڜ)9-MmwBX"[ժtuTWPh%"JT JlR9lZULH 9hdS&,Q'=cK$,S{b~Lš0VӍUҧΨE]kiz]e_eTޙڼP:ed)DnM6A ΄ӱޅbv*~t~ymvF%Tr9l֠= 8KgeR`нsWŹ#L,̳`p^7Z=/QvDci˂YpsP\2#NLszb3lWn8],y͛/ N7R:b:7FI&ػs?w9ϣT8QFf}2?;Ɗwȥs kYWIs_ >\bDѠըua69`ʽerr~h\NdE0iEahM9DaVXQ+J1@.L?c5d&̰ `_F2SRj-=7g OZHMʐF&LY'3f^}|5/r$HB9M1XKQPPӛˢx䖦&0TMOMlTO6tTUQyabdd QpPFZa"aJIA'$A 8AY+Pc5P c&$ʒ K} rVޒArCʵ2ٌ|jXXM3BS M{/).AM>Nj4Y?M9bg.(g%ɶ`4C}rxbi]bM%LJ`2Q!Fi,Vh3v \%{)bʈ]M͈q!<ԮۢڶGރ]"s>FԠV#P@\\Z=9nv%Uì@sE2yPML&&V*C!Ǒ0^!fRP8 ]ǶzDqmfUh`)"U"EdIFP٬4 +dyVN. W;c]zh1cI+]6S8D.8b@P$I"@HZ(+ E@U!A@8S^g5-9i 8yӻq3 i9/`#̍ Z7^8h4O!*0ǾtdOiLݪ TylMiE(;sy޾jTءը/s ];9@k޾( 6'P cdN (^r!E .C#ɑk};D27HfCۇ0Op&82!dȽ5*=L c?RXd Cm Z˳Ö/ZkME-[==E [ɲj-6 %m ` k&Y[HQpjP.Ԛ )G|,kleMV٨rƐ?ae2M&fρi1i2=8/"wS_sw1Zs6{ydžl!+M(Ǵjg X*<~a?W̛+|{tH"Fv=&^#~=<֊pl&qE,"%+%`R?@.\*Gu  (orE,=8%8T@k21 ,-!Zst,N&N=?,5|Hr -TV87熬wYR.i'͊i|M%;tegDFs[.N&\F !j|Hɬa͑KBB3ֻ7ya7yIXeFeAXWQdrn"'CH"2ԃ*)4.T*SkGe<*,2d0c,FgJ8ȗI0R.x6ƙ85S$CR>}_u-X(D:{UV`p#G!\FRM0Ac._jm*$1֒&$r&BKM43`lHe>nNL>F|-\`Dp!"1[G #:b}֏-8q;aiUڣ||OD~̩YhF#J4;˅K&#wgMQ*i4'SB 7 φqeW x[0&U{'T1rNF\I?It') ˓7o-IqzB1']Pϸ =9T<|J}hC csRw>~F$U•~۬FN/GՌTZ2r:ONNoaWMM)C]LmB]qcPﮡTC9V_ivվ/^.ۃn|i?;hז]:(BR6N//`8Ye'{\ [ߍkY&7Ȉ!̫5p<_ut1$n޼Ïs $ >7}5A|"~JOh5lo;n1lTMJ_zq{ $?7^W/?{;}͏/P W`}NWn=7}V6]s ;tXY|~E)6H~a V>zc?#۟|t5*eЕ Ta6ec-?kn)!1sW6n*l6>I܌'G8$&@wK`.q"!!')9ZRZ+E4YJ}0yu>ÑJa\b: m7 $"QhJ* F8cč@8s,>ѻbQ.FyghnE{Α{.-nz.ݘtvbZ:ab>X8a?C~gY9G@ZgWTn?^hgV> h Ձ O$q>2# E3i=MofTJޔBxJ!kq)s4ݞJo}la=xrYM䊁hh8Z v瑠`  8_#ҁg9.vM͙5Ȩj"JńA].Qđ&-ZʐI0+&+. #.H'L,pB4OiTQx\q<bܧqlLU<(j|Cm;3POL/?xw5Fy<npPrǢR:v <ILѲ igRi6r:᭶&H]Ҟ-U1ߑ ɓ5 g;Yπ4su[<{yٔVpMfL_oWt#YIC[([W6㠷r&$cIжˆRԵѨfx 琝 ٙ y&~z4{DCH!J!8e.*^Z:kIt`I@u.Z("˷S|h9E 1㐭4$OK ɡ[ "ǥ,^zp C>uUyW|t>\g.>^U@)WRXڰ,Һ-- LPRZ6lcT|Nڑ@s@sKy $iP&, h YN\H)(VK.߅p܇,YSZG' 2 YN51$4OCF9 F Phe\+K(A bę!8dZ_p3u͂g0s%Lzkx-FASJѺvY*e}3㛓!R 8F94RDͪ3P@TTə1+^}ڵk@68|wR5Ϫ/h? ?̼D; =$hv'Sn/WUX-+b$ro~| c&smi0^ڢC1IQ] oO 9ݒ#jߚէ&,EV'ι'k5"^Ǎede"Bt<ꧥۥύz|ӝ: q/XfnU^7ۦ=A7YRKOο׵Pp✐5޶ ?#;F{>n ιq%, Y_/w^[ <+!n&;N@eV;m֪[9ͰĚt[d_\z ǻph&4έhTgWg^jD5S %KGD=쫀 \u\X&&b K0Fi*Hn̡0B3]1t5FBJρ.dς&.axI[ZP`s9TyF9wgiΖ3w#[z&?=,)YلӪ,PƃI0tT/i=_3,'Nϸʋ6չL۾4 nk/{y՟^]v6K,g< q#:;,ӭ=Nl}6y@eLS Q[ft-ht~)Y(}wy'|cՎ*GR4/Qcq#+6 ;V|Ni ?B\~3Đ*ˌAG_ﴝtJP,=;9vϢsKKqF`:ye L)[>X}Wlyb7X.W[%5E.~}(c( dr9Upۯ̽I=xGrزVC B\>J̶_iJ.-;̶(2 s[IT@ELAS,+m"HPNR1HL451sD Dy ؿ[D~R9& N*63nGMiGICf=5l[*9f[=1H`ZHY Z"H)H"K<ׅF%)@Pƒ|sBjlhM 685I*$ k2J#ޣ%@q(#[rtl/"nC8guLgŨrq:eh:ō5 !D/cwtzz)PR"hDj Jt4kurrXa18y8k"lԿ4^6:ri?抉|s֍ ~|$G.9jo]:ø=n( gTc֬?##d5+U7j6P>O}]/@s|VS夡lJlSvY0cŤӜ$:C FL2g5tۊ!ɮGwSWNenqްyL;\少ܙ}󘤎O6t:"gZ#ciKئ(g1JWFkt ;2 Is&LeJ=2?ܽQ׷^AF$|]W7lRXף|d^;Lni[v[Us@=M*SΕ=78&мE:k#(JX++,TQLAGu)ˁm6{s`3T,16qf!p<fT˵Π2sR).T>J1T08 YC9}"Ykc!/6L}`v[ 32oiWtsݕ+ՄE\9ߵXk8t"Ag>R2ڏV|A E 0} ]%h4K*!1Aw0vgº5c; ~-#DhN9E[< ET>Z` 2saY2"ĝH?39mDfHi:LSK)ۇ +H@ۊklZI>D_e D嘂 8˰`swPLwPE 1kw_|4G;V9("u) q!c'@4RĹ 3XsSlyz;L#1#r!X`h@"o \Q'ZC2Xj;eklDzK(Pҏ-5X*˛S/6q3`h4bhiah͸[n(}$m?.Z|ϯ$\??ݟtjU/aWlYv]T}{*^RSB,a&7e7n%D܍6M4mn!憒d̞$d (5 F%Ӏ1YZkU%["PڷaZ|>adM<6h\fOCp*+dLIƵѩ<`A3"t;lLpz9>bģJjew\Z*7F.ݺtb'hsR6뺋R粒&iB\9TSU@뫠@B#"<ϛG=z, S8ʗfm .GmT з掠G!Y3Y,sI3dՉVkn`C85r/<:įm5Yca*9q̪?}w&} V$5$>Kf$awO#AuN)貽Z&j. <]W^ĭVy8a:EniRH <f"käen"Fq_2\FSD\[~ÕO(~YISIE½ È8x\K\>^-u$=V+98j|󼍴ؖ*PUUNhS9epgL]Q]@LEK{ƋIǖJ l0Qha1a't:c4OR2=2,w)!EEf)9̉`6A݇L7PS}_hxj ׷(Ά ɧϫBƃu?Η ]Wua)_ۚȵMJQ}Эw]z`'GD`wj]QKjz}4ɆNm6rYwH-+jY͹unMoh 2yz~(u[\]qv}hKy:}᭻^`ɚ_46 B6u}賡>>ZʹC]jrlgHW?ƭ (b^TaEcpUAq7=vRȭDv!xIsc6F^*@@Ig: Bdzg!jDNEtb(B=E2G3MEa1C:JNB4ؼc<2/^Pl+4N4B9wS Ϡ2ȣJrddkϘYRF4)Ր4DmJ 04 @9nvGz5!$϶DD[;fz3E]9EpY(tlD*U08Z!T#.:gS61|}xxFӳԩFiצyl①=et:dT"qmӤDQex;ֱ/bAo;tGq?xR_,Q22'gQBq<#q&|LE!$6;m:oM }-7En? m"Amˮ6ٞȖRJ.<쵃7=Fd"B9)g$ٍh&ju 4gtjj1Z{N7mY .ܻLSi MN[U\яO7Ύ7GHywF'ԸY9ϳ828c3>_Xw;ޗ_k 6 U}ʃ9(r2 Iˋ"= #?h/}R$w_= .xäUV Bٻ޷+oK\ռqoQUN .荆E,{wb[QAϾL`;#f9q~9JՄaZLd.bOG!U7뉦is~/I"BMq_ƍ3!,?{ǍT^6yhU7{UzOWIǣ8s4RK1Gٌ㑧|JQ>|y^}O0G67@\uq~yuhPba|;G|աP5{}Y;W"08cBIIK)=w]rP2-Al,3? 4B'Yq8O#э_LNWӴhm;0S2 OLUHqJ;0[..Ok[&i^(1Xjz֨[ތWMPOk -ndީ(DԈHϦUsgY$ȧG ^ 'c`BU&m`|Elާ;ad J* |g1hS4MLBIEjsSz*!w;\76>?կ]R?s:*_Z3[QrFsYMYtg(5 XoY/;Z!}[@& i.f<5P}\3 PIB\"*D|)2K\ZtJ' AC@} +,J Ghx*阄99ZZ5Ktgc~yO+9;%,tWX {jmH uz A,"hw&ϔfeo² a!D>b/4YQ9D=j#Cɢ<&9V<2Yr5Urk\K]v%LYA6 S5ÔǮ`8ngqgZ`f` ា\eHoޑ[2F^wߌOb9~3}?HLݿmLZb +b +b +b +bb +b +b +b +=aEZXо7<'|Óox OYZ[|Ckx O7<'|Óox O7<'b9jEÓozÓox O7<`|ÓoxM14<'4@Óox O7g`tps9C\il夎q )C\fU1H'^_U_[lK^o綴Q0ƌRg31E$1gj;.>>`CqeAt!tӬS[&rήZWN.v?c#Zѕ,",0Q(bG#(:,)1vi*{ts+l|L QJ:sH{ .KAR N 4ɬT`8׼ngf`V{I ̼*q|/ɉLBbT^Eo )ʂZ5_ƻ$4]BbS5>4IHbSu z6OGzU3j_&FFR !6HL$h Ύ4{Q8`*YU5Fc3T$$[YrA' JsgH &a \.Pf5c5r׌ljgl ue].zlC$N54 W.x|q|:5vғi\*ZOLbA8,3NA杗& (Qf8{. &E3( h&of6-U&+kj<$p1wEkWMM͂}"s\ʑ+B7*XRQ@KKó#$Ǯ8R0+m\8L!/&LtMLT GQtUdd>EnWQǦ+kDli=74BiK)h ߚ *#P$:V\$AFź-rƴmֆ; "=I'͉" Cp]F`e : d;03 bG6oֽݩjڈa_33ԀkY3Z"5!wN$â>Y!DJFJ9^LP9:ȃ uA*RB(hx*阄7OV=ѲX;r_5}6ݙ*4PGH+?+p'Xq]O /VshYrI Lٙo>\i%Iܮv:J{7$j퉛,w&c#q#a0J sYޑ 49u`YŲOƟ =ə I:4W޴BF=~)<*}G=F7U졛jt2;= 'ĎO~x(o~{7V*Co.>~wж4bhBԜ|W՜wG-whOroHZۛI4£O4GyrkW|Z'qBdry*Uj*^8M *1v`}$8y=IBIj>K!g%9R|.N6? .v/2)vQqzhVe!Oqmy[5=L /R5ou)pjUu(hӷ;;^|<2sYRw^&uGh58 RqZ!;CⓎL{}\!GW8' ӑ-Vۄ7 [Q—g1X٢^C K+l 28˓6BL  %Kguz0.ͻq_ nv JbBc6+,[wecl7ᛩBn[rO/ M;7u!uR(N w^t,ӧׁyͮ<yn/dhs sk hD f0@vRh݁N1NiLuB={VlKEyRBT :\ViL#%H*-c9"Hށd eY4U-ĭxarjtx16ϟ{9wO0<(s b>\̮ymꩫ7n3[k&8$ͥ,ox xV33޵q+ٿ"sH  fk\"+,K$;]Þ˙d-[6 gFj2[&U q{?8ط(uX"=;) >mYjS1}mAjej&R/]5R(Q!N b$鐴#k̹=/P׶}ܣ lX+~K[]>)j^,3EopW xrn7G{rX_sw0$?{S70;ζa]|aKL_}f ~g_%{wlKPO57W(g~;}|m5\|_^i1G|楙cXmTܾ޶{G'_o^mJ͵^{8>ig|yy|#|WSlKE7Wϐs)MT3UemYlhH܃<6$G{/+Evx{{.exCf.b:b jLMb"ONTHQn=h:Y%C1Pg_=\=? T ox'd0y9U} ~I} 'd-e=чXYa=;f"쌏JSy fŶӫ7lcr52^.m/_ %:{es6/ܧ>;c^8EVfG],g-gc}b'!}o?.U*un.7MtV*K-MA(ΚzJ>9-h~*]]@IHs4PKVIQ+ R1XZԢU~X2RH- O9nE D+A^k!Nuq3 PWOė1 bȓչc6b)b]~  :::6.L+TRf"/ ܷU.(հꛟm2< w=^( ջH-@P$<Ad,~/b䰦 ㄼ[U] XҠ"DtpP(k4V Pe>:o eD5xZbe >x䓾~9[bRYd:cL;C8[h&Üv1x'TsK%#ٖg5a l8"k!0*##/-EuUIa@T_EB˥wS@yأyԄl _ ~=ԒPE"eLԴB)>K *5]㲱0G恾SA^X RKFYY tksx P< cx&cªHpU3᫈;*.T&K^V3I|ӗ56}{m\gw:8eMrd/ p`1;K$E UyH! 9z.p 7L ri6k)2{W.Dh @;,JuX au(PAG*ft0cm9pɠ=EȒ<\\5vE1pUL*P| <@ 3;#"L͛o009D>&~H526ӉBj%;Ɵ?Hcu_!]ud1cQH `fY:́ R)Դ[oEU"¦Nc-#Jw+ aQ6j؀$>oJsV*A1J-6ħecz ΧSk#mzs}g&Ib]7@7ל5FesnJ`$04[TNev*VMr.%IFh7nS^Fly6\D֫Œ5z*- A*:ͨT@=@u!%L C3' )7n0ց[:v= 's[QW TK\ \Dnd9Uy9~0U 3I;tZ1"/d$fj#0r PuMc-0 `CL.yM6jmfDnRZĪUi H>A ɘ+U.Us!]$A/u~5(n/o:OHxs9$*awU|d1*W R7 :LZ81E=J1}_H 0!#E ljzUFˏ>+ERUѼ\ʱ/sN14ۯ벧ɭ,.P,vQ,*\2輫 \S/, zG}@FHѳW`j?&lzcЯQtttv܄9(V?Ee"5jsݯ^Ɏ^ɀQND-NRs/%n{?'{|?z/*];׮{`y::~=>t@?+O4t@C4t@C4t@C4t@C4t@C4t@C4t@C4t@C4t@C4t@C4t@C4t@C4t@C4t@C4t@C4t@C4t@C4t@C4t@C4t@C4t@C4t@C4t@C' TѰMQT?}붗t*l*Z#4sVYIzl'*;䚻m9QM.MkvWڡc/u$Lmy;"V.8<}1x۹j+;qfJiOӛG?;^r-# dRKJF$Zxr\aR g}\]#*ؒ X,aؘ0&RGVӛQVioo`Ճޗބ^Uvo0*f*8SKyDe, o ]wMJ2Ͳsymp:iCD=SNyiV%:[ɖn"!`{UTiMFŇF^݄I_kQMRI(B5"Ap0%gtjɺ6M>ɖUx@)u}Ͱi|'΋op^GNNdlmi=@T.qF]σ?UoP}RE'Q--J}:հX]&O;#1GhDT7_]7sy%,az~8)HىԾa?s2H8[$}xmq.`KEXғɥf-=9-g&>E0;7:#ܫI{a8 x'5g5=h;GQ* YeeoT ZMw$h*I?E݌czw9eci̲F8#Îv-<dz#\qdbE&'Nn2EYZD/^ypXq4qnj'Ee %I}:z=f0p؁ |PE.iiFL~lĒ1d+-q tjOU(ujSSрmP% S ts^_߷TQⷞ5Uei`UU+xv>pĒ>rڏvެŰ{ő[ ľ 9SUHٲfse-{s<󳳨 RQ!g)L:PXiP:N 3''}qKt* Dpӡ?3CPAΰwD#[_|v0k騪`]pHd:ɐu.fݴ }ec\).y;?z:oꜛp?|]+Mfnf]d799󬄧`Y/'0эλqD;ɱ댗C \yN:$W t0X2l:9-[k,o+;f =mp7urdegВ3x3e7| :4d,GVn]B\*m۹sqC}rm1J+f u1l] u`1y"լ{E #O~Q[3ؗbU{"gUfkkCT#t;Ugp=E2ը]:VKtڡPe,sU1ܿO+ \93`v꛹y]zCaI@;z";1"gMgWunL>FﻭaqЀ[T`n6@etH]J3z"HխC Oi|n͠}S)QdRJ.JBj ߳a{n܂|!jK"~S6:'Vcd2=xRECI$IJs ,3L2lb;qmlq'5q6b>Ma U(P &XSL)Ӧ5T':Ë= 50+9肳FGMB86tL)Ca,kPۨӢ&fC WfxXpso?>>WH˂Ϣ' /ZnD%b:]5oX]Irwo{2waZ oJ_+zjpG7L9G0ӎzX@$@QO-ǖ`ELJ⸫Zmel3&fViBѲ/>/J6%dKrs[Nknt`{ CQ 4Qp/`T Jp1A1yTUM#B 'ZKg Kp$^Fґ0N*-{ms!xִc_-[2{`i(>irмX}6ke-e` 0GM,!XK6x͝V;|zܓ:=~t{`s cV+_҄0H廮&m`r!RTZLXˣaIv[sE-{#E {d7EPNTrGiV*%Ӓ8pDQ4H<.UXgQ1G EέRȰ, ^"-NGH/B">v5q66傇#G}L;po;w@y|d4S@9W-9dbӿ5662gʨ+Ԉa Dl>D2SGbTKqA[s&өa^a'9 CD"Ԅ$c EqT@b&y$ ViǬQFYlA@Y/5^#vnMkl-= ~e/h88Iq%,)5HHG2ۛSJV7۽o]àGL_,j!!(8 Qi0VG'aoG_kP2ivR :\(]w4H;6<7^NqCxFǤ"9f^CDXEQI8&6g!ܮX+(UB 8~=(c詍uwȱgHLZE(*t)bTIO 3Cmb`YDSAkiƐ4 ),3RI#="jm/Tq)>086ِ5}C G2^;OʔC?#!->nh˳ [/sW'tCѳSGNC)Dc ;t}w*m0,zF#m on/jXx l&'_5)5Bxڵ~sku/gP& }?]45kM{sz|bF|T۹nƖmUurŵ 3Њp%ƗtW63,/}zLV^8MgsruZꤓZ]WM m⇉9Ϗ~];n4|Oqkv1x)G^*@'^7/TunAy?"~~//1Q|sXuN^%BSA"02_whZ47i*Eӄo.o+ڽo>\̖:Zo_GW>NLWejBTI+~!tǻRQ-T*wkCV%֥%'cp _FHώ_'GKW'Q!걈H8S@),8ψBP1 0bh>啵|:M?9Kh0c44< $Bc %T (8]gCaJ{19iDJ3΁'L?*Lع锻VԝIIG+8OiX: Œ9K , Vȿs+xkϵڥc`iK?*Frh O=cF@̇b"iG4nURkPnRx^ yjڇ͟؋bC7y+|zO&{9gaҶAi0.Ymg"aN(EKf-L"bC҃J+ߧej?gߐr䑊1wH !`ol0aH_72p)[>eeBD~۳.x9ͼn>o5x_a})@F44Yk8|牠OP1V9n\T b)˨.w%mK4lH}\ݼl>%)!)}3EqHZJBp\ 2mJIH7T,gk8V5ĽA=29m=EtعǓeE.mY7B7|(lަZK{LJ*cw\Gy+1JhTJgNqg@x)yhBI&$IsmOQF64 Kٙ ٙhBJ/8̞ Rh2r&*^ZZur.Ae&Yjѹ9hq|a<5Gָ-@FPdEm&\.?snD/]l8ӦZ>Ã{JC7kZWflت:KiUd*Ѻ,-bliAXnH*H '9|&Sda(+<"{'ZWds$ DAͣ⠮Y28&"FƦ uEI$ZD9BY=yrMҸÕjoJ_M˥4$X`g9Ƅ{"Y5 1Pb9%9)3P} >EMse:PFWsiQMa{F@ȶp)==BM@E1] P?/sAs77^ {Pa sJ ¹.MtTKkɆHP1&%H X []_-GԀ(I';YKxƓ ԆOPvE3oG&S[ f#B@)д fbVb cu+ {[J #sl ij +}7`x29=0YhƖ.-Kr06FrDe5YaЕ ҂߰1?m_Bl݄al/p C_[͒wa1,}VqdBssW9:O_;lRb0FVSX905"`R" E‘`~578kU@V-':YWu||] XSl [YgD hI@ ipDjhv-{Z#g|Uyӌa }9 UVH= `e !miGl'jZГ -W11L]V\a]ɞDr«ZoUNx]xU=A|("{ y6p9`1cVh৽4cT<]1G)˰ 8 V8Iy4h1րUcVZ 6VGα^ RPit`C^xX -3Hk|)͢,sU6g1Xʒ&1*O;Dc}0a7WOHg*u䴫E:3(2Ɵf.m7NBoo8&p~\>??|e9Zк|ܺ= pÉ86k_VޟtJ=0m 8,SʂX&| cbs8Je#( %[$B-Zq':!( 1/68`K FCQ),kɃ>v[f$z 9nv)&vNsw{+eOo(+< 6,xqgem}R4"0$jJ:z`%7"Z IySˉ_Mfg=b G$y)1bZ9^+\`:1b.i͔u *> L{05zN>G#qds;|~]M1.OPyO3?,NbU~A5K%t؂L )+M^xfH}8z= wntNҢ￿{8dLT#*8K9Ga&P%:J>zL_|ד㪪ƨ*q6aFX` - 2۩XU2*u^ fVt=c CE,eQ.q-;/sT Ѧ`~cQۋh` ˿i4έ?Ҹq$B馑Td y~/ Ք_(u]Ŭ$ƮWUx1c!Ux߸ -mVa `~ˏtItb]Vނ6R FX*;8$ vG%@_֚֞/0<^QlfSf%Y]}EfFFOA*$y5 9bvMMCQ" ɤ04eTʤ~hd?{<K2 g'Gm/'VSöuP9%xkUiQ {M sr7_vr}hKlh,q^tz7>p8p}qKA>^~r|Xygu~Z37߰_<dħbgU=;%+tKEͱGhrŜ]O5>td}IE0Q }]*wq\8k^yC߻dxONvˣNxw%:;$#RZuoi<8{iXG?/zr" حE@:Sn7L;mox4>ʟo?Px\.֓hi<Vg[.`-P[vJҟx?^owߝ:ʄ^ Ljĵ}uqj֭9tMM&>వI`V \Zys\薉Ye:ün}um]̓ ,Mty+{7F~k;sZߪ^hdQIv @&meY2jKޏ?>lv%>Qx uf}Y{Z_-VHߝ:fgc R_KѪe"fK[r?FVym)ddF4A o/b"+ʼDʘm2ʸE~,[-~Q4R\ٟoõRR(OvNkx~ǵ֢kA}9/߾?LRk䒔fᨨf*LF'()S4t-I+s>Ԅb &3)e%lFZ+ j1SՇS]&L{}}&:$%LE:"l4B&UBE%Sj*D&b~<&*ciͻCsEïB,a-.FL7(z;The+uH(R4֧D!۪ ĹL.Vk Ϫ,U%X3F"ΉT,mtY ׂdamo ~x4 yQƪ)Aesi_!Gu?iY)k*K\0s;s˯5IZg͋Y=⢗}HJ4ap1(պ1jD䒳"ac4V@fY)dSH! QcIҕRlޑ/^q%l D9%[0݊ NVh 笀<6\jvVi። eXF*xD68(j\d4$ÜXkC@z=5(!Ȯ$MIBnTzCkNq2FL[Q9 ! x&TI44QD&+eB>7P$褱#v n>TK*TzC%ǀ:32*QLkDL75d&/bڛ(SѝEpi7i,Y{CRl eU* R+~2th $#(be~%d0OA_u$$EDjyR"8J(46B je| A$zuH*(_CU+cᛯyk&N ԙ:nq|/G6n{1/E5 HNW!*SN.0D!07fͰmgpq4#'[Uд-3Xw!4%MQ 3 bZ:pĥ8 >o:@G*$]45Z. p1%gZ7mq(hVcQB'!$RFVWe2NOjL%17GDC%!YZ{[{#fH܆`m ]ťEUfUSY_<{U|ΈJ!m+ɒ3.$Sƪa|}]6ϊռȓhOYs`I%̾ ]{ #B|u)]LLAjy1k"!p9X]%\? @1Bik ) zc 2{(%g&AbDEKI(D׎no=KPa2S6ѦicA[{x j1kooӮ㺍W&tlhD0u 4tLCGs ncJ`aM˿4ENiVnY{sDq%RV ]"j31if0QDϟM rVZTHpPaRD6HmJ F EӳNJdt AJ`PT Y )A)7 !n0֠7bhlAT(c "($+TC(M ᪝,Yyaè=P(ˢ#HMAՋ9aushp?{.]ڈr!cS;SvV댙cZIf=P(-Ag/ߒNFR!WT ZWY[EpStrgAVA%Cbf1TBmEѫxڠFDO+ (\u"(I&EzQBb)"H4֞s<zcY}6FJR`eSFvL!jɎe IB q:ts't\&!K#ָr^G~wECPQ!"`B0 ՠl۠Vܴ\̡ Ð0j |3R-gQ|p;Xr jaH m)KެW8_aܙO;4пK86pmg91p@4giA|Qf9{zD}9 ~Ȉ1 F|;Ч k>E0s@1s@1s@1s@1s@1s@1s@1s@1s@1s@}y!p@0}= $1-r@NHb9 怘b9 怘b9 怘b9 怘b9 怘b9 怘b9 怘b9 怘b9 怘b9 怘b9 怘b[0W|d_ k5}A8<b9 怘b9 怘b9 怘b9 怘b9 怘b9 怘b9 怘b9 怘b9 怘b9 怘b9 怘b[o5q@Aꯇ|L('S 9H]Z覂 ,bF;g#+c^2qry5# >xwX/x>vn=5(E=[>g_1Ӭ2KJ7rZo깇V;Klq4 U<.ibXRΏw;n>ԞS",j(r<O܂ ":F5@ZB,~8ݣ]]Aym zn J-[!<:XY6 b&eo) Bdk(HX7{G85r}%>nBQn ~02C*f tjyT/o4e&QO& j/?^~483244U@Z!8CC^}^L;yA~܎iΡJB" ɤA;~;J{l_nnn,A:xSR/ve0s}NπX~\Ї|[xg-׼x 7 vQΏ6Ͱe[8܁Wޡ??Owx}&yϋcL'{Tpb+iY [fv_a+D(oݽwּ1P\wz}dݛkG_]Go^]lǞ\^;Ynyt^ij/Ɨ- G+Ѻg/txW{́ }y7\cL_ׅ@z*fL}\c918d1]Z?}j=SdtWPw[K *)% lu019A޵cٿRX,n4E>d3= L 0Ag,XԒeaVYLrR XU"{yyy.)o$ep[J͞w#A3/y/RO).]AG|g~]4,W2l#3@RvЕ>Q,/\۝IRfK9x8+@hbB5:qfkI S[4"tfj>>6G?~teI{=L22Mr))+JzM!ˀ3+K0jfȡFَW'7 x+@Y^{ jLv/݃3NvÏ,lo_|Ԉ|.>!YG(tye2AZ#)ZtH5/r 1+ 8gxMOp~MsU+وb|g7LCoX|4e99P3_ja1V?X=gzWPX6@;&?=a_v:}C5&+/j ZF׀ԃCi RPQ/|fc~ЛL~ E>xV{`Em$1u0/D1/b_XKitDlx*fyd`A!b\$] ނjSљgUg]M P 89x9aHo4N -GKB3-<+,+f%hZB3vCrz4ktqsG7V>I?Ycd݊JqQCn[hռ!]Ivw^_{6w&z-wʫɢk?Rz틉y2 ^R%A'T $f)W4YnVi$vrN$iA2"C YyG&D%L{}.$Sh4Pk bRʎXH@gʩyL䌀YrG(hQX"WAX:;G|⤞1u%{=EbY EQV/zQJI9! _<_<Xw?\Bx`|`s3Fn |7܉8 U|`|%xN;w)Q*Bo|2ٝsɄ 6a@@,Ι5Ȩj"Jńf 6 La{؁ K0H#Y9J@)TqΛ|INF4 (>3W׋@#h8n?'y. ||vvbz©Ό ]4(o%$PFg$Q2FZ7T$5^Xj#jU K3J1&V礑Jl8+p8M-^}>%͜b0yE,Yj7,,`dB;sn}sJX}Ov<ښxec;x=*0.R!)qJk (s^TVNuh`'<jq!(2Va=76p9lWE >1cJ*Si9bَ=Ck+p@TG<5,Fw 'v뙘}A灘(/^,l,#5V>x[' Xm(MbWt`pPLb昡؞;(x^uiޙگOА!^ t EECDhc1*i8$\ISo}D: "U6pguNAE!2ai.(CS0$)Vhү W|CJS|P"w٢[^˺:[,)1$U"8 \3K`\ x$LYbs8c+}Ծ5ɢwt2L-eFzR'J3I` 2" -"Z4(Fʁώr_jMŽۤn)xrƺK$hI2!8ح"&H4D5蔴n"tYNav1$hCI\%ʪ LXL҉k)g"DDDc;C*kHDE+opb+kǟ'3ݟ1ʐi5n"G$wi>㬿ec߿sٟ?հ8̹4頛|z[vS8w<3ənj$yR0J('tDN?(lS;K׿9L ^ƾkIYG'$:P([k@-nAz}39>86|l maI/pomɼe_/>xWΦC5^-ɱA~7=R|q NJ%vJ5=ԻjWW~l~9jz>?vݯFsmK7HއYrs۶+$\M߷vl~BZJ ZK˶bX{1BY~CC q0l|`n<ECOn]=j8F.um=o_$ODžvM T3zJlÛ M%nNW7:x~{ۿcۛ޿o~z 4V]_WǴ/_P)U4(kءhJ>w)W|rCbCh{~96ꗫOo/ϒ7 RͪG^l52z Vr<7MzVH ш!m~̫Poʭ>ȑjҷ$8.8O"qCb4p$N '$sX¥/" fyǓzaz>Ñ+ %0.1l6/ $"Q$ z#T⋍@8s,,Cѻn]Ⱥ$NFDzXk|Ŋ*,﷗;G\Tn ع9vgBw't=Kda2EFh>4v^0p!?)B#T'vajbY؂9u@udkX/7]Gz lġ_#`4HMd3Bym HwBfZTqeED@]u @D @8e z BA( yK 'XN+-Wcuh;$bBQmJ8}@EQPZ5 "]T7 t Y] [lh '-A-ztq:v@p} m;tLHgYϢ73(ok%DK(!E]m}4 Ktƥ;Du2QG(%s()' -^ZdVʨ2'L 32iv1Z3h$ɸָDpXgF0tR"W [|\y;q+^ Cg뇓<ip.S(M[ѩ3Ut (^~+[ccz͂oFYFF@6`hi FN\~ё@MG dNdk;ݫ.lC2'gIFCTpBpʐBMEI%ZĠl.[fxdXLO+k}zG$ ,x6'&Ih J8++ C (IE8g҇ kh`$K)3C} !hHIaڈQ8K6Ug/Q|}ƜcSއV;^h Q||VϜWut@oPվ:>ѡwU8dR?M`8Zj2./rOդ㨢AQ?8Cr9co}!&^j(gXƩ k%-e#]$z  N~I%Q}!)5 vs@"So|xŷ ᙶH?+uĩ !lAH3p#Om$^$B6C'` p ‰pX q$AcJ]qfEJ0/ Au0V`Cm'KM eLJKz9MޝVkH\dJP\M` dv;(z:IUGaiac`ëZ\4E=CH(]f\\1F=e(דѧ)t^3bX}=EOԿQ7!{<"*{M<"aDVQ$|"{9$iҋE :ih ECӇ^7דQ ԥ#Y%:KQA:Z,ꦸsX- +LOXȭ {C4,+2C Î'3$?7C_KP %.q ,\x[FEjԥ*,WMqa9_ƍ3AR5#n%d}I0^^ WFg1k ]pn3$J4o~jٗf %MwڅӰAX3)]c9C` E1U<¿+"%atrV` JRٔL*d'3>`ׯWor'%B(2d ).*kZ2NOKf@qOifeLf;7Cܔ٨`P`|؟MLUL Uk~ѡpe̛}wO4XoWL\&}{C4ۛrAV#sc&jZLB_]{]̒c ц'~(rrO0+@,RmJRp'Hz'a"0eURO&x^>NQhlmMy%7R$P6g:wr̥u\(oΆ0}7=L%raga5i7g:Rjsb!<=1 dp>v,9ڍt$&Iڿ}'kC C1F: lʰA2 <906KjUQ+T&uZ'uuqS=w/eM1 ٧0f/DTzYv0ɥ߮MKXÏѧaƑh_Ay-> ^@Cv 9LZ~z]%mn j֚ԵNNN05/Ϗ]Ni`PI^tԞIO'M-4pR/Q~7~iisp$#Ȕ)ASl4 Sn@\:ڪsP`B<UX-zg՘IDk1hY!-iۛ9;z;n݂m{7䚤. ʎݰoC7>Km3Dg4IN0;O'GӉO9.x||61>jY  FuMESLƖۄ.`t>DۿchNjZ avŶ0&S)h>)tiK5zIe<.>/Cy4_O׹/&Y xel9Ty~MEIVӑJ¯aK^VVݽ` 2YY/ٍVBizodU@?S>>ގ]DNk)ڀϸѾÀ[` hdTq}-vzt+trȍ-R$IvJ* d.&r3Xwy عun4Fļy06WnV~=mRr-/΍%1Kw֬ok@3aW|q4q>1<7+꼁mzɎ@cp"!M,Ac>K̔huOtGђ wz0Crί:={cݽH@hA #HW%>M9;$Oq9}.C*Lc#8'}G_k{􅚩2Hxyg4 85! !M`^Y.EːN"!jw A0D8x`::5r-$hL R" H D!+n1Up?^ʯP Tp.xi̱2.JL!S"8(B3*8`%8<Q:b(',xFٹDO$Ϗ&yT{H Dp^"i CH(0˯waV|.k3..{TI:g/1,߾(ETԿQ7!{<"*{M<"aDVQ$|"{9$iҋE :ih ECӇ^7ד2~9´pR瑈=PNļ)yY K:aFU Aȫ0KRg:66.]FU=EF{Hacf+b R pV24?l.`>&|6:}OJFl\w엤qLT W˼s^_t*2zϵNrGxDނV}wV w| t]?z3y3&/Ǎ>Q~rR3%Rh)Aa7]<(-+)utcj “W]z~uPt9/`=?_\cN حnפX-|,*O ZNe+ U+!J϶+=[Μex|x%֬n{_30J'EX2Ȕ1:ҖRq 8B _i>3#<m5p ~h9-6VToLE)//C r0٠PMmi帛n66J/}=K?Fm[\ J۷֓gJiΒm_{)kv.>DZQe:+YTHN>1BwnQ97}xmq39̮RL1H/R+;eB wMċen ļ27"e_.>}j5.G)Lqds2J (1zuZ5+o]}]=Yfq\okcP kmo|bByZ4^JD!t2tDi좵uv^<ۨ+4KyfIq'/PAV1RA0R8za6L=z&H]W i>j;?~5P2HS.z&;,:g)KϬR[VNGń~ #4"RQ&A҂H@P(EsFҷGtr˪$Q83>2ZyiJ1PQJdu֠I(4gFA3Aրbu_){2̬$J3CiNGcEm&/`X/nǦfd[&'MK +Ŕ^ȚV$X1FJ$NG͝?p2-vBFh/F'b"Cf"1"EA5ʪ6"WTː2 %*uet^f̶n&;/VjScM읅!4$+1jc#j\ xH͓ebWR=,ڐj̐ʚj"LlkR52l4 FRID&ՙols=oW5bm-mlmo{x's#m,4HČXHeЬ伨; HG:M27tŐd:P )Dd#Y"#3i@R$F=w#7n6vJaкle ƒБ=^dF^{8YJUT_d/o1*e,C|W+N^Qq[b'gCPh}Ɂ1*t %}`^| 91&M-zK‡󞏠;>ɬL.@Sg0mk9U<X* FCsBcR&Y YVӨ-)Wve3lh1nVFl&QzD 1d9R2IɅXHLX>R&D-h"s")$8T[oR:A;-0%9Z7灜aeM&ܨuJ*ɐ Y➟|o/Xo~u*r}u=wj{kj'-gZChrbu UJ4"2֦pT2@,!vc^ZmV@&V%"Yё YޡΒ=+B6JW 4' VgqWذcZ,lTE˄ŠR LseVa(X-R,!%'2km~?Er6uZ/S"WlQVG@^h_d |Hmi ƄW BgSB+i]ZFoRګF#76,ZiwͲr.(Jvp9K[u%8Ogk -X E2q ٚxVy5U|)UrkR?珳 ɢg\~[ofWuOր8&AW#_4"'\5"vsIuQRiq`~y]^㯳e<;rz8*8-euϧḌ'UK^Y 6r>UGbÓ$'^M͏zc7)f6{bƣOˎ|yt8qq ~겑N&ua=/u>XŬ~WԕCy g[vo]75<&g~x~z'<_޼~y_^^{+u=0ͱ$G >zxkh~ вՋsu&oxm˔8#ø[g T-Zqpkѧlףn|5+zi+,6-םLܔJ9cVR43!n fpcܛr3OE$%$ hQ!% bb"eu(*GfUϓzai|E=˥B[|2>i/ 颳 fB'yif k<4٫Vqж;{TsՉg ]wum]>6Q^!,2C"Ou9nӣKƚKQ.6mmԳ8^neMz .>ԯ͌b &*)7+5,:Z|q_b TXV8Vw~g7qub}yPmﮝߗ*nz\[ϧuM:?~ax>ilJctkU^.(L]1Sc',lHJ$l $uF,p9@e &jNi-A)r>'QTȹ,kf eZ @dHҺk3q_iً76Άw{]=MO^\E\ӕo]N38~@hM>,(c8Q|cs!7&rc:4:Q.<0,5v3mc; Jf61o&2Fz Q aU`Eqg+YY3 3+Puu|iDf<:Y:S#P)d ,=}@>g n,h l\%x S 'rFilSd6Ho-L0o}g[Ws|oa|cZ;/Y%' d Ue52Ydd㬚 gRmj%ebJH,A D$RlV+0HF GckLx T (sՎ+ 7~=o'eɋVu2#-=w}&Igki67o+>eEsϑ['Iw6(AFIs7;C^GGm)w+>fN8u%)M$@VܽN՟Ȕ ^7(HQ 8r5s:rI)[e ¢ѐuM+qcRu?JAHL)!CH>hR)>69Ҵ9?IvN aAi`x0Oxv:ƫ `d2~wmY';Rn25x`(bSԒ)GK/%~rDe YB+[ \e`1` ]\ϊu;sb,e"O`O/<[j$ψa)`#D,G mZHMLw!fw H,%eh%0H9H܃ JI, p#2kRn9TG گvd}NT:Wŀ5*Y &аuFt9|:3u]hc|Nx郕%oګ_#p \sYrf|}wJm-n~?HDQB.R">Uy α(\6 ETls>ncƟ-8"˒ h)-@8h:9C{1I@"(pxh; DeZ1wAl"&$ygLX[Me#J0X)C^)v+V_HuYUndʏZ6P\ERgڋx^} \/fU)ڮ?ap|D5lr9 Gt yxȿͷ-`c)վ`;! w$ܧ ~9{2'a.cؐkSs~eXهԒlVyZeMON"FZpC^kjE4;BZ%DLї%7^$+v1PDEV+"_ǁ{=|8pGcƬjBVϞOѰvdHU!۩ybŜ{)ŘiX9bTνWV闉Ya_J=<̺Rk͟f5#^o~?&3 *(x0T`7WhB_\XyºE hޅU;N]V)}d8g"$pԆ`x[LꀽN*FdՔT6F)X;M 悷y$4J zYì@c à{BUH*Ny*8Cu0bs@RK d=?ax;8p'E`\é>¡yy_go|Ȕ TVf\g}bz:!7!@2Bp& XWf|K-:Eq2dOP=- A7~0 簥 D`ѻ3{3gpf 1Y fwO;-g Æ#;(c#Tf;?bƦ$]E<}ZnNjt2ls+mbzݛ\'Fſu}|7=_,WO>ʒ-ѐKc+uJk;y(ScS]w{ ߧ h]?.ۻtf{phf!,eQsn;=4,%FqC+-C._su{͘wyȻ%sOּob g[3 gwK c\%5i8(R'k Hi3yrAn_h8qnEP:y0E?@Y^do8]4Jq9ҿ]?_N-U69,aɗ0&:T6򠝏*/P^"TE&q%NjYcIN(<`td<0a$EKbj} . ͯZ2,s@ܬE˞t2@QךO`;.xu]so:Fizd 6D.byLyi \s#H,b!.]nܔ[q?w'pZ~Nv#Jԑ8br9|G7jsKt3y3}W8<~5'RvXotnUL_E҄b<\̟?u*}7C{F``.Xrh!5ޔQ%?/a$54RQM[wq^@Ny\\L)ZU}M-84N<tcƒU/8zMQ1wI}d'fydupqۅ ƚI8 S <*WXhA11O,F ?'@Ƈwfa<} g:D$*X Bkk1`%BFýb<_P|(gNs$ZQHQp̹ I;cHR.מV/pߛ˻8|-e׻Wz`.לII2̃AS)FY1y,0Jn7pZ:r=EZ!03ILXlnW%uw+\d,їYTe%+,Ioq)=5G >1*O;D,OY@C־Ct7a[(MF-4FA65KH U/S:(Rg/C1Sv 2S])}x渥zj[JsLoD9N)ǝ-xؒ7nY;FFZ!=KR U8khc6DL,h.*Nwu`sF0:ZqվBxڄ+PQ(e> ,! !t0ɥ߮mK_ͯոqd7yog# \hZu6:~rKZmպ mi}"%yu )oGq2/Wu#*ռ @YBcԣ/34!ePu7ruVHWy9+EK[NptLj+Nl/zP-"ܠDuٻ6r$4x(,reaa'(ھȒFzVdʖd__=>sQJzsBȪ% R FtVpy 8\3/%"H}&%1B*rR0GeE֝'E&Kqr>DɶjOuפW=)6VUH!FEhv fRx o#^krJ9V kҭ, ӃFʆ )!䚔+plunXT$c[[+B 2^c~rO8Lؙ `<@YdMWjŢ  >H!ŖVfX=!bRD AC2%ɷcexc-vmh.bIǶVTڦ=Q:z̉+U4Qёn\\=7o,9vűUYicyȄ yN ٚ05GQtUbj4Wպs=Fb"Vm-lmo{xfPh#mydSDL(q 3*-w,gL H6IؙB5ɓ`7vol[_} 8]RhSli9l]o{xsHa@?ZȉHc)xRPvqvq_a5{ȷkv/벛#>e'2O7H(GʹqG~\:5,]=BgiX겡~uVg3۽eV|^mLGަr|%y\ UxnG6@{JΘ'dԐ2m[@/ o5goT \1٦{m땝4吳R\>9g3_2y}'l,^ kPO* ``HJu MVi f~= >ico =RV#ۭR!Up:}VknFjGH*-S9"A2fe²d"ĝ%>rwj>Y7;gbHf24@pQVR"UmXM7$s%JZegSLI ) hN+/1"-ڐG% >⇼ͣ_1bTLf&OHo_˫>-~40Žu>cEⷯhj[6--lд5z1j>vT/Ρ=S )4|>L9%iYP[(-&yŧts=LMZ ulN~:m]zWiU -ًm~fʅ,lI84Bfȓd\; sM/I!_ש3x_*VU*Vgjeb[vugvVkv=tݹV;cJ9Qqd9C2]-)4+wBNVy"Bd daZbm@mحKhWGfbe:G\Xf,GAIG__eʗ -+.R[矇eܚnBz={$g0{vZeCqUxVXu6mР\YgAL%" 2s>9ìp 42UCb;#~l^AnSwq /]`ry^pVL}ޝc 9]rlK֓OsƄ[N&5TNѶ+ӵ8J+AG& hpsJADN }0%Ty%tUSp[6phEnijRH <e 2D֝ .p7\["tUw8Mzgn! %UÙiFeE|;!\2Z2v( ΃bKX}ّ%kE{4Ոn՗Td&d5+S~.Et.f~K>*d|7|{fhZ ўFBWTkԭ@R m(5/\H ޗiF9lmbmYsNGOO?4p:2)4ml>i}tڴf~8/c3&旆3IO?6SFt8nx}n.ǃk:iq]sғ~,?kOYJ.|$4]fٱZ}J3Z5=FV 6nB-*<'N,wgstxxt8ssgaR*P|返[{a0a¦/us׳A4]ni z]G$BlGRڎH6*Qk)[ ~ tpZNA.亝u[㽔(FL4ИZiX u?B?|lxl5KWr*l/|m1zAw𙇴.ߣۛl6 .C :G'afcXr7J FA>)+{Zc:>e#~cYWm`abtvʤ{" DR/ULtu5aI늉a>ݟ԰0Q:@1B7z΢ JM+IT>b uW;+\ioE *6FxIVEPzT ۧm~O+ョi?tSB 1n-R+5{)P<-6[L7dB1?ik8{x^EרEFSHwwauebxv8X5~=yu4[Ú>tIRcڹ;{V:vstFHo1hm:ʎ^Bǝ\:zXE=ܭ_~|Aje0+2#zH ! D!p.s?s_Þ~goxah_QniTdׇ^r7~*|Xm4?hHݧ+S3G{GMtz oa;L]ŽM@D(Km\hSIlm;>;xw3WI:/v{GOX1mm3:g$d3]\.8<¨!>u"+;҂8,Kﶟ]ݡ+0ܚ܎.OcKʄJ"Yz.蚬*K#Mfdq+2i尣}h=3awM[ݓVMt?m}ؘkS="DhZnEjU}dLj#mTHczS624-j^be*ZTvsUjEyRァ}c-8I+k51HqSZ[KCU![%kdDiL QR*S)mv>+7j-6\G.GޝBQR֖v[!nLVgz8YMԄ1V~n3:ŌiJ9 o춸ѹkZ,b/nh g6ѨVGNYV1+!ǹcFix9wL"ʿ)bw!0XUVJO%vw8gg+Z'A裃"p˨ ڼoߥ vE!մ-8cgDɘu!gkR|1>5 VUZj Z8'RIIbE1\2'%d >Fk*ÔDi;F7Eע$S&mI!a##+$V +_'LH/"YQYZ壝)v^k,֓tRO.E#OAQ ׄOQdO]Xw"R5V@`&Y)dWH! QSI׆RDȗBxBv36XAX3rVb{UPQAmxZXC:m\qLOJ$[PyU벫4 se±&6Vʫ']g,5':i4 Ņ5X -cwk.4u *a 5VX4:dZi7m3PѪ 5$#{*IUN0¤T)cscS1N[5mB7J V   ʄF-G$&.qJchRȠU %X;.j!ՠPwVrEALAAXG ( j "<*f-E"3|5gt ](∛tavD )_JJ uFR'T bXN:2]:Xi65lDC2GRF5U \IGH( ;:I?Ն?5WՕHދ0j (be+% 32-([jkIk ˨DpPZMil,@  ^ B9T2-Кe8M236֭= UDbMDr|T1&bb󢐴rbB6!dUs7þC5]xc#TYd :'JrJ2Ba1%OpHv9k=/3(39՜ $rAV 2*Xe(6!+2]7xѴq4/e YLZ[]hx $nGfp6ᥛEU'US Y_"|ΊfG6T' O7VEd?=ŤuԆHT—Pw@1 0 1B`@'SQA`,C)9iD*ZuIBvp 䁀:t kR([|Pcm!I5MHpZ{65^ PP2Б5;kڏl: CIeXYv wMϲ] ʉ6BkB Lttґp64IxTF+@T o-JRSҭ*^"a i ߤn$BG lӨ;=X-)K UK F^Lw+vG0,a͇۴Φ\GIôjD0u1vtrLS,K# J$v4Di Ú{)DzI(ΓDCakBm1&$2>6"|Ie1Eד.Jdr A2 5@JAO!Ƞ rwc-iÖ+QB|?aСOLm2X4kn3X"Oz7/V1Z8RTFjt0x֣CLz +~(]IڈJ5(Z54)hc+7 3E5iփ*M3| gݼLFALPha3\BvµuZ"O{ *U.Z'f֫YcTqڢF+ (|t* ((-l4EO(+Bb)3"Q%qmM곩\4bUf\"_%d,<¬$ T$1DЅKO0p$,X yۨOp]g358 N}TQ~ՋAXz=YvOyr~VO+]R[b?ftXчggДú;{Vs6쳜amC\,ۍKEzsYkVoCEPq[bw< 9?y6[]ojXL-SL n(x0ҫIeW1e׃~Ne/X/=<^,mYir'\I .`\,4Wܬ7n4;[0w+~F r-ELTRc2kB0UmL^b-FJ ._khdSIsxVQkc*(^ 'WEWu mblŬ(wb3yT́gIg9|`\lOY{bT=(W]L'S5n^a'9 CD"LC! x5 ;\ ޤacչZ' ve/hjq9L)%HHK5qh;l/"5٧۳( Jf{Z_Z ~BJc+.4cϐJe4"Q !JҜwD%ɤ61??Դ"i@RXbe AI#="jmCZ"UaJ謉81QUQ73yu C G2^;'e!N ? 2аS4Q>#t^$ݐ mҐ"(]F?I%NmM懶.8|,$7qc^*4 /CY;0K)\`zo׫GQuz(bY~8H>M K ED0.IhL]EV1#kΧiJ`Mb քģƥCRj&C:\_^$) q+Ɇ?ДdKг/g= ˏrB5LJw-(r%.3M&0W\F (UmlMӷe_˄Wepv2 m=ӏUշT/viDd\^ pٻ$7ԫ+ym1чQЃL>w?:Xu'GRg+3"-⇙TEϳ2gRYZ n* ;xTѶLĿ:*9s]yݏow.~{NjWo]`.^w?Yz`L *`rp~=xM ͊P4!MzỔ˛lrM7ՇŔX:7[KM}xoS?*FJlSϘ8P=!YY(<}-X Tnq}NK+oCVKUgcx/MaN ,[6triS}WccӤoZX 8r oI;h:GY!a^+MÛVbCOV:YS͔7^s^B ٽ y?BGz rڝ8I[. {p0EN93X3D>d>CpRx^atP-kJr䑊b91# D J/AכhO5rm =obU&mmx^dHZ.?RW?7rGh &mqHq3 fQY S /F !r&̵('J"?GtG8#ETn+#yc!,KSϮ'ɥ5BkTLV`WGB]'B OnOtOV"0ʕ2N XcP[Ɖh6ض{Uv:_ta=kKG?pB{Q|Cؿ!`*Y@>^%^ᯞ·]ndK!s9ji49`Qn0;"$S8Y8}r{=ٽ/wjWx_ϦVNts^^ )__ %YKcs4uK2PP2`EEUR6K.K:W7XD!+>2Um$f[\R b/G8fg5T:Lz_(ZЛ^xwbγ$,\/rc}:"t1!\DӴoapAc9` |REN7VFrLe2j=CWN+Q {[4ut.|l}\}ԗ<0htY4]5!El?R/+fS\84]t<{vMXc@j%9" Z9pA zlj[?ؼNȪ8. ~uXE%V1TGQ25 ޔ7ylY˽^XEߥEfilt(L&:@o1 JqԱԱQޟUh_}[;Rn@[sHǍGH#&сWw "D ^E3oYn ]p} "YbQ)9 $ڦ5ũF.`~Bj@@X:攲zM`Mxؕc ;̭j^ʗ^zw=#"8pArG 4cIVQg&D*|R4&+3eԡHNB?+[ ]86vQix`,(3xn2޵q,ٿ2Чd7#AH \G0)qM IYq)Ք(ؖ氦tUuUT{șBRӉ=U'jԝٓuvプ4ZOO|TčX&O[($\bh y]1T)b+t@HmOjRz^sXe3[J٢Irhp4{hK/qy<-ͱXGdif,k[&=>z#S4 N9 5 FRqJg%ѰԴ2K1$!"s+ms4^RE fTɶ˗m@Ӓp8 @ZR$2MP6Z$Q1WD!BCa%ʫ ܂k !>G/Am5U aAFyʭw[@2+H6q -`ZR9LL.xChNGmԜE\ k"eهv+W> hjp%'[oL=EB$he]Vz7=[no%7uo󑻤``0Ɛ[[fm89 K,-prz?.2SZae5,Tk"dbkũ/$bPbAaqr8w%rDz#@.PBh# Jcĵ#Fy䴰BR,xЙ x$ )[X1c2b=6M yBZ;֝M !^ȢcHC#SV:" N[p΍ bHR%(`-#X\qòwnط]~:~j N6j2>|@7t'^"U 1"mBh*ꘊ1qtg/(9%_;0fhNvZf\Ս (`MM&`d}tk,W]kLS0>j6> f^_ U^2o _?t3˽1Is1s)!dzJ?h9q*p8zlY{9{Ngl١k9i+0ɵPQ O~elONF9w\ڶOiHĬsIv$. kKR)mZȵM]sW_֬tjboOn5n?g_s8pŘkg]_v޻G޹r!bw_o{f ow}볏npK|C`Ǵ^s*X۟}})͘LC ::40%L6Nӑ@6/sfPpL׵pUXjxIy2)q/scN;Ȉ>cK=0KK ,8UXaR+z  N:88piW~A6x(|\?YZO7x9[o?6_4[@22HxyJi86Z4ye]F:nk "!v`(  p ‰%:x q$AcJ]qfEJ0_/!/0|4CmA9Gq;"<:K[};`xzD)\~K#~b"(qQn $AQ*93}3O}N)ڐGyQ*MU1IdG_%H4~CH( OIqϖ gFcKNUS1b8;1 oʯ P>`ORFY:pH&dvd/7ڼ62梍t]"U6X eF[01jOa|zB&\G!" }P!Hd2'sj~/a{wؿl~\.; cwWߦ=bմdn~,'}oWQ7?8]*\zu՛7_2j"ᵷ#RFhMtܑ@_W'oG)-:}>]bN{L8s)HDyB-Q tZ}^U꾕\k.LhxhX-8&+}ɟ04*>**IUhA}afAC: ZGy[tƴwTsv3<]w;Wطſy&&iS\0s/1Gt/9v֑R3|%gS4a?uRcdB+She6,#sF׃#ӱyL*8 ܞ)efd]2g'ş 6:Hb5TsV2b̎ZK|@L30f8,-o<I&uXy"<$$W0 ,3L2lvWBfsSt@be$3f%`]pԠ ^G9ƈL)Ca,p ia $r~ J" MroȬ٬l0^&X9Ħld&~ Mqr8A~ەP1UfDT "$2Z`1"_$LُXS86s?ohXaG!Dbkf9 Ѫ8Xsїґ==dV_{D8YJCd~|:{ݝ$a5*jTL{voN^gyke-e` 0GM,_X𫨉6xp.V;|uI^r>d-<ٓYg& öta1'lgܖ9{ǥ7dfƏqhA8j~!5SjR;gM\G8'7K,y2oY4컆2f/T_߿wLw&k߾9[Q7G߽oڵ 3gi~.0r̗?uۣv?ы=mˏϏ? Sf&oFzi<1 KԏaOEa{Y)oɇ%v/"(A?$z9ߎ<chpM>ԣ+&Uk/˿F}VbB}|.z uk(}qXªK@K@K@K@K@K@0c@/ (V!(1 @!($BPH @!x&BPBPHkVX@ r[E/!q0(Jf&ɎŞFQKٴaYdYSd*PE@p"5T*PE@"UTJUE@"UT*PE׮"UT@"U*PE@"Uw\njĉ?\:Lx/&d4H%]+12h[BHk1*S 4v}S=xdmReEFoTpZ Z\h)ad&9@Π8G<у/Ag};0 Fs[UPLZ'QB)Hr稓Y-6GC1@gB pi^Mr1n oPXeʂ*RP\yzG˹%@O¨R9envrk]m7vX%{jcKˬyR Xv)e[E0Ì 1EP3Xe'MqϨj_*ko.mXD(PYo&;C!p:L JWC֐Ɋ̕W:;1O 2-JHO,AM^{h L>m=P/5M8bμm&ei3XbO(8A$7y9ShE3eaoWVg9LtzZI6Xd˼ϜFs`L0L5 ɸs<OSr8v0 G0x,PFf{]W7}w]E89]'Lo&:xɲe_>-y/NgZ! `+ANWQ_h:N3+ʖؓwPjm6iyuLnř݉N^m2v>&y<:jRT<*Ij+Tm2.Z.+qyV\+Y*΅a@F*7yQ.#I']+$&UZajdFaLJٔRrCI=rymg=Î3jkv! >4Y#I2nN繡I! 3ӳٮOlۼ2;;^TMع%wXUSw&u!`Bv1Dry %zӣ3"WU6i b ,BM 2z;մG$'`$Rq%,*pвh*ɜ=z{s6n/|a<Ąsslf9^8+zfֻ_A=Ҹzkny.xuj3,і=D"ȂrmoVMAԾ AHR !qa!p<fWRzBo,jkfzw]BmKݒG9~[9ѳ##=0h ] B~牠B>|+[O(HZu<@m@m0J+h%BAQz4=Ð!"kȴey*;a6 mP_f (y^p !/=2U~d$o6CUvtel2(α .֞l&(og{H>'D=:e#s"[ӂ,*ֆ!HȠ4N*#|lĢQ/b銃*MTA(1H8x%,&LbPI FB0*f؄1pC8a2s"tR'BR^Teo yh񎡪s_2\o<6N߼:=.}:iDŽp̎Ǖi1.sҩx*e9(^~[cgz͂l'Z5F)Yf@rL%\}ё%sd7%vV-@@v dNØgI;Ph5RaFju"~C ` 5~{UA{PZޏ:Ōդab.V~繬ǹyD)I!R>HqV2K -",kP,rg  )+ AJSRjxP) }P@ɐ0 mZx"{}:mcnkԵ0Du_1ytxkN>\:Ѯt^YnqK]&vut7*f6zEPZ,JIi*#ݔU'݊ v%۲._yrD9u9I $/nR^#/2%#UٲX֭icYiK+>Y1q!e`Gj@-h_?u:]Rƫ$ ˒Y$00Qum69) IǴOe{[.Jܗdϼж wrmqS5-yFhePh}<ѡWN֔8|޴Od:[f>6/Τ~yM61 u9}lqKvy]RfYE x?m xMv _ޅb`qѯ}7^A15X7 o[AnR)k—R\e*\9&@wvddJ\jVCF5'er~+LDyi|.w8j ;#. At3q~g.c%ח| 3AԊzPۦ m n4&C7 1ޑALJk V MCDD>hz9fM0k+^8cAJˬ`-i$ w2&MRᗟKA_6!UkGg])ݽ`DI, 6WEY}x~۝sR`/)Y=**a!}jRrxI*[En1:UWSt`x!W_HIpLsJ>k?(㔆M#!` s ,0E;cJe ?dAAyMTRY 'QrVn);n1QV3A}#2SGH!e9K6_&LDF/!X lYƀLX@g'0L[Վ.|z@w1@p\*cbC΃hYrQ!`$De [7W{L;Yn|/%ޑk~.$>B"_H<ơKUGӁVAMURPmmpQF7u9Hɪw`q˺|0#7VrʜF ȁ6`\:B@F / mͿEu'.jvGz7hL2Q*:F#\S_u} l S\ϕ}w]kDA&#kaf4"WS"9ژ Thc .C}ee蝣˸ 1<,sCt<6xzkJ/BOHp^ .Sr"%Ya`9(#AYn,\ʛwa-sWݼ>y{5YooLR.DZOnA)Ƹ!04֣J"jo."GxPWP8L#h>VP(l71ͳ@ǥh-ʫTf\Bݻ=IHz=}*sB3r;芛o܋S#:. xn;h*&lE*Ą:IaA=`$SE|6F/Vdz㙞zz]O<_#A(22:2?{׺ƍ_˿ff)/r:;3IWZGUbL IQRyyy=IdK،iۼ4>@ Dp^쒦 jJh "0FEW`) 2gr|ZV,rɩC" y5F ǀf㼐pP~Xַ'}BJkѠm8zL'9Emq4l_jÈ' mWYbݞF\P6޽K(seӚKh2QcO}:3t3,B&#q̋ג>(-tBb" ##wd=SԽ "u]ߞV&fҿ6=ޣv^|;~?ū4Gۯo EO7Q7 ߫0 Kr/_ʨbS2Hi6qGbJa~4iOÝ^SIկZ4\/_;ߠwH#y$LbrܧW)*\4|]GŒHh<=;gU߽!z#z[p̐\ ~eFMq=NŽ0pbi gnKG_.ٚzNkD1.USO0"bsq^v++KK(]Dft᠙!IAMC㰸@? sED }a 8$dRR)Jb(`.%W[CHx}JR @&Uخf8#<c[m~Q9LK9xƦu>wMӵGgY8H ZtΛ9'D'ۼ`Ye1oXR[53n侩 nzsӴ%z\̌@2n`AcՖ(&% Dx295PhYQ5>7Zlw)<+Si+nCk3o]BWQ5;tYx=N?+A}o9=d6ɵ.%;/o3xr3`N '26]F^p0°./CNdjEv1T e&qK~هx3 BYPY zUUf!ْjI> \M%b =7:N-/jyQyu# p3[\'bxNը*x)?zSq;W/93 KdVN_2,Z6.U0a]_.[}`YXӜE*{3J :iZbᵛ%W;pٯ}plڢLy3gG>Scf:Uq}x@*X<>\;~㵴v]U#Uv4tPU So?t_ 1ʥP\E7c sO;"I6B]茐;"7 #:+aVHX $iɕQsNN!HQރr+ǝ6*Ѳ&SDJE?_ea2-vBp:jKX@)kmt".5Gb;e9ҜvRE:Ur¼V IS$NVc)\D,8!*S %,˼K҈O%L95VO՚!tS9kr]dҝi0>]XYoױZPX}`J#6ʰA2 <9Xlx-l_T}e$uS̗=n>x@Oi0*P.MaF!;klᷛS~ӣ_ڈ8QTpdՑM Dgq\"[vVý{TDo;̘GwW&kϪהK%nh >$]91iY8u ோEwLkb*].J]M؛[GvڇwU6ۻ*|pu]<6 2닟>RqœB&vŗCn~]tϱ~J\xxL1)"mmF6cH 8 /^y;y$P{uX'$1/ƣp!}TII<0+0>].I] E]2C҂2^ 2mP\[ᗽ^ W-d:ݟ֫??/^Y~ЈXN#G5#7l~N!xtm(3FRy LqˀY,C2Ncʠzx-{G1M3!k/9vS dRKJ1y78'q-'~^YT$<a` p ‰QttV q$AcJ]qfEJ0/`ABKp3"اDQKI DpC@,%0li0&ad,n 9kGz;O] _vזmx<,Q WV7xQۆn 19VEI<,SD_!̿i (CeglxLO,ϧfy\)*8 1*EB$@'4`PK#pgyZ<]7nF6}g ֧K\rxer1`F8/$\7_E-$1l@i r geZ4GCrINQ+}\:g![0IBU*S_6HyٙзBNc|v2q@Œ],X>b<*:&sSrJ_%_L{0^@$z^76%t F<7NS\=sgu6^ǣM HEJgR >0%WGͥ;A^LZg]oi%l>; IFg`xN1a A܃XqPʅ 2R8.458b`EL+frgitQSa7nml<>ƛORo/mPE(ϽlFZx|"i Ic6BeiMYɢc9z2%NLr5csO+13Nt8Y Z7? 3d<ODSdԄZO&N(8,'1" mK܋[GB\&bJ^a`혱&^eb̦ǮVۏGqb׮;ںֽi2W9!H#ZKNdp\h7 Bl`UaJH !/RA$M&&Vv #(:|2@:i*j;R*q_`eX:>K=Brd:-"gbc R36O$`2>ʙ4 S5 pWяIV{:ZEnK:< C֑=2uw쾂>cʁ"vy :;K}OۣmJAx|G.Qя5 T\bтR"f d,%+'^%\) *z5w=A}YL!@S0P,,ͯ-U1`hDh-"ghX*"+~='`ayWӖ)fv5mU\;5v;1Z$7! 9HE|GέW)XP fà uAHRkR, p(3x!dR17OhY=s_: Ko7~e+ڰd@hE:5tbqz"nR멼̅5 :[>*4[AYI%D A!'LYK!kigLlb*|Ӂ6 &ekЭa{mXqV "ԆhtYޢ h5YDrdI Y,@,:T{:u4JreV ,Y̰lcDs!0)7"2b0hr6u]ރ ,ݿ{S M1$KhC+v:ʬmE6, \ . u!5mb#7`p`2[#6@M`RDO @28N$lH},Նaj~ ~9}9P4e'x8b3'+iOӴ(zQD#.y<) Eˤwg9Lw;zZ '--23)b"eRg\08`ZM&?NжsFzVsKޥ\IkkQ/?R nNrtڥ-ibKrMN)6 ѻp#ŷ:Ot||4=8Z=Ԃ4h:G]EEI=֖1|g%0V^@}x2F_i_gg~N|5=x VƐ]s0٬mnە j_37Z60^!ckJ`{KByI늁H&Lq磨K+|<~\4`Ruj97)ia#]a>5l_MRF\f!oƳ;~n*C7U'gvhv 3^o?7o97?}O4 IϪ¯Wi;EO(*[- >hO=oǔ+k>rTCbGha9v$I՟ߏP6}\Ivez$V<(QlOd냒}v[iv&9*΅J r6lT\ϓpH r$'  b^KY0)DO>+%Hsʑ 8󤧞pXqauVLZdKC241IƕJoWb2RO:Uu*`>> u U.U ж;{ a6xC {`U<;( 0b\)o>i @/m{ߩ 1=2',*_m[ ̶Ǭ:~|($dϜ\ lJ<:(061bz Qo:I+k )9!]֥Ȳ^gmmF2fE5Ƅ kZi&r픮ժyڧܮ[}PϼnvL;>{q;-/_?]<1%ԩƆ g4 _#[[H]VAxT*ZuQ0L53m۪v)oGh$`9%8 SA e  D,G@AcFdE]!L♫SC5"n?C#w{vas}-tS#]MfǼzrnEpNGn>^Tup Ԋ IkSFo@%yCݠ[X^+`u߼`pa0@}^, mɬI $ g'g< f@VR. bDM5<+,'ȲK t@6\"!w6VFD]sTsL߼hâ7"aEN&gxK]"S v pGkҊIh(K'Q BB%pi9Ztw1}u[Њeyh 5 :D#vU _$dJFY3NV)X`ZQ tz xMv_=FQhv_xg??sjn޶ݎR*6} nЕPJV,ކ~'Y(~pOMسS\sJ}`sY&"9GoaNż0w]p {MT[Wh pi39l~_DG%=c67.q59r1X^5͊Bi-\PIbMAw4J4 5'o9oe&G^O4D,BGLY3{+^$ Sk>JtOWq?Us =2#(MB"S͛_)D~k&SULή])ݽYٽ8/ʬ*.|]sT=qG"8Y0BR D:XbL󣥇m=*+XT9]7 ttZ1y,)Yy[rV.S=irJA{3ȓg܊ F~Qk4R0Rr_L޵qc2?|? ${76-v\cr%9Mw<&30Zsxx~<<6NZ3};` JrDU 1HiX.(% Thn{Nb6PU ~U^`Ue-\i1TܔH%,41eV-\ >,W=1۟1eVZy[#2& ) i7[0h &ck{kUgU-֪kʨT+sEJTV 5AW_LyQ#2F$Ƣ*\Ua+],–VesƉϮ2KTxr 6mx JBydes)gBsL)hձt.wۙmz}9%]涚҈jW1 neЬ& fjW#t6?PIa_ˠE L2OAF:#,5=PȖ|ů7Ӥު_bf9:\W_2ٗ >U%3 LrxZT3RTޏF`>ČN&VߦNh.J)@},mYJVaL;ޱ\UHƓ$ߚEԤh-ϋ2:C./Wb~qI#qM̜Y94WW@s9{W,%,J$-C($~-z^Yֹ}HKcԸ\tUuF-,b@_LZ~] [ a͵$?,<3!l,aH ̒/ʛz b` 5liflcMyAw4Os,ހ&ѭ-sBjC:%X==T7ѥxX| XHy*`QRm qZ;'q{}H 8kt/'ĪtC>%bGx}qԉ☦+ % ]*,MGn4.l^W'ǃFCY@˂O?#IJsѝ%]^u \N >w64F_%G͞B[ U_nE_a VV+m`ų/bra,gqI"m F/daαuJot"X޸Aop]ʏ,$7Y[WԄB[Gܾ0X{׶ᙯkmf q!b*/UaϵzU\eRg)6|$=45&bfv5lNɳW/#f)lWxvLquay\e#jm)Cّ`[^.'q Ewݟ%LC`*9z$v . V𽼙,nn72S*WT(Z©V,XD)sfc *Jy ȦB]VzM͒%1hv\O*>/fp1Xҟ L]o[p1q`/ȋ'Fqi?0]N183b~6- R7A;owl*_$t\(;Yr-]@c8R#1J~YyZɍ1Ȅ1?Q1gc] KM8XgRc@iuc2&=T=X tL`unwXj;ObJ2kLbdj"bZ6L:v=& nr3ZlG1))--f`L5uH!0SCWI+mth7>|h7KDmcdq,t0jV+uLm M&MEɮ/v !,%ri?/hϵTIۂ% XZ昔j!'ʤ'MRRyz&%%U nɴzH&d#To*Bt F:MZ?ys\{ `vZ>t`Q6QZ-c>U j^-٦O/d7r0>䬞y2fTmc3TnWr똜0utLTZuz :QR ׳Qv}_-lY[u +*Dћ`UJ8@ USjj8aKid0{Vgu{<~kf&/M}L_,i9ʇL[ϴ_jBBx BJSe`cuz(SW6:Ro-5W*YW}vY'M7,6 &eqSpfTI0< ʒ2jUq}[+~ ,qYXu[d嶗|FLV ÿ77F(Q&W׋wyd9{п.??)%FSD9xƸ8ϢlUUQv)A%?w ~IȯUxŦ~;xY m~ZߔssJHzE14f)Z]rHVP4;bƊ!1g++ZCyYJEc$])u| Ϗ)Jܽ+' rW? 'f)ĹĹĹy.3=$'dZӪ gUA \R\)*,!HBB)"Zj^SiӺpY[qf4-' &XB `½̼}o_z=mm\Y4ڄV,iwKz;y>P K- J} YdPwQdrbةiAbBLQ$s' _SF3ޚ̯/ 8&n\d1N'O;Q.8vLMB=vz%XG*6 lِ(Ҧ} *c̵"F83ɰ&ƕex#8=di%j?=ԉXZ)T ,-Hһwz"c{`Cш A #\< cڿ3)~mmtA8 '_32Gj`cLGHA|K¼8rӵ-h>[~rr=Zd ՏmY??:myֈ=*֝>zxNVw[wR؟\.cPS;W9[[j'mݗGF'8mx)Q"Zo%#·`p,:3t k{>^C9^l4Rӥh ^}@r2wah0D!C 㩅qV AqDǣ3'=9(]}grzM .ɇ}Sr"LmąqVKֆ#Cq.CЫgp.,6Al )q,6qw[=С8)Fd>HxEo$` X0H`pJhCtg{^_i#`C423kp"ʘ:aU0&k& ߽sNm^\xdVOY@zֳZ+ D.j,YeG #aPEVХ` L #wp 2m<b{vKW&{mu٬at#S0ˮܲ bf1%y0.Hgk1`6k mVϤjS{$HJ@-/∑~v!TT &s>4D&ɀ'flpYcU`OrڵbhEË&ӷn8,29fs#K7_`mlt>'ɺmr^N~}\+YJ)3J5{Ƒ$d9ȣzhT5 ؙy4xdK,GyIʶVI2ڦdfDFƙq KTZ6y^P7^Pi =JJ{5^=?۟ʦ1KjV9 51R$cǑJ gFb0GH8"F>_ixm#UGnB._ :SJ֋՟|ߋ|nt|M|ŷh{^;V(#$Bks*#s˄ Dǚ(4#O ˰32sA& A0l%_"U5ɁDJِs Omo mrV8S*zcEY֒T)7Nxi 2GZ,"68]G:!s7e02:'* j<섁; 1]YW"3nK]r%K{؉CI" PVebӉ%@G[^^v_?e8%PHf/?oiv ~qAht@qɁVv_F3ij~22s2TW R,ɘ I&Ԡrn2@-Wj[nl8S/$2,1]~T`/~x+ 4.qݎikU+0l!s -)~8Q7{a!'ڵޙJ%5VUc;rUB#W 4L25UN#W_ȼɝwJN9L3C(ɩzlǷ_qJSE4d[|Ge`}vg 6-K ,Kg57,>,R^nëꈌS N@k?$LIo\'ʪ!!'*!_u Sd>>u2Cw‹ 5 -WW}) ]p;BwK6ȩ`w F⧃e13Um熌}EPyr1X|~`CՍ YnD6/eočvj*qV:ѵ,[;@ȷñD7Zqٓnxqߩ&3.ZgΒlDŽϧ9 &84fti_?Dˋ2j>.Q}NK?+}Ãh62,$~q_c,1 QyR|K-"N^I~RNe;'-^U^b@I~,b b+E?V %e^o2J|1J ?(ra~Mx48JDU5mI5?o@4Oގ˷*GT^wM]a%֪V8x1n|yf "$7D]?yh)=)DSdd\kAh(`ncEEe$B7Eit8 m|CItN bkTntx)*3 @f$Exd1U@ LC盜0Lw}&yK?>f9bV7ܤ6Ȇ3';FNpbi@w~`PZjfzv||:MdevgƸnƏ%:Q Ks-T2^x0|Sǽ YyrXC[c/w!%u.sz\k<8cR=>țԬ9ĂHMͶq O|름|z)b%s$ naZ𡃔2t_s<cٮ=P[AZsޖB5ƢBٍq7ǩqCuƋ +VUGΖڐ]/֔o\Ѡ!k.PKwwlͻth9A`f5@B[ߜaj47TVb>@`?a C~}{e&xM7X%'-jKHcVra@t&Ѐ+|TGZSS)V&w @~K hMfk+׽azy/4(LTDY0ZubXx@UOqb:/ 8- Sch|I6Wrбc&b=@r8Y5nۖhA8 -Bl>a l14CN(܍(; "Du; mIY EI⅁$d1vL>G&奣hI4)xaT4Ž-T`6><@e헇ehӸZL"b}XbK! LhA_<{F8䛗xQ_TPbO$l}/99^L-0b崥y| [8Hց ~|1pm/Ley^͒uI<c3k7WWIp;CVEY+G2L+AohzS"|\ t|?^"zF\O?vtw=dOד>6ʗ{q\0$h?m|+@YXCm> nN%!L$M"쯈(2a~@D(&,~mF-(%=JQhՓhv͹Rc!(y$O)*}%u lK04KOSyFI-hw~2QYd1 uN S369Y㕶hT(j طPo#A6\W5*%=+(Iա^J4(0\s.#4C$DQ}+Yb?+DPA F*#'!8 iQ`3Ep F l|X*( tz)z YùRvJ6MefK1&0&CcT4IpEPu.f=Y}-ȅfvJfCIbƺpD^_J?v&U#bTkqIA([\5)C|g V<^2zQG!͢,Zfl%,1Z qJ҉0|*HBDH TȮSy(BZkS$]%Uh3BqҲKZ9rwQxHuH 27{%qUZUǎB6ԣtv۱V5>fNl1:s7OK4ןo waǔp0M$X\HQF1Yh׃<ȀŲFNROdzʰd?6ʱfאu7.g%9I饼=H_"]RiDqĄxZ#ʥ~Y:qusI4xݦ:*0YJ^Y1)`u2 6 Z5;|[qfa|:o7_:Sv3K@=\*լ;*q"hCe+[ǒMzjٳnGk?>tCrns+ᔝB-5k|)G:ܻ _K[4]3}c0wLBKnC2^ -FZ-%Gj"ׇwK3T!ƶR^ C&D\T~2N2ω' f(7CB x*<5e<^~Km?7:>9o l+*U/s kG/$RmX#͈W2 7og@fxɽ6frZ6:)Tid-sAg>"P^~ /]4JH0\D\݀G""H^o΅3Ld[†Oh'0!a!xau7]$Zs`p Mk}D` {UL.>X~QEd 2-ڼ9Htw-$a!_/;ǣS?{8rc q A6/ &w7I[הdH$JA]{fY]]31{?x}=j?.'klԟzJSKaP)@^PD+ R~!*Z唭ې29~nL_ ʟ_2!҄˰,1l a'X^؉X*(DX')׼x8Ji b^ L%h.W3eqi`텟/ai&O^=C1N1cM ݫ[_J k5obBS)'NhIZd;vG%NtNq${omJĉ"wpÖYxNʠ"!͍,wH >I NeR;J?u=j*o@Q3([&Z=:f㾾'p8N PYn>;v5yb~J!ؕ |85ʎŽ#'!U숉e@f P6$У{Rf z-JG3\FC[4'Rh ٯ_e*D t f>WZ5nֵ^e0IAޮǶh6rYʗQ+oV({ٺ(PH9pHk>Zm-`q$p3u}iPkJ0rPW 8^F8Vr\Rp_}$O<  p(4ƶasPrm=(@1&Wy[ ƳiVO#}'+ƺjEoqsƸ~.{2ª?VIq FaW+/˱{hf9;ȧՑ; ~#pK!-Uso z)la 4zb b17V0FT fE%JvRC5Oqd4 \[_L[̩`dh@LNXkNoίps7#+ ۴ ނ(K7$ WFP YU53:`FRKKp*w f7Hʣz-ŒF0coƒu妉y[T2ŀhf0Xz+{Vؤ؇"%WeE2kwNyO6pxV)w LT5l+k21VG3c]ʨmr WfLVOZ[Rsjծz]hOYQ!&q8}_nulLDߴJ<d0$Iey  1Bݔ{I1A9F!>;Wuw$I8Da( ꎸu8AgqP*ՃQ-gQ-'Ex+[ȕ+.s g%j }rwl:~e쇱6X lV)D$rНV6Q(ë!ܯQxL憒ŔJ̼)sL%%pK)YE'_>OQt6(QRr$t_O_(Q*[Yt#`nY+T,$/\o۬ڡ(1vʣ4eh,ĥɞzhS{j :_gS--)Oje bNE][ÙP]ǻN;1*k}{-8]s6O˖Vt&G1tk+"?{'{6l)!.'d(tC+{3-ΑwQ4^3x$={8tڎO֤zy̰9Q FZ\Ɇ /> HVy[2@8f/VbZ̓mGZѯ5ƙBJ+w#/-;Kl:-˂GYafWW4&Y>m/lfk(.ᵾO3BL&E ۣG/LNTi}yR rCO=b6 0P@mAqũtջX-UQĸvǓAuVz D j1*EqZҩ{ r#>__)%9rAqG0 _ޭ2Uωql8\fA4,V%f,N#2 e;*Cɢ粱AV?cκI!XƮh%$TvJ#.>*I'ǝacB*~ƍCQ[ gtD[L6li?{ȕ`4 qq+d6bwl$뛟Ю@SFSd[fkqx/!CEU,B(sK*ּ}Fv;~-,I,|,Yn/RX)UO몐VMcM4Z4f8d6+-!"}8GM+W1>xoÌص""KWD?˥ ƃ·ѿB.Bm1LƖ|2NV+ =yc1K'Ҋ%Suj'LhAӺ2)Y?BUTXb$?N;% T*|֏vYܓ먨DJ=e VTfק,9/Hyiǂ&掟Zd4Q'閻_r_-E%\;( iH!0z)!  O9,̢L)89i*"xd!h;f>4иRp[ܒ~;svɥFTIE߷:}Cэb171XRE P%J#BR*x 4>eO4s )KY*Zn@XISpܠ Fmq< dm#z+..]m2[X];UҗڤgUݨ?q4ޛk!0|-ʣ՝mpIJ}z%u< Ɯ*.35- ҉yij"}ٱ\H}6lx`,*VX,Ɩr`wR6+!44 AfEaFUxv%.31ˆI1QXjz`9y^i:*^?@˲$)ci:\=w^='먾"oƱ `LxZ2UɐuUXڰiAً%3gZ1I6'E4 CXq"7ReNN/k_E*"m ? ws쵗TxlBɽ24: cD m"qVz^n=c.BuW+th s1PX1)]6ܜ+qߐQZ=y U\ Z6}peGY| .x/?{xdBBsHm֌]?؟-*Ad!UʙT6"˳iB "Ce{AZ*'_pzgblJ¥VXz# CJaieZ0 qvoʶGu2gFZᮗ1#F4HyPUid,K}"&ǔ i Qh+}mxJ'I4﯇l4pI`E@hY3`R؛f OT8d𲔇w-\,VЂ3NTghI*^3R$H?~(dbm4@z㽙BhFh8)GSTEB $Pr4jJ4IQ@S(mЭXy| ,bIOݤxO](eozp}u )CU;Ga.C:zcQ0QZ HuC@iMc::ݶZn@~dG㝗Aq=ɿ՗-KxP W?_ :LGW%-X}J _BQ hޭ*Q Y)og\VK1Hmmik#G`?oMc,̄p2թw$G %GCH},ZzpC4ֻH Bۉ$D5C4oiS%35e@K(u aM{q&TO; !"D~bJpjhl<-TPgZH܂z?a7]thܨۄ_WvSUS}AcFՋC6Cpshǚ=7S-wl1q0J13wggQ1%FG7"IYFajhl0aڜR P_W!m2^ḭb@߷B][٠k2U0D,H0LC Ԧ1FZ] oƗ?sD98 (W#:7p:3_-b!bZ~S4j L6?4Z-'LJp~ZoH^<ݗh!/6r A4(٥J${*ŵ\e%?ϦMa.DCtzVMbMͫ5LBD!.F`PީjP Ju=J+&"7Cάm'tRB۷F ~{z{:F{\(rͪbpT^ l3q.V?iSƒHl4̋LDM-&٦ f9x3ޝ0OA=;,ğe&y̔dxzJ]-vQ3V?XQ[3 63_UY*CJC3R.f>蘆ӂ*rc6Ki7޽ɁW<)[sN\JÓbyQ;7|5J"4Xd/b ҼWMC,\l3嚴YVEQ!"]ޡWV|x J%yE V E67Z]gVa]jPʡɞr!vRs@v J5Xar6+nDy jbH:"Wb0hNcaCeHJ(4BTC2=,JDQ$VV9Ukh#b{ԣANVc/n,({釩AaYrQ#2ť06[LU}6U;^"<ԘZ,`n5^L<} xJo?4 ޣ36m є6ꏣ)zNdKa4MNz}X&3tbqL$C1)~}1x#%X9Iof[ax08]Njx!U=.OHo9K1Os>-=mڦ6 $އQk&~]mKfM#zްk*.c1[]~:~:pzvXֵumMkDn)n-X庮™`+ 1i¹Xlw4*]O~򡜔l,eB:0:a1@SNB4p-2hGv \).G#La=OPZ]F-QAkN4= V/-Ӵ/Kuzo!ERs&SS|O79/z5?f$;(ӽzs@)/9c X2n@j[m,bKJIQQü֊&7p 0Mt;5[GJe^9 2#81s66{ R+u$*&}R/<ݝrWO}NS-Xy]`^sBbD'q`bLˀcdpQZz.%=ཅ}ҁ)22# ,ZZ$<.Z 蜥2Lv{[-rĈS1 b5-v2KS,놜lC .6ck V͛Q(n&&K> mͩN%KCٟ^/&Xi0dx3f*tߵ}uf;)_E9'Og|& O= ']m>@4;̞Ҙ $촶$(`Zc.rR/+WIuxKo+h,(n0ILw5H5 "AE$0dX ڔbZizU8:;IFsp 95X:hm0zf_gŗтI-vg[Ѡ:1Cp-m$"FFpЇk4'7cTAo<ߊU {$nc!&deἹhj(c~ZGupWJ%X&]J.s恮0,ݫP9504qo@.KeJZr)5~|J;k&770Un1ܶ4);]Vdkys@KM`{G Xm0py/ ׳syhGIU ֔>Je$u3o@cĥܤ-AF1ueTVZ?vsvfx6ZPSx+!–{tw@Tb-A` 6.%\ҏlUe Njf? Ec*AҰx0B"&9Eo`[N`>Ngl2%@zx~[e9k%|8V=a2-czOxpv ebh K5<볍.E{[\*N ;R G < N8E\4DRdVV^^JYo[6F^ns{2:to=^l?p6k1]$'4pc5 9MCq+r,543ve:*,7E |hzkh2_]ZV;g/LRqO3՝Ҋ qߛV65"u0CQz#d!G } Gcw 1Kϫq 2hfjk?Ҥ^3xr2"®al'Q0J/2kѕ]d5lӆCΡt]ݑ~G1psI,;u r"U}uM2BdғS 04+*[xDP7޲`GGQeb}3\{Ɓ1JW^̕e}"`Cc#n=fi*A2xk.ޜ)~с) z`Wg(E:xX ؤ]W߳2c7T;pNWqS#E p(!D@PԴE1ׂX6a\-:}w*|c?1te3v(ɗ̙}yX,)*@Zww0K2EX9Ї KV=p6PϪn j/ysޟ0 {`j]>Sl6[\fS#Y"?Ń/w+;b&Y?T~/ '@|)ŶEéR7,&胴Z\b&H?mS [ޡga82IqV$OaU1[Mgz6_!rZ=/l(za#(!dK?lӒhHJr8k[9%eH"]wf&9Yb`ʇm`o?~Gxx7md=[eށgB]Wy,.v ~SvR0GJNJ= w84Oŭ]C y#Մ&ujZ{%!Zs!7~{~ w8^\L-L3Z -E<7^S)KjQ;F7WxRjbR/gY&y6aw;~|q3!fZsD)) "YZ$A#)<J} (/%6." (=^'i$ExC{߉s+ '6ܙN[Iw8IXxܻ(d$O);ëT"8 W#ۡEIǧ#j}'V=&eqPQ},gW< QX('ruRSѶ-W0{ cv Sxq8'f2fq/ pV7 Vo8_?Ǔ/٪.tl )Y*)U5uˇT l1b" "ReSzϣn ʃK 7I di\,ˤVh2Ⱦ2+ e*)+]x "s].kiL^km&*-T9IzV1HJiVȭRRb[K{U5F8BQL bt/8 pUura49Ko e"ثuW"4WZVmB yܸOa=j_`_W| ^bBM(^ΞG\I}.oBi[qb^%_b~8(ڜTj*(t(yc,jLxA19N3&TRу-4\.ҦB\Ei=lz.—P UwGN+//V* ";/Ϡp =As-@#eZLeYI8n.dl e_gGE3FU>iY^M~c!%:Gb*=8.~~\ʎ ;ę7sE"EU:#5xz>U%`JuKAw:AM qbYɊ7Fݎԇńh[B%-YVtXNj@jMsZjkk̎*l8T7n4x]ޫǙLƄ >Rő߳p$&\ċlZIJdhMu 7ȈQW^m NCm l֡δ,& R|x򜦲p\4 2zNY,R:[zksu?tF*W1}3<-^7cGe_Ûg`T.&b^ |@~VզKr2qPeuZe{A\ڃ) 2 jN=LHEQm!uCTeSd.n,׶q>I'h~^ Dhc˸*E^klkl؈w>ݞ8bEOJ8ȝg"-"#VKa^l8Tdəv"yCm|]&Wr3 0<~-n6$Y7Xd2|IdPLX_8Ӑsne9A(4*ZGYuÔ`7<,xxs$W686,_G?([ó>2n\IG?ag pn& V6O%e zÇɨ8>TҜ_cR0C'ĘRB`'P)!(|LC~]?K>FIR3gPTr:O,^xeC+?N'=$dfh1ZL L??BLN|)r$^ɃKHa<<]1q?e~6ꋷ)!A h;ޜ?ؖK~s^lT?,Nie^7xACkֶq \+_h,}@3\m.-oVul+ѻUwBC僊%l)8X%e܂WeȒY)vs E]G쪆*eYǮ bFn-C]NO.=&ԔEc0[g`HhCF|*q.ɦXk̬DC.7׊7-eX;lV ︂PDc;T2lЉGT>tw_Oh[/f5/Dn ֜\b2ci m`o?~Gxx7*25y2.J t{ˣ/ߵdTvN}B=dاb'¬cp16pf6]J@ ~[ dS(\ NnAS`HිAsYVl*blJ0CSs> 7dsYL/7dUVI-vZLf05edeK3'vxnq`@мb8jcV/EBR_- !H˳sq\ZJi];i[l Qyř:0Pb*=FG$d8L$QL/Mܛ86K<ٍǸN_H8I'{-E%*U%<1'e*EHT>h'٤1Z~whn $<7PB"&J43"(Ka2dd1M 6<1j6\౶8Cv#l88ԋn>Neuأ}Z{)N#]5%Umޓ6UJ(3/r6Q-GMk WvloKڢS]dyf4V20hWGߩAF;O"x [} P@S:+Ȧ5NƚeQP4tBjAv R$egFR=Y -2u4x5*I+YdJڴXK6k 2C%Ą s!=Zfg0֡J%2qg>x2L;8C8HlY\q-@cpyv1آ9Aj0}u/'> 4h%h:oOGNX 2KūbaW㎔o^sN =(o\-@Rxq L,U2#A3L"xSXd\LxmLkSA9qr na $x,6lmyv/T{Ge3*4"-A0KK>4XWwwAᣤ-QMX'U8\cDy2*3Yb)_"Vos miI~Sys s*85Ԡ8 e;]4$ S0f"WY|o%U⯵P2AYh:9̳jVB`#MAvmv-)ZmQ P^WTc0ɼքTp:n,pb΃RSc8T\=ZC:I#),˼#*RՁcy[],},6JrȐvx ^Ry{g}xp#`; !` Dz"MqH=ed`Ge,o"chV,Sd-ѿG伫A-(cd|coHThR*6ЇwZ0PdA0QT,rv)m"aRpL9 :0PU%.bIW 6eUa3Fb ku>y&$<{^V (W(RUC7eH U7suMFCqQ-N(}*%j)dUЦl97e۸1ΰa ox'6bkuh71֍e 1K ň^_D.:PUй D9z.\,nX)֬`beAR$ rj嚨v*nR\!ZdK._n ,VpCjRodzWmzz!ⶪ#7jydІ7׊^!rPz%>+:(JsP M 40%(Hc+Մ\ec[mm td,bu EV"~6kr'H!#[t Wq@~;[, =,#\evInvPל =jU~/z:g'ul~|6:Wrv>?>}5y B޻okܟwy>Y=UÚWSsZ5!?^\\ ٚϋA# K/n;D8f>z.;~{K==/7A#f{]Y>r:o^q5.$(@`e:䆐kXA)Kf=Fzܤ]}5EHU-jQ$R{2\~?jKisA|Vmx.?WDZƗ8x[TrhQ8)@H2ל @6LM4 )$U{-ymHADj8Vuz#iVAϛ\֢lXɑ:5LMUEUQX3W]qo {SNPjB"aӛ@nJ)]Rm7y+Kԕz'ޡ"9Y[(4D]Ym$ҢE$K*x|u^FL~ޟ3B; .fc6J]iXw}Trx!"ZQQZb}oL&7+-)Ы-@ӫU3QUQuv=S~E.08y"}t? FѿPHb:P<|"d󷫍AUG.T~?2;mŞhO}q!~o(YKdg>_Y¯̿_o'w7ܗu;^MIXn9|qFM?7[c. B\쮜6{sW#'} VC:܍j,w nmZF(C?k~0O2Mo&Ke&I_ɇB7)I9%o5LO,K3ok< la{Mjps[be3Q9M))0ʹ$^h%m h{{WGKl<EKPc6s癇VggVkU9.+qR6bUbjEakEtz @XhS̓8OϪR鈑 ݆~<pFsa>laώg:;${=맴S* M8{rlc_o ?~kKv: Wg?py.9ۋw ӣ;PRket~׻z?;FYx:Ezۦst!_v1?D5-lr`vC?m|+I޾c>Z:4_{"|M^Z_/uʪ6/]iGrkk[&u3 F0 zauN6Ā&tWc@xs-$EɁӓP]p 2H]ǁyZFm@x<"āJ'PtD`o d8:F6L@_;i Knp /l{@8=]*k[ 3\pf:|/<3\Ir5p:z.$m\*Xyᤌ YfU%__B )A7# O&g5ٙdg̕f`qi(!RBkj!V[mLT(!G0FҨ~6Pf1ɗ^zt7?W}Oo]>~ӁSUx5{H~Z"V; ]H?^n{u7tٓYk 3^HZk)j:F ],׎W.U ~jҁF煊.z?JK(WZjgN~xKc~Lqnyh r}A.=~#6Bפ:5@4.LS^S⿄.O!.OfPX6Ⱥ1xSl̫f6T1zas(gдE+U*`nbA?#Y>{5xv9]toă1I|u;?ЃҒ$o_z os%,ŀ0)2?}:{'EsIqmD8Y >wo F_ v7?{s>«!J(R(cƉ@SHID6{B"AGx\XcSg!YD6 'wd8wu7[ʰ"g+dqsג\sMGD fw'c@D!Ǘ$#匡 |3 v_ 5(}4dq#4`MzoMVTbNiSj} [,|??;}yc`tar2H&~jCo+O-UbR26 7DE+bA%zSc&˽g3PF+%麎yE7Wyۋ=ʯi05ن`0`|Q9QM;2QY}1lut2 ,&ufy;G7IQ _uU!E)GOAL-A=fk,ƠQ\3K/h>`j*L/8/aM{QEbVH$p:'4)z+wh)Ckz(ֈCo8ֈCo8,Fܨf#6TDMp=XYt*s7\Slԇ\QlD.*,Ge@;͝wU2FWdsK=6i&yc,*$]kTag EV9=N'꛳'.9Ǭќ]>JByٔwTpYYdlKlJ#L&1?za#?W!_'"Ud}- c׺_e*Oת!VS.!ry=:F.v;w#ג0[)9]\:GHh8sFU8QH)hMʧ|(w~lK;\z\+ad'ᓿ~LIQ%ծG&.߆I*EdVvLruͫ$Uo(m8r_Lr-1Y80AGŽHU԰"[{R]^z 㨧Kh&[$CIJuKjtFh[1{~) -xR%@ltn^!J$xK !-zBw!zhȼ!rc='oُzbՔЌ&WMwmԑc9wc]nѪh@>DIBO8NA!M,lMP?>uJKk QѢuNuqu'hw{.c-V VjŤlG" u8y[N{֨VVzstLڑu$rCr 2ki.lBb{߬?NM:uA!9DѾLy_8i+;Z)6opa7)]ϱ;8Ive|f6/rm@vS@vp'SţxOh}8keM_OaK{)$|hP#Fe`0aXw_*7sMoa$,՜֓6Rq*6R7.߬ FeJ=MbN +6ĄE*6DY?[!S=@ zH0(U{7vP?<=OR=92<__jYE, &H&SoTmኳ+S.a}'i6m5d]69r>Gi"jA⊙(T GSB:8sg )79v5󵱳U zN:;jg4FW᭨Ole͵;`ڿ=,\7fq+@1']"{Q1}:6"Xx0-eɵC5̠I=sJʕc-gS)5Q_!{tx㎋}oǛ7o`U=0'ӗKh)b`p&IYٸb!ʨu]cWK׿QɌZZ9)& qON4G]Oҭ` RQ}:P̖G('&~6]SP:hޟ}UB*6em$E0r4Hn;XRQ #)ke2HC\E1{el*7K-U#.زΐƕt DoeIyI$r@>ܙ0 ̼th7󞛎өTFg(@f3 MݼɂN#/cY<ĺ% ,)Y 9iHVIF3&^8 ~6Os6#Cv铑괜t8#@v [LM3* ڱe/e=X*jAZo ̀_[sH{&FuK-Ve02n|T7|.ԏ<қfn̅\ґe ^8 DabO^z9C3]YBqtztAz`zbDِWQ/us3!C1x&,itݩqlKw )CS]tdnΖ9ҟ֫+o/ap>U&@Jٺ]瘲1x/R  !v8Ec&PRe~ɀDL͗?ևe5)](Zo,Dr./{Z$F;EąTֱ_׼ot6Gn <@.$I*#vulGVm#|]^z,{ \^彵JZ>N6aK$1WTZ]Qbts9ɾT⌞U6|n}ua4ʟ4eq:v*q.AiV}@f]\ mP2MLQQ?tU<jhrRrbjyZ☒/B74 XV( Rmgu:pCj<0%w- fQ?>6gCĀ ͈kq|SL7!a\19JjFQi*p(xnW|YҔ8q;ƼHQ)͔txHA*c7\ 9tu-w6V>$p1p]N#vPnSzR^Z՗CMa69d-L U>y |O\]j8.!![q%>#?g\@!s@ZTp(X@&̻yn]O"};>vvIf/DP#)\A13"u:9(asXJ`X|Y}Vۄ%Ke]e KBPLTxq*( DzAQ/)tE[LbhQjحYё+`'(MQvO*C %%. 4FBoEqxE)$Q]Hx“ -% {]ѵhU}$u͛r9gk. i*͟R%KQ k{wo ?yG [vGy9}z^OZ9z1\ֿPs7s!3\k:K,x_fBQ{絲`qmpxwoWǿ?ۙwo/fM-_$:w=戱P5'*AnJ<1y:tZ*vc 7WՎ@V-eRc t#rVGl!b ߻kS \e;&hR1oܤ%NLaR(s y7fy̏{wE#wuͫ%xCu }޵$7)92fHݴ'[@qȥztuG/R`*1R7.߬ \Gي2䦒r)u%>5n4=FsIEÀ"D/\}WtW8׳ZUܦ뢮WZ/aHFT+9ƒ(V)"/ 11#qFo n<}_Vtt FkrH%í6Sdҫ`[Ɇ"R2z R_bu8_5Hȡ(.ʔT fZ:f=|sT35wu$pV6;xIJXCbtU[b^6!ٹh_[q8 rp1%[^x9 UΌTZZTͬ@ɷjR8v}sv.{#YPz}!zg(؂RȖF?',@yyT{\r3w2IOFWRRF=mdJRN7i9X*6e!]MGs$Gܹ=wB֪k<Ʌ{`cra(=KI޼'ֺhg',LM9?3hWz.)Pܬ7]Dƫ21,C2ey]jվZ֥I)H_+IMb^KUHu*E{$WO_F_5}>dJ=}ϦqmjaSf.J'ejL^бHuY7Z]9:(Ka$fM'l<>oogf$Y*Go~$lVePUoG_7ұ"^qä,z#V6v#q tDGWM)A?V=)@I֭ؽ־k:~ۜv`ŸŷEjywKXwoF|=Ut2w?C襁+!P(`bl" MTHXF'VE l#$oPR[4p R8֙yx|\KN@s盬mv֝3wo_بOon٩h&Myh(/LZ[ih7=JCIo_%8oڜq,9% ݜ]͍$aRL$WM +PmX)*.5PUr[vhQj|uR)a/XY̰Ί26w`9%PXG#zw (K#%rNmwjZ0={J\VKo=NՔZC)9y\хՏm RJ>UkR6Cf"~azabV1D^ij 'McnPi] 8T5Rcd/GɡUYsU4gRff8CQ4q\+$gЬe'j(JUeِ߽J*2E(d6' aҖbؔQ%3w5Bs}RPdwsy g/^[ŵ[C~P}5áfJ()dlhCTURSυE*%6z#ԢBml nlص⡒1" .a$ n9kG^%K~ta_Kz}%lBQGm5y (Y"K˧~BeRc_Ӽ,ΈIJu d >4~R >mU*%IbUC@G:B,rZJq]Zs?yq)Q^lbKRI:X=p Yr0Pnq=u ɇV9ޙlpȫkL}$DJp)w&ɛ[#Uy 3@vυw51+®GgP^#MX6נ:RW! UNkf9zWU2/bs{ZJ;*W j9\~5Noxy#̿'Bb*1^!j=~?_vi~ưk3E&jσeǖZ+IUtsx9o9k킉Xyt?4Vf(Ȗ q=nB~nM`snޒ&~=bUD7Ggͫ5믿s?7餞IQr>y^@/p.`C.Hvm[,_*ۅ vđ]?HpCCܸnQBm"V>-%;% @i&ĭym+tu͓D8f=HyoyOq2/sT)A6pwXd'ݭb_y2*̣3:nX#]DbTl%m*ࣛ;gς~F4 \D 3L4}}>w=6\ ya߮]DBmA;u5In"=pC 8D&͋D?6`=mĊ4](}#/'v4 1S6/J X0Xݏ , Len"Gu [Fdkhd7'#4ru͓ 9a,'1z]C#$垃2FF-vHkπ:EvKo!pXn|vT? s@? Zu+G`#@B/lž!;חgum^2ǟ_=1>vY|zIH Lߍ/VL,;rxEsߓ. EI>^3kS`+[=2gBZeP_ Ibfx||vߜԼkE{֌' :)gERegnyB4,x,IeG[0ySbt8tYi(/|:;=k6[>o;D}5LڍSmsy`> hTL@z!J%D.o֎kZ&| ԏp4BXWۧ>=:Ůk`}?s};%N<2[~9E|~'BPHKޙ(?%'E\)٬K_ oSA"vM9-QɆԍt즈}xdJVQQlsXSBkC%suRAϑ,>d* ݼכbB3DxK`0ln"F2zI(d3ԚqeDZY 7vz\<]p>T~uNjԎJ:9ݱV־ ~?R@?`٣V˫ ΜE~W?nM,5XkLTnqzӎ8S#0y VT]4cn#]HAWGjLbM_x-pm[k߱|'SEޑҶZ!ZM[I"// ?HNR|m )tɳay6f]D"IH5W_ś6JThs'eNݧl12P/țs+ 9U(>WDps(` {=˞OgCRlT# Yb P)*Hc $DշCKYnM֜ ub /RJVV*-)ryZc*PJje gBCX]3^Vo/;;_E6?0ٓZ*ӗ/ @U56.9Z $[ޕM QhF  @ΡDz,({@s.:Hat7KD?AVn]F! - w5Jz;_"*1WTՂ)pi(x?btwQk EQ~˞ ECj]7A޼g +3rpdH஽r#P> f'4YS寏m,Q~2csD]&ֱDK zI9G-⼨Q9ZM =(HΕ20K}Uxq)GQ8^":o.XŢ?~<;͛5UH_xg|,`$:1*ǥ'J Y1rk),ε&+Gd*xH ֕.E!*[ʻnQ=bs/hu͸Jd,* h7Nn\&}̎ڥ͆]r*u:9UH}t.&FFW 'N Q%eA-үѫ)QX?* uRbidh132Pi*1B!OUj:AӯYm+%?{ɍß6wWI$Ei `q~XlvPG&G_ew]n\U")LJV$BZ _}eQNTw֚6kMO8]gB;iG`/auXs!6ʭ MtlKC'[D1TMA:wZEC'(33(H9ֆ+sieY*k-dņu. ܻ̯+@sM TP^WTdiNX^=՘Иꥅ.L.*=*^jA=yOvru䟻ÝpUS g)_2p۝Lނ7QaeOeb͜չAʂ{ֺ;ݫМ|@FhcgK[l繆0m(nGD% zjS6!lL&ĭO v}ҝsdrjsp &s$p.QX+`͏k *+}AC#[4~yw DW2֡o}[ۦJSN=m YوVU>]&TOď_N,KM6tm۸Bn -Ę!dٶm֑:PAGD#"6C:-zNИ|KQv`#cJ{O*3GhT:gT/>ӌ'_Ox.Ό@b,^Yv_\t ]9SJ\,Ebyƫ(no"ߔ<+%g6dz΃|,X^25!`}/oPE M4ޢ|̞BpYspg =. y'A;q x!"yLCXiٝ U1,*y[#Ru}ͻ z.ׁQ&r ! B=5NF_)rhFFclSX+I[Sjp"UV2\j$5-jr5`RbPybB^<(9pOMʈ^ݜwi~OYX0ny{Ƕ5fhD@;8bv\Kn[J1i)cp5xT CPub@J/C1yfPm!ֽma#;M6sz.eJRmWo{OX6w7w6Pe0K:뗎arDѲ2s1\,YʅRJ5GTz(?FFpQ,/Z7'u~Bzztk4p"i*o $K5F&4yUy,-Mwg4OPT\4Y onxC7fTq)FѩP_Πo^oFmK ͦGQN=M;K| PK[J[;!+N95{Q3tژyW2)5(5QŪ z Ž1SB;D6٫Ð~x*{κXl{iYx6v *.+1Gfuc[e،6lL68같UhdVu#vT>T0B!y *˺9Ѡ~sF 'nlrW+p{ quu휈r纅uHKSz<=V{sﺂXWQ0k"!K#]fΗ`Ʋk{]XtYJE~)jo~MEycrq9;<ݗrwDlDdcu\bqf3y{Egc:FV+Z̽IjUZ/2/xCa#ަ!yC 94iھQ}%M}xoŤzEPͫ9.4zVvb0RAiMiFh' &y=hBm4A~FþYϾ==mo{Ax!|޸_>&,_Rl7+yO9qN~_f;%BR([+]qhw¢!1ŋX} ѽ 0bHBx?9,'6߿~N"\"g'G;'/[k~zxRC9SUuƛ=1 RidAHxm[@uqd +;zm H=q Jgnϴ}oc z֢@ڴaV/ŢttzGƱw–^?͵妃fə.|fZ4F[Ud|`z:iS}֩uiDs^mS__RB/Yx8M<:?\} J=eӼVfVy&R}IRx7̓̃mi# G,Z>.9]t2\W>,ɇ`.3..sHTRl3+jG\Yd?LOO紓_7HxO.ZT݄;@9NM8viFkj[kzӓjuFŽ2fc2ۈv"@5/IR`Wxcumk{H#jV4BV,Pv즤axm^?kL}S3v~BH~ej@MwpApVp :\äKɁX;\B;Nd -m:ExDZQe[11S9=7+e'ba ?ۈ(,UXd$s_Ƈ(F1j3 ]ςgjm57Z,Ke7v4"P\\ˠũ1ӼLDxZR\uޔ2mpe.[iY2k>l\Yu.R<񚍡{ F"fzs\mdK5 |~፳[ueŤq[r֛z濟gƸl}}~+mUЫyBYMT<OkԛfG~}ԟfG>%B^ݛ6[|~Ӱ;gmnMF?aYF=뻖QM,‘}ҵtOFV {!C۔[JV{j|IjnlѵZBQsqgB-q"ڄI1]Ѝw7&1+>wc9p:o~/6s&[1" md "+`8f 'bV'"CSxgD̾} f+ȢEĀa~@Zt&$τMoi it}g6(%ʩu;IYd#w84=/IZYMx@O,ϪQbn@6PDl =$eF4861(WъFS.;*,im&7dz7k`v/OӍ<_|?A9V7IcnD##5veIwD{ej۔|f|&!X [\Dl,K:#(||x#S7'o@Չy /^\'6Ep'Ŀ)gy<~w)Tů|0mn]o#`w|9ҿ.B\wN,Wo~G6V;h -'=by9\=X4֑i?j[5`E/h@ Rz[5iEnR0r4jZRh<ܡNNBpxY`h5Y# ?+n44^?Ew{+}g{g'n\4N1^\s-`oʓU=le9cVrfѪ76.qmsuZxzڮ,AEHf31j6X{*p4jS8` ՇrN 6*]Ycsq}=B?m^eX 6$&`5vpPE[`72Q9Sn56J: oY0`}'+FzdgWr3Yb)mvXW zծW0k^p^SWEy5u1,n{MD|Z13S0w_/Y13"h !>͕?pnL\q9Mw ½d8ǎ޾5#-ĝH[ aE<7(*W^X3U N?qIkT/.,v9% %imFmpBe 0C`cAF)ivk(Ѫi[ ai)e`izf#>|}4K{V{4نc :M lVlQT@TIOEj39U"1e sFQU%R@J ,"3ttx *`^лyyλ Wk⼷F5[c5 F !]di؛9:yÜg_.H_z%t} f_.H/W{b:7'Uomj1*#% $k.l#Dtxeel'FjjUC8`(9ۏ+ >?,ҞK4ŮfGO*\<|EYY Ww0JȼIբ.Z^ނ?W^EVQ&[yot2ўY>%cPim߿|}E9*!$'"2"7=% bݔΩԠz#2gfE~N*&wt$L%*6z)-.0YCC[q-3^@_6%4/@5[N\UR[D;-ʊa 2Uu&T"z zVKk{h3_JgO*ÑNI+Z r-z'3r©M X8'ß7Uo_Z&MT勞JvYbJF0i 2!N$ոQı^{]_\ݭ>xHɦ=1:?=iV6{ՇG3KGVD] B5Kj)Hk2Z#@Ě3BH&Խy `}s5ed\:W &+mcj!5ȱr,fɽlZ{oUarsr-Ph}`L/|rvr <#ݫrҧ$a]PǯX5q>Zr|:ip84YN{uȡ4~IUVH2eBP{N֙hVSҧhĂZ\f7Rw۱1<"I#,Y/6'SZ9iͭ8\ ˺H"?3!OlE9:xV܊Ls?!Mi~?srDjX刴T35a"%FM)f_tm4FnB2J1YcȔŢg9G'DCt1x6+"!K&/RtB';VpL#P:X+P@{}k&Y.uEu= j#3dw0<+U :IX"Ӟ{-qI9g%c'$rFstW)rTFo0R!c$]r?m4n?g4#ئz"lX Q-M }ʏkdo]&]zqpyM3^䯄ylA]$$g/k@?iЁ#J&*ِARsCj\1g"3Ȁ H+ę#) Yi0[2ŻA'y3bƦ pMyuh\ l4x%l;i9qn㒗T}@? ݧ="ЂPm_ﵫ/TAS8|nZlt މdð2[afoTÓ0Lw 7*KT*@/}lcJwYojhz ЄF ~EY4V{ǁ㏥M[r-儑 $-J/w/ N/ 1]VHRѝ|sx'F8wy|,BLket] ̮&ya>bNלˆ\*r:֘/oCRWC{GZ+2yWA,pnK$Z]ɑ{lBhe"iśy}n~}!JmB'=pmp4g}pggjV U?]?N=?߰jV91Tድ?/7wR 5vg`9|qv= >!9rZQ7sSmwgcFYUٻWhDzӻĽLF6c^\A6ZHZ*%x^˳۬Ey~a[yDFb쟫˿PC-գ(#xfJwg[bҳbh>ƭ/UCVAe6?6v`ډ>י/$m7N,b~³ 1ʋS9&:JjAـ86`pC>3p놬!_jeWJczm v3`#wk=PPY+Ux2I~\˜8MRj5u܄͛.JD; _@1toG0}@Qc=ұg,,  伊&+CDdR21abRj8`xf,мIb>^u}gWח? t4AFl?g*Ckv;?|l̷Q5q5t{n 0Λmsfd̉$}C/E 7DsvpiKlp0.?6ÿoI WL͉O[>yP8W6mۛ 6jvO 吺Cu4i^m9g/l奺nF z`;4OwD*[77ke0> !6n\ZqՖQ5QZ]fBnHЋl['e߿?&,\8_ґ>jpEq7>= :aUy91sČctɸ!s|g0e."J.M偯 MJJOIt//Oޞ.{XI%[3yP(,k MH3hsO}<;1ѵ3X{@'nR}-7=Ly24!c5KjZˢ^Ҫ.o|ãéeR{< L+ΪZbʐ޳8jo9L՜g*m]%D(/Kd\RjK]e{o׈6z0<#mv%ك(|+Ύ#v4bN] WgjYcn<Z۩7V](Pj HG׷HFץ&ԞQ롂2?[ OB.#dW`\ XYdej2,F "sLۤyLD6U9M ALKn<̖6ˡFZ탑Vphsgl'YS9EIPFU"1A˃ h$l/fuu&ꭳ~& Ln)fihLo0, 4YLѨJ1"/%mdmn߶uma"Pzwߛ/oۻ/[h3>&3#Zu>tm5Q(9;%Nq](8b9g(K93Lv\FδOeH7\!%߷Vai6F-Akgr)1qI&*Ϫ*Ϫf6LTxAR@ɓ!+d1aGB5!2[&i \P; Q~hwɨڗRF"0{yqa-&~;9rXc NVDL/cŖv.`mY$3ڢC[(-AG!ɶ(*l:Fhs95"22mk`^лywVFޫXf{o& rB7VFNӫp ,S`[G:eF꿗?&dM B>Zvq 4YjfԜ5#ի߿^s@ mcrɉ~moUE_żzѹG . oVȪKר~}'D c5edm{'lA~ԶQM2guXf77f6iƩst:f.A, [d s+UG >RXƣiZQ 2zb6Rg\0aFjTk #J&*60V*{nBK뇕3҃Y<2+,FŤMr CY+=ga-!WBAw|yZnaLWlvFVJ|%38=6%Mcr X tR+Tok]ajyLSsMSt[>{<+sWȇlE Q6I<9V5Kvf"B6&|vJa'SB%2Πv+[ *km ȶ\SU<$H0l6QWK1MiH)"}Oݤ(lmͮss9`ф즑e@vG-|:Ͼ{w.W7ד՘Rq7"N׷7o[#\]?O]z\ .=B/,c/x 峮u_,nջ_D;@/M9>r-֟f{ن|}k>,bĢ%g-FkcFOH #.q7g9=S,WXI'gL>hPIJ$:RhRqf1͡sf=FGe[<= k &?"NEo8k?\E\}3{ ͇U6$].7og34ޜ᧴YB.o/Zy{润5|/'5+]wnpN%П4eJ Of(]K.s; >\h3/Fi*GN89FFZ6~tǓ*y<2%"cIA?QOYh-@97/˫8D':|.<퓏E1Sj~XQ5j,NjZE.'MEs-Lj4dܬƜ\֠P=xoff@1=L;A͸GnNsi9u7=(BTRI&ZX߯>|l%43 w;C=cTQF+sIYNRIp!6kɆ_v]( lHqWqS NBQ ۙk7w@QT/ZiFx~k2zF2})1[A,iW%H8reJ1*Fxir27?,?/3SL"$zp=M!|13G*i(XlEPceI{Po۷*eaN)m;^IZ_}`z\&6wuR) 85ד9k a;ϼ:s@kJh+&*N"NLP)zlɅ8"XLRhz+Uy2RR|0'!IduuvNpO1UCg $ڼ.!Oi渎i똶: "M^ 9J1/-h-odmӏBG:'}t$&P^1ܫ% rTƃ6TؤRB|pkN,> "d0{.#$4nO{ObOY7 E{`OBݳ6NOVd:EY%Zր*{R-eo|byse<~g_UH X(aX .2z&$Ł Mr;OH,4i4ItUqvǂ4Xi۵P fnjpQQm."yѐc1.DV N\h 3ǣ`-?0\d)_mP5yN aszoKBW3]}¦DG1i"R5%- SpI 00>>@N=J<4)tL xZZc"N9DFR1T~`˜*R4oee\ p3R(`#E\Zi7Cgh44Po)\OI9< `.'woa#B5tIՓ5tIVCm4SS畨lQu}.}NmT{v̇5 RN)~ r&ϟQ ed/ OP"3.<P0͍հv\nkM5Zr$)4p8"X$)ʤ!(%cCyʁZVޛe_Iu)1d_KS}ىd:e_v($􊂈)9C'وԖD;WLF`  :X*scFAu Q!.|bNi\6GB:A`Ā=0⬚e?-c]Bb.f$I`~o;E>U N#e2 jJɱwYC[ WXPy(1fܘ#w!zCr2WHmPL䲎p4 D-?k+@DI"cbMoAb XllYR(g,*.(ⓘ%" :MTS(娚 `& S"Ex:r+I5t(R#$1[p!|s"dq}ϓ}*W(j`5Fkـ q.9JNko`Nت{˱Sro97+/&UGx7-o;ʗ@ "Lhc .1\'>4Ar ]|s|&]VBp:'¨5eq>|w'n^ן=vvOVPN tG-zd~d1f2>6͆ o'iBCzYV 𾉹Q~C)SAQ0g; a/͜<*ʤybE"H3Oe<_z%夕fO]5 ; >2͚m&,?z;7ի80DlďaOĬO܅8iE\V(B gLTPq=܃-G`=*̝ۍC|1+ݣ63˶,sc=e{7cuس43utE7WʏHΔsO[jT7}j';,O(o]:EU,NV(hNoNWmEjuWmꪭS!v;102!QD NIzaLDʆeI&%UHlmw˫:n4V^Qu+Ti4yIɍ~*1Gs_P#eWm\Y+F~سmrfOvmQA%8k9id[~(lvR-M9ߐ ل"guGu\}"_*`KR x@HDB+:cXۉY}MO?twsFFK\A,,PABA% }@߃fwL޽s7eH86Fׯbs?xWt';-wwjxOgmJyjN+"6`;e`=qJ_ͽ+3^޸{^Wf _VAṢRkڡVP\ ͠MW9gؖ/+ d U[(!P-!W P] $}=1B%[wkOs/aBl`<u'nPNp LpybMnpwT>%Zr7ܧYx{ZJFmftAanP1޾>#BWƴ<]6ݧ)I?͟q,=d6zK1.飒."OZ^<nnV ^fHqk~}2|g^]d? }{r5$g׋mҨ޸{v8|^~|:1WͿxcNqܝNEt*?]&%_rmt{׻/^O.?|?N4oo}@7 !ﺑ׏qr%:Q8~'#ek{p>vg(q>-zrjxLe;*>r]z SPlmH=ELz'] "_2`P>A (|i2 \LuIL\uM%_NUb[6 夺 =!w^=7W~$ M_9zr ;ӫW8UU-S}w|\?H!/L0Z`SfRԄYH`]eT šߋ/Z h ֹ6-0lfdco`j̵^Tly`i@}$RP Sxr~.X{mʦ$RYZ9{a$ǡw;dHv8 S9V3EL(qri,`0Cp՘bhGtVQ%l%m [cPTpxa mWòӌABrڪfl $ewk,d0LXؕE)QY1]mMڶaڜ⎏E HPCxs =3|OP)"D*B)U+DmrI%W>j_a,: y󨪔!V;H=`=skSdZ1Ǒ=ObR(dl饻.ܳ6$QO3B|آ{*Es㹥<č"n<oorҥ<_D7qRT!Tk|)0FjĤGnl_#>6E=t^se)̿Q*w}|+sMe 9=Z)ʷyg@zE}2[ycNf|3ƌ3OpFG+F\ |=%`ǒr;.tLY,\=;QO GXٖ}BHpa@?+X׉f ^/pQu)Fmp|Qc"!qu(fٔʟzx(9ԇ~ P-((z!לe=mMjqxmMk~Bo ƆC>}@UECe"5'O!Zf CMuŎ\ѝ%.8=f %d)QqK$KrKfr4dcI6dKq1($e)=hPRr)Xe`;m,-Dw斤,Cm D|;fpF})(g 0*0ʿYƑs:UBTBLɿzW__]6U_5U:`ۍ*rp%H/"/" @~ VE1lկ3% 'gK:HrocK>$*dcK6dcKndK#uؒ-ؒlo/$4T!Ti |C $X"<3[R2FhcK:ڒmI}L36$%!!mjƊ,kEyXXM,HBi%f!rY "I.fB*d@K4gn?(20JHƟCs)5 aQ},$d9[섺!yzcd{3.u#EɻO_f22F1yZ_Ip޻??l.l%fzbh.ccGy}]`>`6oL}V>A#eƽ!#wmm س:BI6'/3=#dH*?3 ɡs- ltW]U]$vj D:XDs"F& G88jruA :=Wu]zC u{w&m3X}iƦk:i<*NTqOc59[#U66[Gy5o正bc8~(`2J24ֆ)teAv4whFFӗח=H"1r]$FuHLb% I4$2 " S7M_/. dFbq:iEVI$O+QJ ?$;_/H6En?#w,(} P+eB0ɱ*x+"yLx6yh\jzN7߿37 V kі꿃^ .koUUV[UoΔxƅo^y ߀1'T'5JCD$J(# r%8Kb%Q8‚IlDX# M(D;Lz&*qWȹo_왫Q6Fk}_y|;?wyOa`8^^\4^9;dܸ>8Rq>"n%Q B21:&+蔱KQώ.qQD.6u 5Ś]ơk:qX"R&0gQD-MXpKt݈@aiE/UצF|ՈV.|8 F#E0Fg,ZbsX>qL@5'B ?pb+@"pBp %aR"E 0Jgϋyԧuf(j"ی"YyXx#1$d^<-Rq3}=WDfp ̀Aq"{5u@`!& 8_k|\-7d\ c0/.Z,57xa2ns[FAkaRm_H6$)jW??sǒgJV/bSIE IfnjdayRr7 ̭2ظsWwj]¦q\`a4$XѐK .J, "+/kb'ejm*U#Ht xe̾u-2?43Y9A9 뙷r`n%Fw;0pkj]D^F؍| ;* hj%A[0-5iWLP8kluBgljc}*9Bކ%bLî[{?{XZ[W;F`[AtGAv;banOZ TL矋8QJp”Qd Ne-/I(j/Բ}b)UO\k|?ac~sT|}4puw5Z {0hTVHnwc$j4k錵YXC|Oq7 csOuG` ܿY.$yOcݠYB2jP< !|, ik#Xc{pymigTYi8,9"Wb~VTy`}B`GdIy6`@<ộMrW4s}|7xaxx7AoF6T~" 8mFπѪkEkgP1rLֆuѽX>pUDD'4J5g+]x6満K:[n!hW l  \!blnGzY4/'<Y a#Y q)8m̄.`:ST|x֫#Ƨ .v;|?+cΪRL m ApAmRJ`ZFlh:iHI:."0Qd1*1B"2{HZT<sWTPbMbb`F)ck0"VF0L8DH$ Tzl"$DcWdt |t RI!,Ro6zOW,īFgVhq4^^xE8oWd*+wZ~ nRG5ʹHWyAN Qf 4vPt!^E"vqm\Uc{F,K[/QR˽ -'(eIȘ!.ZPLs"3 ,e$r|/OւIP>L C&dcdceİ? D 8.oY[H6֪[dXUd?ʻS/|/kP>߷va?"&LkW1@B!*fȤi,1M+fz%v S0I 1\kRwMqBWSP-QtDܙ`2]Չi꿷*"}AfjGr.χwՆldB=g6o+K;dvm-@l|[\D w7Ч{ ~;B!r"SHJ+Ʉdմt#5J@$*:*Ɏ%8f\v eӖw[EM(,;6>"<3 z/l-2!C.0!V.D>'1wA & wvJ Ί̻@9+eϵ_bX29RzPU[k[Z} ;>^cT'z>mb@ V(ZбؘZ7xSy7.'EZqF[j`MQE!hc=5.+L S1SEֽvHe4 nc Q1Ħ(cb8BI  ĊX8 K$I,´4a[BkF욾 {MRG=VN]NtkUOO%Xӝ#8ektRD0<$YEg8k%hÇ?aFSz,QGR}W^̠U dXF@kG%8BV8=ne* )XOV*jVV*& ZzZGb4ڪ⏨LpՕi՘dc&<)KztZ.. tCEgIcz~n̡̡C\(B+)g6 LkBMJ`!dFXXecF(4unq2Il$RH%HG @#bΧ,{^{*17CwZ#LؽIJ/3x/K L$vrbwq8y+r,R2ێίD(kG*慸4OhC(Ҋ}A7UHs6hhdZ]Z[DWa*g^JcY?k3$&JO1RJ4(65BL W1RFKX b'JVm",E},nӐŒwzA Ix:M2ۚ'm;d=A!]2ۚkGmmHW.eJ#ET$-5,P6w$FT%K9S{bH G)_-]xLMƱ% #4]Fj{05_@ =ɀ ,Ի+}E+R7u!~ĥ!"!a◜]B7|̇OXvbl|I\ZRkDwԀ_V>\Sr=C88EfMNΚM2S a~FE";4Ǯl/8*MwGj``q<׼.Vn@vI-Wz1⽊6&UԽ<}B @Jlƚwf>b:gɧ^N:ЫϺGcVx~Wx^ [ !x0xyPe7y+N ).0ih7vtH;~͹21 ?!o tmA歷bh,1DV!#$/f|s.TQ<"ׯg/Cцbm<ѝc ';Y׷l@sahs7xMp; U_}9(LŢ9I \3b+F0󮃁` (M6&wW7EkRYT_kwr.`w7^ga?D.s;u!|:|Ҥfz|:u-dl2s.i-N2.*ME]Jw'4eLեys\4դGڦ62J$J;,RW{m.i5EOYL~qz3.sc,""H.NyCmbhX!T𐁂e>~~ #xHbLQHBlEb5rq+/AeV1qoUY{XqUtaT,)W xs&`+&ixK&b &b2W;#׉? ʣQ.anr| *79DtE/nltqfdȢ'ppnZ@|XXLiS 5w=ao2ٷ.:jf@I p)faQRB6P)m '&.y+ 3X6UCwΪqgw]vVۀ9k-!$DW .7B8$3 P@j\ q۫4E72q[Flt&`~)̓(ZȲuU"RaV&XSD8TIlȲHZPŨQj&$LJQU(5'UeEf+^tOU뉴L8HW" 3-Ɖ24hz/0LJi ¼,(R-gPh6>O܏zg{SW8%C]{ e0=0DyLbWS1ޟ!RqWb?-4D{gCƲ$](,p5CJ$La]-my[LE1E=&TIe0%st ָ̈́h L)Wb>ts8I/gH4=R66k!p >g#X >(!~Q\+mDU_U`߆4vGw_|;ES{7<"yFa oR֋?ULUgSzaK^0~W;`&F; J㓋gMZ#q2M׮Ƒs5q\EB JE0ظ2emLkmH /=~Wq>`EF|~[YR(ʯW=ɑf8Ҋ&"G]GwWQeWv25nzuNM~xBcx|TCRhW'Y7٩N:a}!BҁY`Rk2+gX[C=\;spDqZNr?"mђ޶hG§Q,D#FigH-4z8c!d@)x@ABB'{@TW&B$M"oIktb@ )JEtHlTs7,NVA%3u$ o}6' YWc+Nbu}6).k \WhQZpkbS1;e5NMWp}A-|K8k&V8a,b8wʩeNE1e4T3IGT,xdܢeuQD$A"F #bAbyTT+uv aŎU!8ҭX4΄vk~bt߳8 As \QL4Ep S4Q;m/)IKhDX0t*LGD!uhI5Y}{!rpP|Dp|jtIQg0H꧷MQ1rjrjE?-LBcB-)RdJe8n; VX^ S* Ohw&E-)jcRm)jWRځb/T׏EJr,M)$SɅ`E- B5f zYpE)BFn) Mw`Ĥ yH(0f"@kN[wҙ>{p 3qg.ɦjj[4˳[+ߔpV \'Kݛօ P f)p yF3Ņ&Y=ʦ%%s:3S\喟:XFwƞ$hG J߸߇l\_mHwPn";r[[Y1"_CYhp5rHC4. ]>a+7>ˮP pr\+h%Yaz"*IkܾaRsA܄T|ػM"WE6gQ#~PA)ok%CǶ"yJ1¶f$9߻EZRMJS=؋ԿUV.ƹWwTe6P]6x1\Io??oLkf$T6iR˵yb^o5z.k3C9[w?n=\VZ`PӬk"efRUAE.uQV}T (FLL> hzztö=%&!W\UZdj=͡t%ӽFֽي^fV'tڬg*]'Z':T9$AI|-#d>A%>N6jvq4zIx#X>@-vq;38 (7u&:t&͐XB֨jC >p`cHcgCΖ0^_]V{}YL0ѹ㟉 ZoG.L?81A[`Iv{媹;tA.=Tzօh',\d1S/E2%nD p+/ dAbkX j]goCR!`aTJM r J)h|^uG#ǢhRFV Xz*8υ8IրjuJh6yS]dkK FQHPUZ~zڴ(ƕ8[mu8di^]Ձݿ~䳷$->P'|)+)./j[z}|%q*MHk&F޸i[y>~]nT;?6_c|AF\w;LYz 0q;]^iaÐ|`3@kZ?_]<=*]=< ,VMGr@3{[fGd6ǰ/ wQd:ZwG&ںLJpTm)(4iQzo q1O_^69J a`@3 blה\(1irBHGRN˔cǫIe`]0#rmc#vmyNW\-^lhyC#0lDc{SNz_ئmϱyO0A4C!5s=,#!vjRfȞ؃b0+R~k;2ep3\!A}0Ҫ iMtY@eX-:.#@C`&d66,&CVHy<^ys݂eLM%t$i 5h[rIݝXΧ'+NhաmQpV QOϢrdNgI$wKy m_̧W/Β[Z: :x4)O ҒF JBߔtn+ P/6 '[^x͒7[wJ(( fbiaAO5C #W*[Nxs{$Loտ7۟,q/T4藓uɋ1_27ֳ!,3Ln)tZ  0zB),wg4y0`_A]а!%8c~$qkܛ凶ƨO-nU6inU6yGBe49b0ȳ0Ηsi/\yICl0h=)Jڬgw~0gh&edM:?\} '$$v՘o_g264+\=:lry /mz_WwqKO-ՇҮGi<;Tf̉-UF"\1kk",l![G YgFC• Q]&wO!FJV S Uh\("בk@ `N CSmU|&Fŭ8vr"E>IUT/y( jkF`3-c.) U\$FZ#Q6yR(hjc YdЅlQY/FR1Α iXa#W"Wsh mB%n S=1eOX^|Jrw`ɓF˒ɕ컼߾s!!ق oA1`S 6i-$gHŐ5Br w.$QrY9YC`]3Mb*yI!iM&vȅ\ޔoݵ8#75Xsl3*zޕql"mR?Nf0dLyA]6"}N$RUWuuVsrͥdihtY %ъ oE0S\5kIxG xd? #-HON󅵸s#PgXMI@_v #c) F"}c5jOyE11F5 0&/Ht յauC-@L53AZ&kVk/<IR,td%]']Hꟙ~`|E`-d'  5HA`5Gu#^נA؃&X|N!CscwWTM ܊k&$)wϝ%ssIXv aSХW %*1=eB'Iع}Z Ry'+YB1!f*k\㹧Z X7O9EAԃOV-@0rDB=ÙF5AE!:1CdK;:-aEıc:ج 6f͑f1dl LņgxlV2opج(tYÍvpl(O#SlV #gYP)6+?M2,i;cÒ6GeN'`$_0 0DS)Wj[+ꈩNيK'8ص5sh0`WZ*L68&!τ5Aqd`B ND@`Q.K}AÝdn=)Bw(7zHmط)[+mO#:B"0 R"𹠒"B FP;!r2\Ath UQ{і sh L:Sb tkl7 nh¬)*T @/"bǘ/ JEMf %c ṰиMNI0r̜ 0lTAoˆ~YjkΕmlX ZGme=!$`kbGo?q)ՈV^x Սe5kkeY b\g@Ĕ_F}(LH< ' LA?>]z\ubwtOԃ?>Qb9r!>}q1G6YEbpXۮ IG1.N(:rwB1hi9>Z9v(O#םPR fwBPN(?Oz^-,W5wUם*== E|j!xS)-YRti%Mޱy~YCeI ywG%\;=V5kN;n1Mۡf` Ba;Ԝ1 #]isu96O]w=BgYpY\<:e֜\$?xOm~bY\ݳ~,k?Ԭ1eSrU݂ 8׍/3iYZß2u+7{rրta6,x|$c9᱗n.ԫ16g9V`=Esd$ts(0y` ahQ9g9y{b{ zԕn.{)|QHhe,Qhk|ѸEE[b4K)J{W+,7Ut%fl][|qq\uU!m'۪Vh3%1Ui*`.1Emë*\J{@|إ"rv`& XM fXy IWpM)7RK;BЧ̅Ҙ̠YKɏxtB#uUcL&Հ=/eRڦQZx P2ˍV s-jpQWcEJXPp>)IEIG3Bh`簙HAJ-utKNI79hLk+#Fq!ڄ-!:Fup v:mݒ n9$;$zkcvTuKA褎c&ܷʭݒ n9$;2Ddd(mD{% {M5j%%pIs&9b7 c0'K넧>E%.Uy W@L(rGt6W@dL ݀u $c9!i;3.(c`D85BJH55BVIFpICY"usMVSQnq,PH j|l"{N`*,Dǵ%WHf6nOƂ%zqP{cq,mkj ,[7mLb2dۈo޻jKq4ga 4u34' Y2zIp3k~ OqՉ1Ӥ%&-Q,g`닱|^~$,jՉ1XOdBl4H bs'9rW񚜙KXuEn3f-GdFݠ~U(d_?c̑Vpd #SBs2A|7s.d6"fsDv˗c_6Cy5E,ZzJiرJiç SǗ(%u&XvUc@rfRhNrz VDrhE8P ;IvIw gR1 T@҈ʹiR+OS=L9w ? ޵yO#S8Dg~.z?)eX}|ž&ޑ#0>0 KN/z9FeNZ^?P:sJ]~TJoT&R# KBތ&A=˨s>) TN: ڏ.o`3Xk]\.oCp[;`V܈.  =.?6Hs4*/~juy_ZI.6qm.Vlu xǿD/`Xu~yd^d_=Ǯܾy^=.!Fm.u"J<,I8o}vWMV& 4)X`Ĭ^=xP I򀤇mFH+l?چRyo&< u}2ݷk-Mc &(ij6L LTi'Y ec Ia&_E.Lm п}jCb[k WNL!>nRlDo Tp@ 4k_wX1Vd~1?6\%ؼc71XtDPbG|ծkX'{Y;^V̀Nkjhe)L ϐGN϶IǪjɗ2^QبNVNzw\)hۯ廛kK')>,-)j?nJ0 =p59$ rZ|Y2E$+!NB$:9WaItVڰhSVqjSh>1at EX* u*,٬r8~M蔫_S29$9l BkK Oݶu;V=]iT])LB$ yS;,E)|d(ZNIF$l%a/]T(g=@Y=%b$S3Kv{c3)8#53XILPFaˉjYGXFz( bL1KBi(ڄL6vtD`8m8Q`MWlsPmq0ͻW&/_?ЌM#ۦͬzuyh۷2Y cW#ebd]3|e][-W]/?KJJb^d)N_R̓0k'| {ׂХ^%Fq Q1Sa]嘿*@曓Z UV–ELn*m kw~l!3Q>ʆe13[y8Vqʰ{JUR#Rګ1OMSKa( KOi{xNu  lʁ[wIS[ߞ[6&y 2P/bSq]P3,Lyb[WՂL1uKz֎1H3#/j4yj˝ +zvfuױ}is= V@L&4 Uʪʃʲ^cUg*vgê9=Aiʖf[\J+99:VHIUmA ۤ_]m~)XӋx߶a<]!]0 }R |5?_6y9jN- %l&70{RVVVi1^4OQەp" a3S*PnT~Up,D@ Bo1>=M&@YphG=')<><Ľ}:{]) w~Tg >g]c?ϓc(wr1DŽB<' 4DaYz`\;gcg4[DG'$Vѝt]tݵbHewwdT s׷_ߚ~}kӦ_ZЀ2',k\Ps)UK/$ ! k" sR?E;/ rRhKhi& C*`Sk^R d=Nw+U[xHZ ņ*s,9'+X8{rGp(F om=^ֻ>Mg/{j@ 6#M y!'ZzQae |gep4Kzd8u'G0)N+~mڪԞ3+8}2fQ/?>w;0'%VB ,še6L{~xzyM-r#{,l?laR~VE9q>]C R@8)s6`}ESTHw|*F5 +ڌ#Q^΁m`&hha&izNr%YE'/h((;!cn%s u(,SKrzsW43g67a5o1.VM'wqjx>fk1\#V#$ǛFYLˈ#Ys2f̖IRP&y&xK$*A9[:hY оl/a+F+/4K^|Tb>*KZyEX xK裱82eRTL&wuc||!t5Bqfr&nuWظfxMrŐ1+Mw^* u<=FCcl/ff4BmWyuY6Vm?B,iR Q-j]Yc^Wwh=Wnz]CLIL]m /Ed&TԹIwW7h`ψcт Q~<GIQVbu}DJeᕘ'57âBQzոmo4ʄqVxc@Ea4 XgQ5T/v8UJ:M910NZzFӕȉ0+56kכ pĹS\Ӏ u=(K|{x )SbNE+d3WW=Mˉ߶&Ke|z۪1P *P)$< C#Hz}An@RBvV<(4cS~c(Ӓ(o=_ BLUJY.`dx01k3 ġo<{su,۾xQ ߑ[L&^C}߶Hq ;)}~u tPo\-xP/IgNgGhz;͇C`,DH!AVvnNH=x%3$-` V9XV'K97V !2&#bMűlƌ6aŌ CI{o[hlBF)!UtJE)2iJ [4WZ ^ wT$jYs杤La\ϼK+>T0q>E$W-TP)`(UYtĩuoM](-h UXt 4SHմ7Q%;/͕mQ0WSV`j|I9{| -w4E=g=3d:h&f͐ M=!R< Yi ,ߊS,#)8o/W;O6Alf&N7P= o|xo~וaf0xt'&wgdr&6? 9UV9r"~?V9bLUt $݌ƬU)? Z6ҥxW"MV=*5All+֬%&֬|i4,-{,{#iLa\FښgB'8:C =ag*s+^ŗ'XSeaEy$eC{ȉźk;teۛO.o`>h˥rhDT y0$ .  BBP#_H)O2BO"PPń,5c0zK/p(W'IS>5!f8'qwj2 kte!uB'[5<{Do TH|iV&<(Uĺ=((UFuQ:@+E!C(4ÑG@jZ1mM.Y (kWd6}S~O@ʕu_0!EjE~Gv?qkpQ&ݥBkm9k3"tT.F-3!>F "$P% $Ǎfa̟(0 $UB+Tīo^.V.J+/'2iz[ę˒4s4-l6]r&:o?c>>ttfەu5fIz "wt`Abf5p>< %O xm{1 _l_(JgNt : UXbn[]ẗ́}wg[t?t[hqR^_{tQaWEV#4NjG)R)/"ST%9/""?Jb.l遬"kcV4>OyHH tKWDcQ"p*z3('v\xqMMi*@ɚ+&Bjl Rk/D fCfls $URBmT3 y!&(ZPͭN.>"d(;ӋcJŵQ σff{/D%Oޒ`8Yf:H5' kN#I(RzQF5CU)m ~٨Ӏef?n]iBJ]z2`UU hMYq}>9n83g:0f ?~8g/+W+m6X,voG\.󛸌ۏf>=gof`>EABJc s;?_n~$< i/΄fbhގT!*o4sz~5-:o-V[Dk0y{ӱ;+6|&)||00͏f:u?ʅ5LOVDL +yquP۵<%\<@@Q Wb]ΌnFfP=11r= (מ }qЇ`3ǥ]m#˿҇ɎH}VYNWd>LRFwbLr4.} / QV" wГuEI&]p4 ef9*ZG'Lju `{q8U"@‘pݚ|BC:|ژ`JۤYNx뗒$4 NhFqԴl˙'݅ugq<4t-cJAm͑;I5~ Ƃ4e3@jD ;}.xeLjraGc"bcA[M\zYbd#,fHKk4kY#ԛ[?R=s q A7_d'Ʊ4ve!RBl#@TnpN`Õ+ Lpq u΅ R BLƜbvcca8\ڶlMr@iMrkf /*DT)#uRtGcK"6rrZ.BmZ{7u.v9/ܷԪdbBiyX|9#8CĞ˟'cꑧ%L'rb ~#*Oݎ QfjرQ }4ƙo&KJTE@Eߛ =]w [f)$zis04HsgkF98"l%a؄!E6⽤;u$| AV`L,V[eA3 :ZjNsxJ5- Λ9BD5ϋx(EMSsP͚AVV,[m(TVLΩ(^0 uLxh,y*P$9b1P+(GAx-#c${ e'<RSo"[""d j텓J@-f8K?j:RBJnGC-`d@ͥ@\ZWqhc{$#[#m` 3$"^`KH;1M b3BI\k S,eسlc8zmjv8뻴7ƃ.KWKJ_l-K7>cU 7wC?I&ӋO)K\ZJ0C8$FiQkǽ:.#UXr4T#!_)&.nKneiGt>v;xqNVѼڭ EL\1ɖv]VyD3hřt[iDj*$ R2/?wݘȜrc?w w\=uhÙGB~gPpadrѠ ՘Mj 0XZw-hzp.Ÿ=!<w'nXMx{rgH0rl >BpyE#PјJAEh.O@9e7~ M*Ӌh8ݹ_dozGpAݏum&˕\<}0)9eD/ *8"E %!n87prz֏{~9:YP]Vq{?0/g2Rx'Z~KDm_B&)̛}B,6>BQFäV`v Fҙ-F.VVi~YIDfBn3? 8CvP1+@䈼?x=M‚aִcl)GB CT 9G|?'&9ܿYxxAHD iH*qqK9mj &oТ.F=#S2:4<1Ws4#'IYEyJb,j%[ pIge;ͨ~ІGp["#" Kte; F(d#[F[/7a)poMcL3DCV{Fem{06:Ӽ'>=WGKPν<ʽ'YIdi*=զTC'3|򒝪'~Sq 8%E$s% w& PM]y]RݤϔN9;+U]`aѨ;,H-bC'R#Qc ܔ o! `=HE`Ic!GU8ٔ[ g@Y;"ʰPF"-$#i/G>3'DlRKq&`5 ?ŸL"I>S/^$Vf@1 {ڨDl:Eqȧ3_% P՘DnbJe 6`4Vb)L9E6VT 2rȋ "]x\[k !)S)3Z"t . yNViH8 w ̉oOԯ鉚e>*G>g')@ ҅Sp6ͦ>mII5|ׅ ø 4_eܳ1 $.'{{v!CUH#%O՞!Ofiztuz.ERS,zS8H3Ԫe%=rГ.U V"l} A-VjO#̖\Q䭁%M[IXءIi-xOlEE`?s-n|s TU[iwo&,/'mNTf&6o9(Q#6Dr҉t)U[俞4~wDfl3#޸<ޙC edl kbb#3o&z  `=ycDZaXe e ȯ0:D #T}-l#/-KQSp=@M0c΁.IPڂfL5"Mc2ɑ*4)^L/'9DNZqG 6 wI][T)I5RװIET" :ƥEP_C's:tIOE=NdEe5Zmj9\7ZojB>`QL í!I Pn4KV&:pX{ӣIa?6C!%S"xJ~Aگ炚>NÇA⭶?~켫 .t@MqNdޜ0w$ّnHFshO'SkIεPQaԘyYok5Rx$,:PZ4Z 0x=1|}'gjgxVO,@OkAb(GׂL}|ς8[0i\E|>ThhI|s  5? !Wv .xA1T=&8B㠃`v4?`t`c|WTƕgmރK; Acv眉,L;H4Z]4|d=&^x8р&H0AҀ1P<14v 64> ~ n>mP`Wr~"+NhQ=:]X'|~`%ns M9-e w4l>ltqNz]{t}Ą$9ƈD,s{])BBjjcY=Ke8Tllm}j( ױL ػ;3]T_(U@,,)RI"h(5? +ZQMSum %aT7^'=2Cl6#YX~{mX)Yh0֤1팿ڽN@1 !FS\%sJ`.2+ 2XdqIPP;f^~qrQ:)2#;z^/Ah 3;lWW4t{⑰;;G"_JR!F`U}Zzy6CٻdW=m r xd^o5IIԅRS&ȪꪯM ] )L!͛Z\UOղ__Xd@ӛkGTyެ֟Ȋ6po z(4RhBn0E^lv9L,_'f^j6 Gh+ǃ)L X0E ȰoZw 30UqcGQI4cog>۽5Lوs{#jF4DC1F9S'prKp:%y ѽn4TgIDd;$[*yY"y;דJI/rIIxr:\_Mfc̮ka t1Ԟm>Lt ײ\k+?WӦ`V3ôsc,<셸qԚ'^X/ƈ E/'Q˶ *s*R"QQR>mBa9x~ {55# EDH)79:}6gϞ~ԑ&w.)Uȓ1|<] гh,wa<i4M׏Q-xZrɫt)=ٟ5H]{%G☚3|9 ?.r[>Qu=+}`ON!__cJxe:c`wC}9PKuI B ֚^[a%2 cnN^lz0)^<]{.?>f9m?]E#)POΪɛ=͌L'SfM µ6Ftܹ r;$Y )wG3݈`m`P/aG&v{ Nˆ`ےn~I icRǨiO0sR6XxyZ"pN[ D3_kh9͇ߗH' FZ-2 DJuRD(Qtii0ƃD`Pbg'`Cn-|Ф*L|]rs öEM|Zo0/A3&;<7J_St@#|?'>a_Ta1,O_'"-~qyD3"~p$ #}͢|: $s,w(k|ׂ^cuFcJ?S]T{|TcO暓G;-c$}oE#'ᶛ{-]Gf=.y,S`9StNir>OsD靖@|'b=)~l&lA:ēY&}{r\])NdB>vN )SO9f`"`OQ}{@5A^a[LHν+nAI A=Umg^И[g/ YJ͕# dw1Ӟ*'dY S<ly9p)ei#f1F|Va x[ z0(j l|نb eIJAcBT/ '͈hi(6 0}ж;SDa5knKC \iO:Iu*Q|<̌6}z|*"/X᫳EIϹp[fu/|ΐ9 G0-EJ$|Dn-оk4 <(' {$CWED/"B ebk3δzUOhm '1woO,rJ pr!$AalM|}|9KrCM'+mh7Y"-mWBHV9ȲcDƪ ;+gf>؁aB#GmmXu&>0KNYtSD$ުLSS*FqB`qLZ+2M`"6ֻӬba$7ekqs)H@u`9UGWfpQ|=O+,)pghqEW}*ܢX/E>DQP_.[ KM]xZZ{hN5JZл“(8p]+t-0T1{Btn(*D7:^-sp5mN~-.ɦ!MI}kYH%AWzrΰ@<9̴BP"32٠ZQ!x(.D:z|z<\b}v5broX>&_(UX#s2چ wݏWσv^rv9a_țqsTSMT~m{KQlf&f;S 4ǔ~s41mCj:z?%3$qS}Gpj ")v,j5uPFq*?~;d2Jp]u jrhR,e4ÁxYf^?[pIC(#(D8 (Nbƃpɽ5JŒ#2$BQK*'ARJ@ȘT&k8EDSFJT(W⳥htSE]9^O,s;±1E}`^l'r('c-jFMYiy VJq;T#͸g_ ]v3YrO"PWQ7u1R>nzӫ4g>`VlE2c$f3(Oc{ 'J`&kXwJX"VOdi:/ې1c2d@"Xf`n ^}Wp$\5ߝ?;w+4p fV԰Ls :ȂYΟQ03m79l:Il fɗ|tp1W3C =ZEGBq5XN"DOI2S٣}>]=&h!h o&VɊ0K̜Ñ!Wu}wߟ(>+G8 RMœe})eB }&Yu(;_8Y 5{#dVevc#I U"z1L_ITFTH*%yƃgO־8-X~?(_$=Izvhz>PS:;9_RԟnG6WH_lF}}7eh3[}; j"FfXi8ɍޯ#4 L*o$&;ϵ3=.^ŰSu;^EGQy4{I;ѨMc|T¤V?)Fg):2etwjSi%"ٲ<^Ĕ,KDi sa]3 ~3=Θ,ꍁ6tHj-AKͩ>`>XKWGU.Zi}LUX`| 1r U ]=$4fxLy?T3`A]e&T y: k1˚SbL")3R#,-x\ p` de5[_5G .G*f#N8:ٱOTK4/CW04i0jpփnk5֔VKa!R@(LX8%:DkS%d`FT&a!r5uS;q +1)n& `t'^̭dQmYW3J$zVĊ%̲ډ@+>n (CX,DSPRJ*0Tp#%S0?n#QjDZׁkup>$XI#F^7=)le܄1;ݥjeŒ,ԈIT"ML-N.hr+|lB_.@b>D_k9:J?Naa@: h4" RE(ic[G0ƃD`P'S Φq:1{[Y93;%49Or(#F<3Gk'1up1R$@ JGԗsO+l_9_d}Pbh8ğd!Qt0gC:4"9Jꨟ_DIM5^l_ת$HFH}wäN1\_"X lvZ| uOs3bGK \ $h}|5}YИ]Jթt9&,@ۅ[I&BN]qh_AK ! 1 NCNe'|xk0ee2]󯿀77t"$`uHfKcLjaEҪ>3bGbFQ8=̯aHևL qIKx5]XقFP"t 0Ex~xoRNPΎpktG2ծLjl1}FѦM[[D+pf;d8ɻoyJl`?()Z4D҇ן_p]Km` yʖ*|>jvnQwk  1 i`T\'RsIKqUDso:ʶJ 4(֯`X;oy7L/"iؙR&Sӧ26"`DŽj8h*Vidp;𪌍Va5Fpĭ AxƉ w+Kl#3(/$iEݦpe5g_JkP'V9,4(oA^);59ЯSv&Ts_2ߏ` $ Q.`gKMo3?k7OokTF0Z Mj8*ƚ7Զ:8<,NsϥIyɶC҇~=$w 9%$|ajm$Sox+#"xƁ> U)sA.4\AAuG;EhA3=|f 7^%J3a(VN?E\$&`r>\>%$~?5b"(jxFWNO'tT7X4xsVќG9/:" 脹h4 ` ^IHTwB [#a!7$;52hK*BSp/$]B!qaAl1Y;`utm'$REn@})0 p/ _mI@ځ^.:%#1w& #; I(#F$hgvD)BXci*t%n&֬}7ܾՐB)ah2{M%x y@Ljŧ+ ?pw d )Т# -ϩԴʼn3ds(dsqƬINegYPJZG:Upm is Hj/Y˸7O`4*.k~j.P鴚vQsBr.|Eg̙IvNuivٓx-Ŭ]dI(idi(J~裾!)'Θ1RtiưB"hټknԀ}@t`6ˏ꧿4<Ť?V9G̏c$wow9,8c$MC[&{Xx|k3glmFçl 83ik:CفYz ]zM=ҤxS[^sm@t JRO҉0rsbAtoYϠ魚 qƌIqJ ~f3V[plxg63^Ǚ ә _`W )," fFhw-0! JC8' Iwψ/kn@_\Z.o׻ڃ^\q.b`W@&݅_f_V3/Vu\[)u~%Dͣ:.`)\™ P3Z%sj_jj_ Bۦ3agR-S Fw Qg=P ab ˀ9P&zR'*t [-sVhHygݍ#&{0?E=NJ*D,nQ W)PR#3g3 1R?qkG!eI9T-kh{4 ‰pQ8jyPW౩ǜす*è:Y]rD8]VQTBUa땎ۿZi%%Tek;BEƳ@k [%QTEΠ x3 'h#ڡ.0瑄Us8x-ɘ* KJ}:X%O(D+t 5)`<qTi:  $V@yZ_c< }c Sv[κŵ\C@<,ncU{J,'AHK2*pNpLTb:Ӟ4Ӹ.)\R겉.Qp!‘EksHw2Ts IH:h[K8pW!f{!jWiH% |G BTں1x)'w,1ǫ e8),Y>ंܠK;Uĭ'e!`$ *@1hS| )AgLӁg޳{jj' X8۽q3ıa]],wvSQ}-yz񿷫Ewm̓Gj;rcOb%fGv#B//wnxrwټDMxX#[) Y{7U+_\6op=6 no*aFH32%"eK&6d_Kdڸg9^{F6RN|q5ڸ']EZZ\3J,t'T,I}a%i!!\Dkd`={|$EM- B3Y,FfgH3*}-&}[&]-B!Ƕ4Sb%",3b}2(RPu€s^E]`,`L=X`YsER}(GYAxEynlDᄒ#<7O@BT-;WKr!`AV`XyFQb:8E y"Z#So>yDP<VKѩ=DMEpݪn-$䙋hLa޳z@KoL8&0"t]˃aq zeVP5ZH; |ISу-}-W^tʃOo nǖ e; . BDU&:m ? Ar1'4'QyFēKǛxRhēELɬ ;-?|Y;+8*q { ]cA4';K/נcyfji~] N#ڼ yŏ? \swwzo"$n%|̬ xaNs  p_8(aInK SnY7`6W%Ћ%#p% M~K$-4e~VEaU4njU D/:Yُ/#p%߯/c US>Ha +Mo $^izS\ytS(.Vbr)K4%n-ibOh{Ğ0DjĖ} X#P5ֈ-Ih <rrv3Y=D Ia56Nz JIyq5JSG kFpC}0ugY{+ZX$3Y$H6YXI&+q[[ݒ_1`Kj"YV[ (uonuA*McGbK4? N3j拞~JNuuUhS0P?|ݻ.kZ}As.&F8JbL_1x%L};OAkMx !B,j~U ?6/M#{%[%T8~}EӋbhƢ svaZZ6d2MzqESr~z-+o$_m}\O9UOIp9]pاٹAhY|8x; qJl# ᆦ\/_AD %=Syt=ޛ_'(^2 0@V FiKf@*'zs!I#E䜗h^J|=2*"D":ɢ3&1< E%ozO}0;h;} o}~Ld2nj<޾|Ae7e8 fxzωȐj-_>?n>?=0xKm1/&`%ѿ{{4HlX1^ǣ7G)/c'O;c7i4ΛbA&z>9PbۃoB E-x47-\qݘ8ѬVo6ήY%B7Ϸr" ;K`m~n~Gb|p7x1w 7gkC o9fo2B[&M7 އ|zAjfAo2%kdC!42ܽ5}fg'+*m+XKڽQBΣk V*0Pb7vdF >^PL\Λ{ ^*Cf׼dQr 4& d0fnS5GAit:Y,7bf Nو5}ϘAJښV9G:Qd6w|?m0*ā 4A2b`cVh0(D"Ioʪ_$s&ei.0GEoXЩw윻jDU׻Qݽՙw!YpKApaFόu":X8SJr c@%d:}{c憅=8z8J=5OI|xg"v|Sp&Rk/g?]؏Kx-6TeO_fa-T_Y?aJ+vFb8䛗Ey5%Ɣ^Y˳,gY.rOGW'˒zJ Dヶ;#렄HH}{q_jr/8e...~mv2?:,n2rYP8 g zBsgRYƕVXH<$R1ْ>RL KsĆU+i'QBj|h(OFtzǏ Ct?{վd'bMh(sN{.ET`$旂3ND<XH ߩ̯g#=dw&]({~\͖":L1e dv 1YI%&lN;ǃV#YӅB0Vrga/H`Jۜ(f;-GdDZ-كQSیD,C=5Ƣqͦ8HR;.3G#-yQDm3K )̊ ե͛]26,v!bn0hDEN3NbSFǨayH\$ȡ(d'Ҋ])Һr!cȵeU~ktҖ5ieURJUIʥpIGi+Jd+{-פ]3&*iLBCZXOw F- 2DǣQtBCRetVH|(fV*/_=PZq.G@x@E҇(}PHBMB`0(Âb˘K2"aY*wπ FP/"8,!3JJHZkH DτB c2Hm v:PQ1YI 3ɴP Wܣ9*('RB.+0DLi`9*p)$Ղ#ڍLfQ֮BiMk9kngmM9:7}tJzK,Ȕk^;B$tcf0:hBޯd2X7J;6$d=(]ÆD"Wڳ:ylV$-coE-hhUSZF9f]H:BQ5ٔŨ|V 1MJ imP74!"&%V'Y46%i%&IfmX x@FETCӮzjCk5Z`۴ZhUڜJ`i.kxjȢkRC pґay8;䆌EcpA2j[.)0mj4+ Zzobwz6$ |Th!xb CZK!bTNt6EƊOZFx>/-Hf Wr6 %bM{Wnb܂~59n=9=bF?fBEXĻÙ-Ya=27YjWu_'XgsLQZ)gV=T:;J[Uun`msp_ +\I0IA5W5szbN}9Ů_U}Wc<D1C0&z9l0DދOVؾk|n5nhwta5_|rfnafj-#ͺ;,뮞B4sH&yg4 @ TNj8mx1%)t^%`vS׮/IGJMC%i븶Uɍs tFHEQc{# (fcp9}D3oK@Le+USYIгtvbzvB)-.:N.k4O燇S^T?Nw&?f׬ H=%m_Ƥ 9K¤f6*>TV+{bf Qsz߉ۃz|maǴLkx\2U_f4uj~Qٔ8-U d?-5on -s#Bm:h\uhV_»;),^ZG wbjIÇNvjjʰ}&{-&>OL,@k[m]hJa0Ǎ!j֜fM[4$ԁGHf}lś|ٓ?w=D &|Æ7L!B,jR2٨4g(Y-R6#S ƔNĒ8J& FRxzZML4ܟ0O1Fz42B%Vg1贉&'8G_Hb$)$S^ޟŠےsQQaTr;H&rfITN1u8軐% Qdڝmҫ>* }#ѻI~O'<[JgWB@:͟ Pw$5g?K/A]I7l6V;NcE$XA;xBJRI)vrn0gA*2 [ ܱDA}˷g>_}08yrEY^T1aBvj ba,do @N>Z4I]m ` 6يg⋷`av8M 2c!ս$.XYxdqVydߡuʙc\ ߷$: ]%r 3Y@?}N?>J@֪?X" NLjlfj/G p#[7A=rzo)9 ^7g=뼹1s̏~À#cUw۟8G54o=x/QGmT;ss7)ϱvL|5CeZ<=8Kq*H/e|eZ))#ƈ`="h`!x1 ,_%{AKÇ{E c^FhsPdtfi!ýc8Q_SA赱J"5xV!f U G="t֞reUv#˦!湫єC-\5$[צK Qhצqün.>î}B.@%=w+Zwv\R3(Jjk])H.z]մ`ӵ51Z{C" 5rsFb9g(sHk&A, }*&'u#6Z?!)Dc~Nr:cFKO=t}MIWut5e]pRLMD{N]J'DefN it(*ײ hMQBW ZF+JS0fE-cR3<-/|\COW%+_}Wl.B,pU]Բ0'ϢZd=0geNrNhyc|JsFh@9̈́HS=wZƈuڠ |\jB^nHroJhZ1Y@}L\Op`.H[N(^+W'(~$2y!ƔC-}'k vgV$_2P MUi]e1B+k6@ТDž hA"aT%IS:8 Cp1UF _G 9Nh^:1* : 0.~ŏPQ W vV TfݒZI۵L7Xu~I.;hOG7i4)Oe2^j椟7~3 ]њ=q: `mu/;7*IUЮP=zE6ItҦd2ŻX;Rad5TV[_1kX(c< 9m[q&)1Ex$Ӓ}^߸%':JF IbJ?F?8 Hb$ O+jJ<~գ'x?{7ׇ K52i&tV}L\WߜRnT>Zi=֖{FBP( 0#ņ2qdcm%,42ej`8Ӹ[X`V#.2TJWsdV+os)FԊ\_G<Dr6imS1V,iAoo!@åbveWh\ze= TVӾ,mTCzn,6TӹF2hMų+i^v|4*K*u7<s֬g/ڻ2`95Ȏ*WY #=aB)Ƣm֛zuf#YHQ>|ʑ̑k 6rzʑoc&I5TB| [bm@kV+6*%D ytQDid ՠPQ5*jZxRx% .s62TC?PPm2af~d$lS՗U=`2 ޸䀖!`[tCP-.0ejXP$Z7ۺ ~ߖXvUp (_#)xG,J1Q`Y] ~ol HT;xrf# 3>G"@G'#07{F+ͷcB+|TܨsdY:R-QRAֲv1Rɬu kYK%79,v'gYhz8?|@$:ALNDjjp)C B,5]qVO6\_G<yMSHuH5ȍSHېf+^\H‧];ȜS"`#3ސsF?##y2y$$⨠DG,mGZ"sS>GgW4-ǣNeT`#l-.G  Z?"@%AWȳhL|>y$ [GLf^Y5oOV{OʰSHVK^?K-Z8,7,?kT8kgZ\ R9 ^]nE|n[pE[s*ZG\`{%5Q(K K"D A{v*nJt0k"E9.L6~NyR"\_&7=G :m8oGB_+Q%@a!n^!UGy - u(HЬL&)oO~RF&UM6RB̄Vɠ'qnQRDu?ɛ6B|NފO~i}ɛ9@KD TDw%ԗ*etk@z?"zu0;p<6!vO;דҁIr,/7Pskĝh%񵭘sAJWߜHO ;C }87sOgJxIO+RA6c+Egk9ceqC Lq8n]|ϭHA"|3[+͢cZIv[so9f+?SZɡMN+t! |K/ (U'Mۆ5@i):K@7$Jp3N&vAeI Ӝ=Ah1. !GmA1T# ^O#p1g:B}*,..`%I&e ?ߧ92Kyfix6rǪw?q$EIb\4% nzu[q( ` P{k΂eooFy_̓*h6D{w?Mia>.?]+_<ދ>`>)s#ZE2/xNj\S> + Jr +%' HΝ&LY^ʥ-^25^2G>,؋d^$ "YM >Ud:0 X` *j##' hm {}T>J[IpAL2dДZH<8'ϯT$CL_E2ǙAt!4!eϙe͵%db cҼ[,I<"KqXGkFpk~F*8 !}(B ^j4W"w DGGyBeˆC-PQH,7FW0i) u,ef(IƄ1VϽ) }ڔGHyMy) nY|zq='~aɏ[޲1傦 Ǐ߳>>H9\Eaw0觫TWsLUW0ߟ 7Np?/t>(#ӷGo9dguY`^{`c!Fsy6] *Mk))@ >nybs&k+s\sOr4pQR\~8 A$!X*X~B{Wƭ KOIUhc_\s[y9q|5HS"QuIl6@0\Ht'lw骆ihtSy lWc vGXR"n:~/W ?W;g|K*n7"+<5Bp k9\d6sM:'qlHоy<6v˜KN 6 ,e^+„#2B!8 ,ϋ 6i^{p.KZbk݊!dXgߌuX`; \8s3KP&54\Xf`P(j b`273x w8oB۪}vD rF9Yn k$LN芔n+(hr#5/Wn 58&ۑ|qUp#1*3$sN4Ccn2eM1M6 (+Ji +iNpߍ]'?Ucs# r)3vK~H# +eʍ4_V!1A18gr bkq!V~K@i"8N1C_"ڲ2+8ʖ*K`Td3 δ2`fJAWA bYӺYVRcwb KPrF`HEXX)Ë2/8(TG V!)OnoW˴e '>^gb%R [-?ͳʉe?ҝxV0HG~[3;.щ?^Ӷ0utA8ʚܣQ|Hvk=f'+E3*ٙzO*k\J>lʺ㟀YY0 c$~d }9@ ߅mAWi)ީ96 Ani$n H,)6$jR>۞ۡ 9~R~Z%- ?ښ;y4aw i;|R&iPUAI0' ϻ8I{;]]2/K[-7W˭_77Yi׋ 8wΘ)P$0nKV`)T5tJ^ !0RdsmνUOrش{kŹN65N]6ySP3Exșw+KJݿ Mal1+$8wLz7gvcVbŊ8_Xg+E/lb^|VeNbMŚȳ5egvVbj<Ԯ_[͑k/_:_XD`d2[(Fz Kam7 ]Ĝf>Sav~1E_ sK/s)\bl>-򭚲ˮ`0~ s0zEvXwWv a8ެ _;yxo~QVU -M` mѱ['I,v##+ncbW3!hM,|:;۶Eժ v'5G{|Wk௫+Mw4JVB+XA[AVu̓iq8r;Okx Z#yF"-ݾ#\zWbPnh_>VlxvP{%+ew+,rgo>7HslV!x kuibu(T|h t=W397ap)JhW1ڧf; ,Ko@lIq0i64*:fqfrws@[G!X$a]JAZ44x9,b{qmwbg ;˼.߉Ib0meWaOG =CͿl6?s:9x? ;_[kLy_R߿2n툑vL׿΋FgR]/aIE!2ߏj`Q\R"3{91|#: uL݆=q:nG}n)*ZSQ2FA7Y4jݷvPvk}qμ[!yNjA^?8Y=Cf]-^v|`@Z{g;R>폼Zzk u2*?2>| |$mGa2@;iͰ|2=>V 8"]@qjw/{>8¿1tul#X{Vuՙ͗[}Bbc[wW}ygYmgJe;Nj;Zo {i}ǧM1p I\ Ԓ{G1Nss0^Lzs}?&`3ɋ\ieB'VT~QSn?j>j5>S~!Q omsD_9}J. ;O]aOUbXdڒ+B1c:+#pbb0q:[6Xq1"ps]bE̋=T['&?:Eˮ[gdCW}!0OG^O5^Islz=#'reZGN8n?>fS}˵UP#'ޟEɞӽPuTC#;kѤR@R^D\VQdˆ%/GP%W8_y>r"h>^F_g o;x1[Z$2FtogŧǑ}*~4wZ}$Ћ%\&sod<Љ(w/£Z_ 9s )qu3e^m.vrW$KrysVI30V`% nʫ=Ңu Šquk-I&bj<Ѻ-'%( l354˅2%W(5)&ĺqD'[ D'[snK 9:-ɔwģu㶭ƝB1#:}cګH@5f݂'Z׺Ő3uݖuw#2 ĠQ[{U !nC'B'Zw !g.!2|-&ŴhB1#:}cZ zR[ֵn1$E4HTn[7e[(uD'c[sT^ںOu!!g.!2|-(]n>n-V)Ƭ[DZr""SWk߶nZuw ĠQ[{FnE3wU-.XP NhX_Nys-xu[ 9s )I:Wac!Lp}8O@Ɲ?׾=jpz8U?5jecjXelp V:oSDp/&Lzz1 =lEi^ĆQ+.=+Q<΂ gA"heu,%Gù@TMw֨ic>3Ų9f&4rC9&(c~!<#hB/9Cy1G.Yr\7xh1EY`c!<䘣j@9fd19昚 1"1K.crQ5A޿3`c PҿJ 9!U6s̚":䘇sLMД嘵x19樚O'cu!<#jǽ1crC9&(AzcƘc Ի3 9渚׻3&;kCr SVj1 VoL 9s1cXw„O[ 9s1c AJṙrC9&0zcTJ9䘇sDMPJza9l`*2,/fzjl>[|9Wߪ)>^0|Y_~F՛լZo[[aVc[H:fLgceakI񳖿w-`O˻|/]wuuM]~tVWr1VOc}I^|7 ob>1 5`(q>Vⲱ2<5kY*Wpe$d s6\m(2-5m hǭHS۽&@tjx 'ƇXMv[{ThBSw`$kxť`5v2ΌU5W_:|okzLׄx'QtiPUPp{.>܎%`a9o.ͤq]J >k$hi ]=]#* PZ5 ^ŵ]X5*pӃ= Sq=W39& zJI@.^"FuV]^CT=QMGوf %x9,bLG*ޫ~qnH,2ыҮ/3"!Ba$2Tl2ѫҞjl39rx^>%vnj?\v3d$h!c;ѩ<*?;?RQ0de_I1,0$(9?M|:C\LFU6-n&rf]_l4fS>~Ys zuV31)?{Wȑ ︨ xЍi6<[(&><>EJ*Rd3ɪж$_DU_οz;^.gԦ9Y(srG6P^h- cqA3rR2r-RY$PX2Ynl(0O@h~ߋnY}{ۨயː^~c^+:=/{)eC9Q0,uF 5'<g:EAţ`L@1ׂJ"a8P/YobSބ7LZ"SkNSкf8 ?WZae,Th"dbkũ%ڋ>\GpB;N_#'?9|QhNߜpIo6pv$tv{ *|V\x4,YٕF$zLr3鍧$(崤q A D+դpj`VΥ=]NB9#.^*xG$_k2. ƌC*H{ (pV1KBޑV2}5P`x, `>8M *H(9 +qVu.\`>juD5!/@ 1$KD B;)<aXIYQ% a%l!4M]* J3K-ɰ1 {/#Z{hk4u!/`#[bҨڃyp<"1&dHT&(PewZ2ЋЅ3$K8ԡi" cvA3f _Ӆx1P*CBN*\&Cf"haMKyB !cB/@ 85덎{ 8:J-Y:\(B`@6< h c-) gdtl7Zyv)Xlvl0xvR3/H8xl }lLC3#e0wʜ*F)-2Y~oICQJs&K`Kaz99sIrezKMzrvBE57lO?>lw7 _Wwf6MK~g7|>78g+~dA1'Of.fy1bX00$5GsyC_DK?ˤ1ēٶ{d;Fycg q}`ovƒZͩP!bOAHH^"00sQTC VTel E)tf~JWgͮ.'d//'[/'/|HBkeS3H _PocL,Zcgik<x%WҢIӿI塸/3S9<"XbO@LIhθx}ץn(X+ot!(f# T @ڃ@?"2!) $ ڇ>TЧс,)TYɼͮ: +ɰ[j({ 1v) <:gq_Wj5Aߒ(ւpɥXڏ:[VQ6ڸL&ʉ*^Mŝ §ן.\ ٚ{$*s?{'SOD`Jj7UKRo|8VizWK̿rz-2p*$DE3Y,j e&g%0Ocivv}o04A;>`x]$>WezWYOM|\N L۱gjS㢮#%2 ~lML`GGIj@"G'!Me8!N"YGaA`>DUn sl+ `Jٖc[B|z3t=%lKl$3PbF6JE5b֘h+-)##{؀_ O#l1_6YDDzU5U.xu;BuԤn'`+`҆R.l{B G|]2Kc^Noq>=7J%q5zcH煔G@ui P[aH:бu:N?3zY1C|N0({fpmF.f7%Zk( U(K \gB' 店|[PcB$:յ$G{G;?D?d7e03}7%2Rh2'jL{jYYs+i"sXRYg ] +q/ g%Y3ƒÃRf }%H+:>'/0c䯃j| 'f1% 6ķy5#9UAF0TUBMR+Z!P,(( 㙅><ӡH%Dׁ#NE6Tp:8LWYW5Lfn$̍k`̭tw5s{wHI|M!KB(p> F=)Ҕ`)TnT0BѤbL[jpK(4/iqyL>\Rm tt"Wl<-*IOZĦp~"#҃ݒ b<*x[ qU6RWnA8o]e2U m4!mK:G$޷C|׶GW?v0{MfJN.+ Ũ-4C|2zEW5˘-+n&#kgURL4?U-Ñb .LTIXwCJE9dyxͭ%~G+_[xa1ђ!v6xG +TJXŔhH\ZM޸#5}Fd-%mQCyk r%E %Oo4Aɿ]` Qr Zb˸gCkWsR !_g}G6̊YRwx0Bϋ>,lH#ǖ6Lu -է˞#I٥J`-0BJNyUxoM971݅yȺ@7 #Q q1K6:hΩs[XgWLųDdK[ wbOt1}?jwfd76<;1G|L~o7\(Luul9id׆g|6YKf yάt qLCϽ^"4zӚ2۾+v &y*k- --魥ND3E.jhJ?[x'kgv1࢏L)%ζzupL]m=wR]޴YVI&vѻS'J_fxmfY$$hA*ZӖΥiB =,Hhyڱ_mJ*l?F\Xy sm]TŦ neq۶H_#ɏ־]ٳ>g g^yiۧuwlxT~,QGd|z3"Z=9ag |8s1Se(o^VpZ(o Z#YqR q*1 X&LiZmiGj J8MgC0aw"0l}?o}}P{_গ3lq0"hƵV *Q{9k XzWZ ^ΧqyW_y߫;ѫ^~31`o`^(g/r"T&t9]ړ,i:ܹO9kVCW:;MpS$;hH!.??0˜Abh|0V37 c!`6͈.];t>jfQg35} -.ۋǿ ?Mcdo RAj2H B!,^@@IC<Y &PX% =ioIvfT݇Hlg1\YU%){%VIGbջqnA+)s aP1ʁ Z1*&1rw /|OWwȉ@  \>Bag5 褄~? `iYk՝քc8SseΦw:6\0FP<_N6!*M⚖6p+@Lm}SIS{x=¾27<]ɼ`-ZƝ $]|i~Lei(gX}جŌc^5fܵ J|YF#xz:6цqzv4 }iBB9gBj)QVCQmVyU"ex"έ\S֓/=@sa d/xILA|^rDUR ,x'aRa gd sW{~+1/YEMem2DhCa#)&z 9U-T@ڪ@rĉD7Wpvr&GdgAAE$KrtK*)nTݶjsŸ_JRn (Vۅ*E)ռSCΌ ޹ c*B ޫ*qD5PBobʀL $MhxmK2"$XX76^3Ku r'PE\`੔fAD.IrG3"vDn㣪5.خcB(t |2ܽA1DI5jf(h}M x(:FE!p X߼yB5?PPosf0#<(OBN1Hq7!Hq7 E='*׹XqI :`PyD XZKA - `9%S_8@ dQH^=>cW1g*7OGP28eS/dP%*ys+ aUa2zs%vTc&1i+Rw yJ> |/ WbL#y`TbJr Lb^fLQ87zIpa˭1`Ѧ QiD6#,AHF"!ׂm|VGapR؆1@ID(zs9BG%Kn[EhJG T|SU#8LO`[a֭_w/ 5kd;ѽE7LL>g3;@"`凇Vي`ɺx}S;?zߌpc{WlJ_{jWNo8_lW|ݥls>NEV!;BB 'WK5Oyf|#`%ۡWU~ïb*ߤ0)iv k ֙QM^F/7:'x=ׅ w^ ьi$Yf$zL>&axW %ܿTsL!Δ& ͹(tX ,G3 ūbVZ{L8ZQϬ;F$%'t.>X/m%KB@o'&O?zY6:Q+aMՕ۫Hidp@C;3#0HA:P9(97~!0)S#79p$+7Broba]-QR+77\A܌_sq7+ٛT5_pbSBImH7w +5& _~r`}f;`Bks/Q_)$gPa4?}tB pIwjR'RXkbKz ~~9¹o 5FIxeBlm_":m"^,$ x JeVK^AXOXw*ޠG[ wBU҇|TJ%v*.<.0=>4yB:kۮH鱨%M"'`Tג'_pK AyȋJʅF^BZaXDV`]NWQ+ƜT,/4 PX7$Z ?0-&_07n{¾-$JYVRQh$Q*E\`l ;9%1%5bS8_wny'˛5A~ kSOyl%DPJri,| #PB-/XuMg0sZp*R bl.ÝB"ca-/ޯ5#)k}>Sݛ,>TjB}Jo fbobH+47SWGwym?)EOzC=Cbqo\oz2a- MX;Gds:gH5b| mwQ.m l/ӝ!9a`xVղ25 澞r몵$-\mEZ봨SzxW+P{BFMcp'ytmp *rITL+\9ZVhvSs;-7TJhFY6[>hF[7`iAfU129>mvIdQS46YԔI8܇U6 t^k*\t]pRop8b{[Eti9UJYxx9QK$Zccy: A7;:jBSt̑R'dt$)RLFʑAt5{m6QMg )#,+ fw;:M列콲ܶPր>y-v4&!/8 4C7k,u_6AzPEu5󢰚QBaaE;=}.oӄ"D:Ƒ24>wdAJD`^B*SOJSfVW I 4mqފ2(+HLIث Ψj]gY VO\n,d/os "h.R$ϕqiO8"B(*xDWq .C`0ķ\*|mR̭ ؠR18eZ!7T!ʏ>TFl@$׵$C2">xiv͍aNf %s0[i*8e\g &(y̺f nЯTT@^7"p \H0 5T8/P蓔#- 41BK*8vkC}6\T ZJ0^[Csd-68օGtPiIqqlt(_z,9q mSC49CI;9> wV3ֈLŨw/F?m4h $뉫N`UO܀87 j'N8W,jXR9 ǎ ONLh!T1:Y1]h^zu;劚|q;74ӭ Փ<]o8_l)5L]$/=9;qJl@^{x',KHZ+IU19# ǩ 6BF).]8M7!?㰨zp,Bیʻ$SfekOf'mկ^lW߻wU'8gf6t2?: \`G.k/A>$|ok"JpUR^ r4֙sUF`˷ydބJeꨄj.t4uG ؏iFXՏfg粂l/F 0}7y?̯-1{\Wa >+6j 5xͿL+Pqët"G4 Ϥ",HTlΥ1Bs$^2JQ ? :E 0kW-ڛ0{l?̧&.X?%u+{C8J׹ o71֚K:+y >q!2Dz_e/?œsC9'ew,.J{wQCNcJ2lX~rb Jח FL,0x Kmm[@ km9,B !j͒5][Sb$ƓX^kw;FyVW?>rWY8Vg\`h N"i0TZ5WsDzea-؎& ''=zk#![Oɑ__#wD'íj7J(FP'Em !8} *B[upf)yL4G~be$ֿƣ$-VuLO.L>,t Nue;s2-4L5dKXIJ)X"LJO߉O>%EgFYF*’zfpǛ%Mjս@}xڤ. k#jFuc鯓QMuFB EGl_g kyjcKvRo {()/gB-P]ƩfmHZ4hY4HXWvsKJ #r/Ϲ|4W V2~S–dvӰ@2.Oc6GJ1\cTЪӇ.B$)S/G׳HE37&NE|4{$U|Ֆ1gz_ge^y8 '. ^ce|Q,y2 }-nnudy"_UŪ:-S,:oc.%AX06Q4E"U& H(\kZ4r4=j%02S|3JIջ@QuE<ٔ˨RB,w>i*w`HB&nAmuQU95J6D[kvU5grDNI!׷D-& LRwIú;& B[)UNg !hJl) L,, L3z[9N1`X5ň=A0L0Á&h3yaiLL㤫XtCKrtshs @i`H׼L.ʋ$ib, u@F2Ж($(0L)O^2N=`121΄#%7)W6r餛*w Rk)Wm"i\P܂`:մ@=)HmZ@ pZL3*4Ќۻf8)l]cM4.7Tz*5+_mc?Ίfcšn ,([]?4[w"W;?6\jb ^* ؞J6 yZcluyf"6t5\ /sqGr[2;Z1_ς5םy360s{qrzm1c k7ق'Ukٟu~lJ7fFL$(;0fO]6Q& wZ!՛uV\L-u:I3ÙS!ޚik6BBpm%Sʼz`]iPAŠ贝v;dbbڭ}?ڭ ELI-5H#8) Y"Xt&nO nqMƙU.=`ѧEf#lEyϭ['e쥒т`ID2T*bV@{۔]ܮlCfY_mҩgc5%8&sgd> K}%jTld6jh<.G2gӼ֏hL a.=VZ8%'E4"3yD n{-EC>^;FiأAźIǠ?z4=&ʃrZRX1D;VSԙev"sٿ5dO,>8\j98q4p_>G-eθٳ-;t҃98!&^~JƎ`(F4T1G1/(}»ƘK{e[ `X sn Jj]SR}!k֊XV\mgIvNhF]X9/he: I#${ eoCN!Ć)8h_R| &~Z" 9݆~ .[SLp] vS /: ZȊ<;ʌF|O A-OO͇y!jF) xF?=Gy:.>*`ˌ*XgAYAm4 o~=X) 7C ʘ2n"k(׈ uy80|ަk~zqƮ|]\lW댏CSTX}U2/Em:"z4OvfǿuZk#5puG͎Ʃ3ƍZ&z ݴ<[R&;ON֨oXI+ĺ9^@+}{rXK[ZiQRBzf)ȹ2ris!i봑ΰBh f%KA댹|OG"4:\@9>E_POZheD!s[veV 37L۹KM;`40ϫZfGw^UJbAccucCR뼽SGqm9Ya}$iR˙+G H[ɎaؾSNFS3pcN:DГ9tה%ڛ ".h݇56$tukadd+$iL:?"# a$[rL.2I%{IRgLDV&61Wfp#[+:V3#A[u1$M VN^F%OvR}},hT^aЮ6Jvwia֊u̐gkCYOqeЍ4ru057Cg+7}5dt={&Hknk'0 N6>ŗw5 hdpʂ+1^ADPiEE$9)9f|5"ɇBH9"IO:p*+hެ8dagHqTʫVQQ _=vaG7=ۃ_sAo??r 2(y]P NG>>orq"Pr%9Pr%@|?+d6tXgB"RJ"sGȍM%lt6)*J8WCbKVK]m'0[[Ƞ?ei^~zW[BwdD<ҡ{5 X2"T%2;ՖFaѴk}LqrUJi4b԰^?4dͿۙL0@zMk=E\QeK.!kۿzo^f2?Uz$. Q V8tuJ,cp$X(s)zD^ɗ僓zR)hnjɗZ~͇|1^i1fC.,W {;M[(h漾fKI/KDts6ZobRx*A Z!xǘ=gA X}UuBR>LB[^!zOb2V8,4j|`Ej·CPU*N-).Ѧ`Y2\NQR˭E9tG `)TK+5+s;3||M薜V 㼗K9iv"\cWI;Csѹ}p}л3B.FUndpjFQ,o돹qØx1f`wD^ w)צ[\9\!8to*,Gc&T\)CVB&DzzϳAԏO\do>%k0A3 ]*]]շDn]Y\sdml|s)@E n Q&DR0z5&뽀'Nf[~'ȉ9g^PCV2KGlKρgmDLsQ"~ $%#ʥiL"KJȨ"I&R[CM³h]HUЭ5/p=d Jf"gJ.j\Foqb* cs*cL#R&3[yΗmP2â"*e?{HFrPIG^p^FjIyFfA`<6i ݊̈!-ARc,R+s sMN8O1D QiK#D7xtTT'm$1RFJ c;%JۅDEM$NCH:@ƨ vU%>)֐R~5I_izR8:%@Ք:S*JZrW# IY7P*mn\: 7y hg; %f3AJqBA&c2&K 1b* 'gtH"B ux0"~HG`[`4r.RZr&,A+r ɒs>x!g$DUg&L(E;0{97YT]Ur(\iͺ n~X )!B-* 4HXd=Q4FZp%hD(; c*J"!DM^RDQN.h3ЬwS_䳡4"4Y]n2hDj.,ZSvØ($+#[hf^!׶] Ur͗%`ct~rY1 sfD#zh .϶ 7>TguSUHT*f}aǵR ;|cЯUjDx;eii{Qmny/Z~w*Z"ΧOSjzR=n&uCKH_dYIvNٙ|@kmugz[djI[@CMNV#C~(H:hY_e+dVZ>5'3Q=?h7F >\Dh ՟(~ /_V,p4YǏ<:<:<:<:ih;$JT9f"y!iɩJ*ŽɄ$2<9BGJy/k7:|;q,/;0r0Ds Tq-YIv]VO7/;jfF[&9h' #hmfoK>wN'7Sǰ,C~9Vawlϔ^VY.Ej}z0x^I..m9?@}wFُ77vOW_<fVx6_.'dEMnuܶ/G'h.L0$ nBR>Ȭ>Yn'hl|=ՙ|fSw{YnsOLϤ'!8S(rS;Y2 [4ÑrK^XEca7rқ1%<dБssb\y=5dk8#@8tLB!fIN݋rL"łNYvsf 9wȜ1m3ZWa h5r.FSrVZ8GY5N|$qL>xMk !bw!&ͨ`I{kdki>y k6)TV=J.5$wf_gAY|,^w}ѻs/>K"_E+`=ˑ6HFh\70VZm\VO $`jPH;Ƞ057NNDsiӮ6"!4+ #k,=3}|Sdnʫi"'d5l Dgu%Msҹ;J^{x|sMF>￴"$n3Qf@4`[fG$U)tm ߋ:*VZ.qO/]M\`^I/Y@Z/(Mv:K>etԑQ YjE2 쏬*LEO]!A}(eV8D"@ꧻf/|ihj+qXIGґ-r0*sⓓtaNv3J*>pI(zq.\ Nm;־r,VCWn<$$$- w9% O"!Sy J9h0MJU#8tqFFU)?7a2oW7ͽpd=QY> WxiIӧ]ڑn}x>ҡ\a_}(k.Gu_yLꠇ-TV?Hfg[eKF5yDr%vpn AZl{j ;P!O]k[S2';2y1D;&#YNVzZo9/b2&`2DUh8),g#D"P0Ll5wg5ˀoC ^ٗ`@qOԱonAbOʾ-siIy[e} B2ˑWuLkcMk2z'huE?~f_r}낂\VaߖH$2Of3˛i'}_e.| FaEooF͟:qYȂx'/|Io<>eݧ;ٯӻ..HRnjwF[s1/̭⥻:O&Ƨ,!3HSE Sq5Zi3*@ʭ&6Rn'&0>j|sUbr7d|j x q mHɒ ݠ)tI0q{gԢ}ՒoWW9朵ᩡ9XbW3RClv7$ 2ٚu}u6[k6=xrYbL*1B7|HLZ PA& }'F}Exݢyч5=_.OYAo%{'l׾ȃ1d`u/..}n85W7Pb+x{eD''D`.J~+}gt³lvp-"L5;XGUCH(y 4ҕhB ]T\d M#(g}B" 7Amo}뗜|f76Ȣ1}'[fCG_=iZ*?ew$/O/,RRe~r>e)G6 <ʒ@aDL߉]HKAjvN-5?&M^}XXkw1Mr}ҏFaJrV*ykR[;Isf KXe}{w*/ހ;΁ uo5 ßz-;8קti>qvLcWS^d@dLa6wVK&{ȗ}#Sl` q`C40#p .Z;Z[1 s R;? O3\MxxF&%ezYZ+Qt!QI͸S硓tI6^ &;5ݙD`'maϝn)qA@`sL\R4r#ގvqĝ)2$>8ds;.]8|q\k[ Ѝrrrr`611M SF{!$# 7a "KR0tF]8nup)df3G]yF>5QU` *CA b#\lR \_IzM-\"ԽP+P(z>Im!XV ,Z"m` EB$z Q*h SRis&˜>{W.jW0 A{8l̰ yp x+>ĬrpO"o>w5dGz9'Im ɹmуvq|^gO@CyշV3 ͼ^JlI#U#('JCMNL<1Ɍ1Ɍ1Ɍ1i3F;jŃ̑c9u =mBT"Hd!1 Ț&}8jUQUԪvrq;8]b~V:GqĐdgF.6 'FEыft"CLفVzA'%ح슡;q~b t*˥o\ΎπspRc@vv!ԫr$ Sm|FDf[֗3AScdnZR*|g3MB^6P=(^_Pㆴ{|0F1̣ 8p垍ɓj#a ?ۦ@ɼ#[w&yǣg-dS`/-uPǛ_D ZDh;Ϻ8^3Y^?x>7J}=,]iÙ:lo!z!%[b(qv)m0ZPy[9hC(I/Aqc,;^:q^uPV+Y16NAL@8,Zb#ZA5eJArv-GpRA56aOEõH~'LdTԙ <8mADy1$]B'kh,,5VRX;gD4Ik؄xTS֥Uayk|L)oA!x::u sw4Z*Jn$@Ru Q2WKim0LB0w oM(5A55Fk^[HXAKx0YFi kr)W~ fUޟڵ17 *9$DpELCYz|U h{qHK-MY^䠰 )~kmdGŘdPK/ؗN$AmE7.ilCJՖ,l=~83[d+ůe-],Wd(Dn)hˆ]:穻 MJ<ә2. ,{n7 }7ȬOQ^0Tșh|- =3nY_l&Њn ZApuD'?Y5n%9|Ĩi_|)toC5>=_,?J{:5GWԼ:^p6cιoH ` fko2LKwn%GLҮיȏ>jH^pEo7f֔| -|6g c/"KghnK->!=k[^ W>o|&X=O|VyB'͊'ӗ]3jxyM:zxVºۛպgx5^}m?n`}-9Ʀ)o˫zyC9<pu!mڮFy+?^;W@֞^\_}!FV۵njD!;F630m]Un,m U4JuS#Z5Q1X1ŔfNE(稃} iBLSB5lLb ǡ ]\ { Dg r{go"P/K*C%jH>~(aˤB+n=\{)][U@%Č*mIK}y9~80*1ˬ}־2qi5 G KΤX^MW[N璺V --+oΎ;G[|YޚD]S@5^eg_[bጏQ J\^ޚaI8&3uB|ܹh ~2m 9{Iw @Ԭ9:<4]5ԂrNQv//2FhVFmZNfPZVjž̺&;bD\BǎѥNbCN" BC~q)Ӿ'Ɏu3hD!;F6w0XOh݆А_\Ectqn[7C5k:ctn#Fmmnm UtK0Pp>c((|þ@Vgܝ$Rh"{su|:P;hdm;o5F_O#g/ ɛZ#_P:W|~U_I\ի N?Yb8?Z]$.$A c a-~dHZcAR?zJeZ' V댠,0z&" ٌƫ@bv 1Ahm)mzk/76*T6@~oUFK3ٵqk_Ծ.Z]D+\ LFeXOExP3 YR qی2! %=c66ݷq76^oջ^3n QA{mbXLU5lɪEccRa3>xcz1٫@ Tb^q@Kx: O@A ̨;ݔ;H /BjHD^2.jCe 5 ꢰ%XURȔlx]S&)遗Dݖ@D0BF{XCEuxO=L0%/ Lc]֮k (T!Vv]f{! L0\~eT jYƀQ r꺨TeŅMF.)5.P)+Nd68Q|Q'0jQ tom7zVUSw4M*Z[W_mV0)%2wvI 5ԦYFe23_30鬮so ;M`,3;3NQT [QeS:EC;3qJlwfV8s>R+}dŶ;3+9c;t};+ y0C#;lϥ.gݹpwăiim8rc@r[ e:0 U\aP'4㐅uK||KH-E>Hw~DϺCh ܑ_#EWZﲕ>c4Eo$Ck.3xiXsOH'W 75YZy\)}'З?~ mN墊*\uyH],HI/uιOa6![ѤPozt&Py6vx#+ ͱ 2F7NUDa92Z_^ȡj<=/a.HpyMQalLtM]'@ŵ=,2U#Kk"#p.kF=}Miu.@5gWGsDLrd9TȒ쇏 = 8A*lloK8A|h 9GշpO/]S|Wi+: ES("„Zt#jfɞ1^ R:YJ lWyx@n˰U TIOtmWD/3 6= 9dT2e:L5kaB0'n HxcF9`GPyz.ΖpK+Em-uQrΫ!lPm-ɞ1^`0G!fub&]zEEe%M7(<ra/&ƍ>ίyqXKcH")~Ar#>=dMjO% Mi#7. 6%{ (1,E47" Ch1g䠊qA{Ġt11p 8! jHdxq'A6y֤ ĐU.c9'8d#Ggi;˩ȴ]AD"''q#=Pg77=m]/i:"1zbq{yES I &I;`ژą|!"mGT@S6ShvM]\Iyu4jr!c̆uXFxi)WUec+\eYAI{4HNtxڀnqrE2VRƌTIʓs;-5c?SQ|@j1b&>6Cy'ך@xQKpƚq ig+*eeBfpd(#q=?bjhÕķp"PBCJQRi^KSb6]xǠ,;8Ahy{2%C͸_FՠHe(DL"GI)fHQ6 Q 76b6#J<3a aJCqڽSӤ)TCtrc5nBE4Mcͳ@M6ndִVpA/uv6@e3˶땺im9zqvr~z~u[ƿ<١v+Qm3ikC}s-NTk'o W?N7Nvs5\ofI}ޅ=6 ڂn*m6ѿl+y}U/f,(_V1j)HA#I cGZ㓍N(܀44>]q ^ׇ)PipYcȓ8Zם|i9)?rA* J.Gԏ0@ `bDFJX ]0o2]LW牯=zDZl١X`Q0"@-V%@U< }mpἙkߊ"rJ!Bxǻtr<,}ac}%]pT@6ľc 5c>XT(I5= M*aT\!3ƋiSв,!S8bY~2JGX;xtVԊhZBU?< !8r;B3z:}*۶w]І3w:kv181D1ʠfyW-m21GOj R\HaQlT2Zz׽۵,m?qT~Hf?g(ot64, Jqlikd ~lU1# U!y\7%]K:xsuJɲ gO:;k5( &q:$Q-t 4 T%%Kb]pKnE 3 }S8q<><͔GIGq(*rMItIMtx(k%su.S_ /LxNinnc層~] o:?]z;4_WJ|F C3n Sc/ _2,ںmu; FOuCR*ݞJh$E9Δ15#E[ "&N]stG'B2.XٮEH&I%sud Ф.^2XBX xGI`B8 ŀ2)ާ0(&^&?=l烫~vr|In|cgk1gWVolӥ}qJ i ēCḎ'ՙgli&mF-`c0ڃof_',ϜX_'SةΑ :xh㙋YK7k~2h]ꉽDf\/'2z(М^%vPhA8 4bϱ#!qjAi6|\kf=^"8xOGs:UKg' \NMϽ䟣R'e(㜉@1d%5VkUM"#3qc+Ɋ#$@뙉ɔઢ4Dܾ!@:u`{oZQ,z9"]ܺuP5:5%OKS\FpT|Zdm̑ߖVO{ۚBNa[B)ԭ D,ϸh]qM?kY*z yPEERD(om$9)j-w*SRY.3CʨQH" 1?/0Yo< UΈQV96 lYɉ9;!Ī^YkeZɨTzKG-vl}9-ޑ%&OX|k]R9bH=05կ>yȳ$>$pwYݲ^Hm&4 2BVJܟc;ԙ|GO=T˝Vp=XNxf}W~Mqa0ΗI.8nnsA7q8{u8 -)y~t1D;'~Qk ok'5+HKJPNca<+ﴋ)dkTv+.1P pVe;Wʷw>rZbsB|~ʥxwK"Z8s`-=Fe/͐G1GbM^d @VٛS@\qz>Y-҂s}3?~`LR"^+]D?'('\CGnYPeYA"K.(gw# &JRw.BϮ:/+ vd׽z!DJnd]4ax}Gm3Y+噇f*ATAf;&DUyx &7P>,/#tj^'MԊzSηG*i@B_Fq/܅{@m;P!#1M=i!|HRQ?N!8|Q1N8sĠ=U;pRe'j 4Sa"'JmQ;_c%O6kh͙46k0|p\}ƯiQ `8*S8:`:2ꋛA #WR8B҃$< Q)*eRSZ1c<ڌ2 !0#Q!ټ@u$ Y*EGg-TQ`Kx1}f r" R`i\RQ#? ň<C\iB2F:K>]B8G (I5w5*OWaLi-se&bk=Ahdf/LMNN@5w8sC1F8C䁳` +_i#(2k8ŠéR\``dAHiHxNPX,-PN:ˮWגr=%jrekZR+T(YSAIK.L*(Q8fKhtao@wJ rk( %Cc 8DfT/g+6{VQB22t$j?(؆` L +JMKK!`4/,cY Ʋe,.0wE8H4{48 ÌF b j)Mpv[A, `5 qS*BlPtE"Q9 .N3c !SE3LƆKga#2 k{E*"& 5{z˨uˠ [BR+193 CJUZEԆFQ,H(,>&cĆڅ pڅ q.{v yA8fBB.څd]p5zݿ2)xrݾwlmY2J1mI*o@TF] Bh_`= !q4"H Hٸ hQ7ٶfkf :ry1PI%^yA{OSOCkt&jp>Z$XXpvMnBhLDڔaf&8CVh$8xCiuDA@ !.Il<YHn J'y ae ZV+x hb ^ Iӑ " 5-82JVOiiMR;cĀu:!(tA$71V Y;6TJ;6J2 RH[,4Z-9$`6-"iC'<@M,ð0 m U0ߩ!& ٓ 6TJ639t%cŶG, '3`R ߘ-C Jz,Txq%xĚi?fOkN'6uQ(0cL)Es2hb,.st ٽNنJy/Lʫ< T?]bkks4s*Odi6l$1ڸOy'wNyh՟Tl 5ZNbBEc˂a 0ciۨ֘B!~7|9f3Z!ϯeq0qmCfA2}U#-cLy?NW'רuzh%k҉)`vsÝwoŻ F_~uNU/pyͭͭݝ{Cnk|?n7ow>M7hhAw}aڃn~~s6jZFߋ:^? 'Yq%q4&5tuymzgԋCooOj^f^L~t㢙yi4Z̆ &'_NOI h0]u^\_n<5`933<*Q?z9*&p E4\|!]8ن%jT|-I1ypq{=<{_6]j|DT8G> Lxʯ'N^|>|w6 {=N;䦵Z <~!8ppy``rB ;-$0셫S~f' ӏu!1_:ߊAerm7V(' N_:GQjojuG5v|dIAQ4|^;qtWi4ytaXiۿ߾!n~}-Sfwȫ~3bgAV岁3kxgurP#WxqΗdOX9u1sv쁯<~0{ѡp&- #8b Wnqt|7;2'L ϋA'9 d6^|"Ьy BDl}}Jg*xC$, N6j][Z:dQ6_],+`hn~(.UQ |^ʻ]8x}(drBއ1฀Ӕ~ظA뇍{qYQZR9 |>jnA,n8λ8);;_sן]zzzy:@uȸJg/ߊ(2J94ӎSV_k&ޔQ"r׳ct%!TX8kRGl)5 Ѵ0:q=ZxyJsvӽ(Vn_~g*?rv)+Z5Wԏ,.p$xk) RJ+i4F #iNfDXGzHY:ZGzHo=Hosww8_!$DzAe=z77ȼ5Z%<(cLȳ8K*ӏ‘;(2lE 'q><㈛f5ìgCS5ƈqf( xEC6 '#\(A2DEv }U!ȓ?#H$HG(b92l$] K %~I za}Tm'tc{P&*GqqGEL5`N܌hרN#E9YEӤ|٭RD$lRcT1QsHuF3RTg{Tgl$ ֒WaM "42Ti(sjsx$S覙]h>tAZ-҅~b'w([ "ñ$y[i z,xdzBYƄL% 霕B8^I"{)BYJyC@0{b` Ń7_M IkAE`"r2ȁ]T*@ Dǂv9/ʹ1naMx93 RV(}[/cv.JV_}" _ϑ>/??勧-xnFmk8i:|羴q^O_<_ *zh|VarXjt3wpwzKo.YP+wxeŪjo†Gz߉zϿNKrd0]Z 9WK _6hy̯._pSV: =(av/I d^DD8<!gR!g Yluo_v{5EX?pPDsV2TFcx0TYK!ӀN N ʩhz&kj60 sa鞏 j4ҌvY]F opdAiVˀI?&q)@qϬP> (@_+!pzE4%18=}⒑Pۀqi;cXڸlu0Uь NHdZJ 4֌N6qf%bDXG1B#vI(Ec砨 Rg֎0n[!s 8*W <]7-OiM: D3')j@;J7q&t@ʖ9Ec0;p^ O}ILJdR z&9OxBP!}@ P)ԥ2{yIyvٽ൤61 )*P}|sDʘEA)>py-[t.oB .ᐪ IU <`Q^K=}Q5Guޅt_l^[v0T' gE̊ycJX&[#ol,: #u^}avMnbݓZiMeT.Mn޸~t5ոBg+%=PSa.PQeԎ5{TNZ {@BC˪z2BռQۜ΢U-# vl_x+`gJPњmm׃on8m+ p .ߊ,*F--ߎ'7zwNkiag X2dXYpCecƃTl@(R[/%@< qHu5pM dLt1%m 6sk5cNx7r¶aD(j{n՘.D6y|"厭qR Z)w^Gk T vg)͓cdt)ҬMxc-p4ፋ) TKjvgk µLR܎k :HK0г9RQPLn]p&{NZ;wRQOLW_7;ӓê m8+iMrgq«ϵfnpQ8^J|Hh^r{bT xl92 Y i ߍ`C g6v{L?Ǝ`cG#l{VF, Az/,%dLdp8qDz,ukSk }Rbskuo)m:(4C;KwYz!6lW)蹔:F^(}T&D$IJʖsRK||cCHf#}Lz#HmGjEj[G'QeYLlTpĖ[y q b/x8#m^= tBϖg 4싚ntq|Q*3c>g(X];Z>/&W+%dL|&?yOnr>L.ӠC3 f#XKhTKӽQq'Ӝ*7Σqrko>݋,o4ǭ;UVbW>}\T[+0jzhǂ[jp4RzlUZנ C COQOWAd2Zq8md qœAɅ^T+ma*?ldL~1٘cL%Wչl%K1O19NO@d%qmc*;(˲lmwW@Q[;Zé{Iwz;iqsÌ*3OTF,Ce"{Ԍ撰{ v;xF;GEp,w$# :DH G ړWF[.˘ei#'N8s->zw˻P @`Nw N.ױ~ B;KiQ8HjYQ9<g*aAZLJ`s,$}DhYmoCR32 =&&$ϯG&42 LhdBcBke e Jo$BhJDDR'SJ˜2= j%}Dk]i軇ֺmodȝ#ϖlj9au6璗/Q?^\ϋz4)!߶v02 -Q7èL&? &,$ɟGϯޤC:o]0r#9eq|\__ @'(GڹgaI=sAe+H=y?hS!p΢'0£p;{w5T yt,f='WvʡOcw7rdǷh}w=X<Fcb#T]1 _%`᳞!yvO>~z8M1(3Ȕ4q \h; o/;'y Ӥ,''}DЂ 9Z\ֺVDad/7WI$jI?[M C_y2ݲ h*Zd4 c{U̖pY8%Y2(LH3wq\pVu8JZkrD4"ǁa4ҹѠjz3N>*# a5'g 0mڇ` Rh՛፦;JRUoYx՜B!>g|s_m!dCq^2?>_8 2P7O&ꪺwYY͓zZSd;Ŝl >Ҝi#V;o>hhʑuW~9Ç6[!h]\+wx2 ׫4L] Ņ,s5^( "P 4d;ɇ?WI4HD0yc7 c҆iðO(S 8Y%8C8#pAX{Qwgb/H)b}P"%g_`ʡ9ax"LB9X["B f25HF .F.&ٔRA<×"fepc3=L7q·f,Q&44:EA`bq%iB"9.%&#cA !y~ǪlYM~,W=4sh)t諈B$Ɏ3)kF9&r" 2gOzȓIEα8)^!coz*sJ0nVN ٻ8W0tߪJy8b${f,ںij)JrJeٴ.n]Uǫ?x3V (8dK]8;L)%p.ؗЙS+-DBֳ P`ogzM^E_=dty.,h~oHqFэwRx \p̠Wa6=S1} yH4)vɛۡL$UE"L_x2z>Ap*W*e\)pYɯ} V5'MhVqԥ5:?15^JXH"@𜓭G{.*xe>:)ze&蜑<18p]J.HF 4N_ QRH2Fo@c-"?0wdQz6İDPGcq}T^7 K '2iW$5(TÇј$( 7[W]4%eu\?HCpm nDΙmCaV)Z5Ik7eg*9߶TL۵@Xl"*iz* 6w2_2 hWZz罔t1G E~_H\nBK]҅eh24d\k]㋕> \6^"$v^@o@=?k Eoz).:l&uTh)ڔBq]kc'}gLW|ѵ'蘖H_̋P)l҂: ϵ53>-R46]p8rnЙ Y;҇nMdi?q`#d%x wڲwT8"|r81 2M'M}j؄dS-ܻaRv/p#V  쐣u ` z"mXJc[ںc_dlof{]ȹl H/}m# o$ \ge$Htv$ёB+))dJmǀg4vhY$DX6׿6osMlTw$IWo 㜩pWAMȄ,+.)ۡ/C{ ۾7Ho exǭV0e=?$s@Cn*q0nhBiK/GΝGXЊi'8(r}8/Z? Wp1`cklA';IL28 `R'𗍱TY/S\6&DzRqt<}ZkȽO !lԨ"U܁n{鄩nS5,(s&k93KE_b0 %i0.BM`pL~/ *y7oȀ9M e?0"ڦ.6%٠AtgS3U&y]"4ϗwRpXPzVBG6{>4Xm x<> }K1E<Ýl5Szj0~M q8tS0/~[ALsIqdYH1 jL0wtQڝE;B|!=o2ARɲ&Yٌt袧bsL: L ,4yw" :;Gn|ZIPywd^%`t\1oʮj4|pGUc5LAaayEuDw7M%h֔u 3.k@,5J<rG_ᎎҀԡȺ*eIDuVèK!#G掬C:1iu[MDS~ ƃEIN%(u&F|J&:U7@}8@t=c:(`k @QDܱEv ,A7[E<:3 8E 8T0kk 4%4%@PVDIԃdDS&1mx C(,2&m& Ѝ&ScDɏɒMQ=ɐI_T6'X{UC򆻣ɔ|(ЙWb1K'#̂;u>7Aok2鰴Oxw$ӘRc|QX2'DP `)ƎKDaߠMGt&LL${h.:pFUOb wIPs7'rd1L[cAL*YaNjGZW9K (t L9}jm$ݱE),$)"/ ?hwT^Ӝ_h#& 0Gu)"D5KRȫK%6CzD>ԴnA]UC@&Lj`pXD.e=%_׵ ;kKm{ V{1-6G!η٥c\pQ ݲRq GF5nr8\x7j5<1 |di 4 ajGF-c(=մnrGia γģ y壊*(3f+ʖWgdIGݭ ,{j1T ,lA=%ća+D4mK]<4ĶJ|I.IZӭ/ zwj=o7Jz$?b a+&mB8x*l"oE >M$,z}n.cHM߿Vo7ۋ_|m(ؕԼJϲFtQg|cyۋ|}#kG4Js?;9HM^CNl&s;4Cַ|K7Q /{vy_,:(WQҿxݿuV*D_9yD"Fr~ŢY$B'aD$,. uv9e"z >&Xޖ0"BPc]N ߌ,غ[#WU_ϧwmy_`3~L߽}3듷g|cq| dgJ;KuUgoJ{Zo5`w#l3ޞ[v/ +j ?5KۍZNGk>/`^.c:1!s2E{tΧ ԝ/:{Tsq@bQpGe>*?w"_^*ǻޙ,Pش k^>l bEP}k|( Їp 6he.C&LwJ|;%+=buAyH2u$w р]`A6;67b_Ŕ#(&De r^ ?_&埙#~٧Y+$dH%JRy2]>E|'i'C=Y*;Tw<>GWeg'rS'~4'+Kϙ( 1Yl&VQԡ@y"ӷbHtx6M[FZ8;iѽy3=3';E~nY?e?Po =/3B9?{ۈ⭷DMlBWJ˅{ΥOro)pw'?vfs~8{}Rs􃞝N< & & ʴ;w/<|._u?|6ɸpnO޻lNɬCdbZ8 ;9 DdQXo2Qy{(O7VNRONOIݹ8~dZM2j4]j¯/sh:iK{[^Z3#M!6A2t&#[z?+ǫQ҄'y|UE8;E2>T9N/FIݥ<Rylwu`fw_N:)=};I%/+g,div,Ãm@N>%֯R#j {yĖw9{{_u ѯup4 !}^RVkqGŘ; v`=, e =LOlWfu /|ɶdS)RlY]&G2$SҮya53=L%ДFv pukX248Z1Q! 'fZ我(t$ /(+>Ӫ@$菦ns>PHp:Y.=~^ZqBRt05 G6 Y+;k.:C 4as1 /8( 5DwY.eIxEXB39Ќ3e<Th*,ECEs^߉z%XR۞4@{KtY;Q`"&vbذhKTIﮬj:70ef|_.$n{:[?|#~H rc0촾jC,ik}]<>fq@o?.pt`Hb*Z6\\ kbnV'|z6&Mi$+&V-cV2$,'^I + J"E#P؍#%cЛ C_)C<+7F)3HV_Z+=touC_V6+qs7~<2&VC<yO+ZE(m2X BSicqRD !jNg*6F+z@`O2+P?=u#.6܂P^$c$H3|?wt;2Wwy!?p5u=k4lZ䜵l<}K<]+|ltC30]77 /9+|r"|~ }x rGh_Lh#qY/q[}Ct{z#d0R|èKPLg4P,L>5Yj8rVK;*gZ\O6i[] SV;@3۪^ eLA]đsVb\]^8^2c?\E.մ"ZsB1BeET"#eT;/tAOh^خOz]á z}{QJe*Ph!qlѨ r C!KD!(EMˠ*୤TР)&2Kmo4@M5JQkI1(;5?n gB\D0kNzz^*x~J|^5+i 9<&iStV8 |_B׿]t\Z/$yz/V=f{PlH2-pz#5UP!?ʒ$1ڐFcϬ[IjYiPeg3TOexٙ 9?^U=o9Dɐ5`s"b[!5fzƿ%1\ ӊw܌ RfJm&Zyu=%b"G7%CN ILj! ~, @PE+9sߢ :>+9VDO hQx8㛅o80>l)Z~Qd~y=Sȸ qgsN -KUf,eD)2L%EHn126C\W)S [eT:~,'68qŀP5/<+ =FSlz c(0=eEPfgQt9xV;xH֋"Aj"#qx:%N3*I߿1e^ Sww-  _ 1.`AXY>AwÍ ȫB%v1tWK~kX5؞M*.y\zpB<3Fy9g)hREQP &@eI Z) Qrv;nV#pp+᫷OUjUZ@iɬa]v I$2xHWYҽwyk ̌P-Ď! 2_V>r*ڸiW$h즼1=TBպV@~*s~Jdڏ ̪po:AZCX)Ag=%1 C10E V8Pnhf >b ">@!BȺU"Jp lcތ:S @& y]7.8 |lO~Q+Ravg:c+{ ZrvnB1rdW(0 T -.;.Ze))Qd9)/Ɣ*tQ{dR&&E^9Ǭ$̥T&)-XAvR@FJ/uѥGLTY̢fDPK%aA2-$ɈN0"18g¨L|.م:^ޒb3RO_.h[$-8\2ػ)Da T t.exV@O=$ 0Pas݅Hs᝚t/igFkM(9vMk4LE_%HQrGic}>|egio8qGBl$9s1r3.1ˑЂG_\BcᅉK~VȐ&rzk Ha}]J}X6#QڟH:L̳Q {}PO}nZ@zS@"I=XTsx*ƺs2I510034@g8@J[ -_FIH\gh!H?/o-b,gSD B)qT :8h<ޚ|}+kMqj&Lm] %GXq)yzI^U\яz(y]3Wo~>bk,]^J`__= DV:ɤ=$P K 3yYY&-sp552rrYsVoL :=N#HϘ7>+ldwN7;B gqٷ1FׅY_1<^k'F+W3+|u<)]QǖyPJ͢r=YcYmYٮƥ9KHKKZ(r:Bk BV%Q"T0E PMi̸v5"L]JK8ȏ4cm(zܦwYop-1]5{JW2ϗj|&TMZ ~$3˕)-)_/d/ B_Ⱦ/ E%36BƈBf&#jO%)jb HdJV}_׸&|}Y,)wjY߉>&Z}_8sX}\13j=lumj>W׳y{/a(a|{)U+,UE.+-rYiJ\ֵH}զH"2( ѥܞd&X$z@jb*ø$v^}'jN>DdBn%I;ʵ rj&X. X}Y$藱^+c]b/.) E`+Q揋l Q2DR ݿ{Y2LcS @#VpO\䘭Jrw239pH$DBIƄyaP%-szM|>)}>ށ.\߈Y|t/Nwku|iiX _օ}#E)sJ"e2a <{Zps"K))e81_}aJzA8[2paji7Aq8l6߂w5%1xK6Џw-Ë[lv!F{Pc~G Zsw|D>U~wC-{ 4 շzCԖdE9cz[A2I.f\< | +A!A9(TbM;( #|1dQZp\o.@$hP&>:q}\%U}#nVy K!4`ehE s^H ۨ-5\v!iI0`e7%'H=mY }<6FP[M!pB49?KLN ] >KbA%0Zѯci2|&Wn(QaCY|cV=]T8V,YTj^MPù%oqQ!,\=Kڷpߊ++f!&@@33+)J #md[d.HT֮H=8pjhxQ0SkN ǜ#*SI.5hf RjRAQ6>5sesa rG}[<0 =r[w S9Q>h8Qbз[A CIaD-}c+N1Bp-rs7~<2@+!:?<§Ϗjā5DuEY06;(?hL3^Ȃ+Okxk<çgVl6xqωtH,VG"UQM^l]:ubp4J5qfY>x8EX'c\a9N\tY|kwӆC`YQ28OܩtTɶi'U$)'hT&hTӖ:lE(v缢W~XQuJ%@blo(R3O\Ei! =Y :J/w5=(H)KF(.`8CLA!WЩן y ~K<Z9pmeo;r-Z&{U)]3hs/Lfk^?UVZļdf2ګߵ2M/KivUdLH&O %Y"A*7TD(A^jFxjgUugٗrSέ'qw5V-f}'fV/eYJ%pI@vӼ&|G%Js&(zR7k%5,= Ioûo?n_7Dxdeg]8q<•N,Un~1.њwŔ>ZU~:c-h{^zZ'߮XWYbZsWW89XcX$I&XʠDr9J`zbU6&Ɗ|tfZDxNg `|\pyg+Es|tzVi$/4p)j!/ÎW,706mE1z$LG=4F,f+ڍ?6WͧI$ vMaz1kl`{a/K Lisv} 6*Z1 Z_]h@}5kY ^ݒFѵvЈџ;ǟ;`)(>p}է_^_r,ӧ8 q~qy1JvꊠFj\o/ j)LݳO;q6E(cUuX[r۲^ AH\Lo*M1t.px0rQ s6Lzwx>8y,߃].qJNni/-\A1v"{ZBe9I218֘fhEZ`8fuTG4*JP!X%.V(N7W9t:cЫtMë%oLAw.ⲗz Nx6 3fm^!<8OYv }wts&-P-P-P-P;7t @i J@D'8T(DrT3(_DC@਎$d/%ʕOo |8qγ?F}U~o~V_ɎW<[٭ u&TM ] pv똩L%1h)i ? Q0N%'XlM)VU(+Yx6y4Ű佖_$Q0PEQ#A!]՝;,cFX|ƅȇB6ci~4=:m{|~OXŒf xr|yyG:go^9|Nج~yr{gʐ*cӓ$MEj'~ۮym]IJtx?gO~Hx'^v;s[ͬ4O?iXYJkdO{LaKi u imFrZVafu{#U ӍP cQ <0m [+dk͊Msh<GB;{@_#7MI!`ypa`/ZRv!J Zr%Ew~:°]}GB޼9}eK,HgbQq?\ 5-jz}rB,ܹخ<_Lюwހl'8Ɇ]ӆ>֋Wb)%LWυ!5~瓃uP&e]Z&o u uj&aaPD#4r)mj3薀_ʗ6h1Is uxƈ3"? K7hY;êH,ceZrzv<‚VZqIKU\iV8A42CH e5(B FRW%sfUxMlj҆x߈:Pa7?Ik &  p5Ldl2>gkYic@)0 D @kœB a."i! w? "kRaj/XCq.2e$͘Sw q $Մ;xYb2.S?seFd)0LAe$װ5mԴ9qGPM|OCX'CZ0LZL1Ɋ;AۋA|Pn9.߇Q?Mt.\q%Eʢއ&F߈Z*"!$2u C_SUNoxIq}:\cķb-cRoqWW8[pY'הK#;J=0, J,D JK &2KMH]\,L!Sl1ųrn1Sl11T9Yb2mGT% $VAH8pj՗)8QTck`\a&b)6)8[+O" 1g4[1Q:xS=̕!l”=70fD]'߰2EP[HR$_﷐b )b퐂ф*"|)R (THFD9L4BR\Rԑ5ƨ{c(*NH/R[HsS8Mh-{pn2@oXYBIf0><IY0 $rGrP ܫ>4^JBN#D0GÀ2ro:JT Y'$oEP~cQT>LLp|ǔ*:0&l3ܼ!UBpsd2(ZrOejob# 9%oo&+XxTJAU'00&Wl{~< ުA0 嚴Di#E+O#m /+&E6rYҸs}j}"q}߸ĀTL }>lbOZU.jcf~6 g@/ϕٹ7?5_QԎM xb*eW!W& sV^bxW/6h(l<!ܹ 1e%BdLxܣ`sL 8 Ά#/}g7EYpfkxyF ył=4hT,>5w;RHF xw[x*g~~T'O dcz8.X `8 5^, iXN@ vRАg0`=.6p,g&BJAI/+66:X::{e#Ĥ͔"L by. "[i[e &1?jA ;~j `QHt(xKH^K"AXRA'FPU#D)*y#X\"u1eSXڲ9%ژ\JD/A!)R&n0ýieC*+^DVc2FDA4"#&!`h#|$ :IK$ЍbW[%$ c\<ТnLF 1)@ֈӔM~6BGL+5lKFM$!nU_7LLٰ_Bv_OSN%pSdJT˜m-&`w*0q 845 pi| *.CH( ZcPaj!'"5{LDmas86=n@=Z7#At,ht  e-H(i Frc" ,#g!ǹ&ύ"p̽Gਹ5t0`p}F"8 "Q0o6BU2HR#_ҀY^Oj0M6\ f@m tH=UrP 8L! ~lӶJ0%\*jJxTA07r\J5Rm knHVVMHs6]%T\[VI@ ٛnlmKlKꒂXNԕ{9 J\b46TR!GN5]$5Ѻv oq/eI]XB09 a R xg1y0 6XpPCA|fNM#{AX mՁLFd ݿ8$po2bfs3u(iTJxL?B+D=M{gNi6pYbIWBUJPԼwtL{wJXnYK-&e]R[žVRk{:5%k9h\wͣn5b)՚4#W+Xo6#>帗ՌN, ;X5Ѫ7b!+(<#ۢALPz3kQvۓ>UhUEi%M-ujesAq'yx֡%|"|/xp/?s|p?'b 63 )i&F A Wnf^ dZfeo~{3?NtKwN{u=;|\mŠhR?Ҡ=,K,ذkLobjѺ홈S[#;φ-u"SWޖY#Sc㕍|PX|)ξ!n%e#1YfQw/sc;>UoEG~:BXq_[f%cMI<8X!qad￾zLb͖g9㔵@Y믓mbڱ ֝A ]'YKmI yrFG+wn@0{6W'B0X+ϜM+<$%ۼA*j ~ ʫ`DٍcR1&)HI SYzUMfLiqj`Ybĝ`&9 W||29{;41|}_#×fncpf kD"9R(2Ak].y- 9Fqt@G +oE:8!B3 r^^T na&B y+N2ː pA[- TbD|PXj{`[ o3 #ofgkB,WުVV~#$'yMUH87}06#U #ozȋwY^=o ׫4R_W/W. 9+x30~~|)ӳaɟ!ݾĠ?|A_Kbcxc:;}R}_QeWo8_@p}s@&Ԩ"$僊!1ZnR#tXrjv6AeqdU{Y Js5dX$ dѫ̎tRA, 39j>#Jƃ$Kò1 Ȣe-⧽NhP(( D . ېMY>Axi bWDbÒQQx?{Mު֧wp[(1By@_Ceq5}]˿ySepS.x:w%KPA8FlשJ)2I|euz(£8;ڃfZxEDmhZ!ݥ{wyk_ @FFw--塅!smdvnΜuS8/y˦p@g6gIK 㝸ڹu+g ϐ| Q ??/$jlcʬAf?/:7S~8idtx}w7r|,,[l+ivu!]8 J)I]cO/.˻G,J^ݳ|];C;3jZfeMvBƄрYD--n2]TN&Xyr#:NwGUj*kp&;IpT+Y&u)O*kе 'dc(M˚`֝ KŭA w:}Pּzld):jx,k 'gjp.dZ (פ]tׯHS?|wCv-[74DJ`iev^@)Hd1HNE ;M J3ЎSJ =R9 i)ԋvɢ(?.?_xaUq!V!^1%D)Z%"().CtWcJ`@ -2$S),^,KՂf:ԦE{Uq6,u1Q!@`0Rb\x0ڃ5ۿ96h؝HTkvY. ("4"YGorRp #GO/a4 d]5#ԊDQ?*UTF"SByM6SL)VIŕHy\@eO DhNrUy1lPΌFyʣu2*)L*Ab) 'CʜFOaFy8! JNR3-$@*fT>eN+ stdY 8 Fb0pIV_7J͝lXczPЈj)(u6[\(.J(992hv[]PvM27-yt6:si3cظԞ}ۧ%A2>0edz?!]΅jaE&x*ADkj1@0˨ӗYC"I18rZ*X霰J0D]i~l#zM>)ЂG O5>]ƞ6ȃq%N2:@6t2٤,RP>9Q oS&`b,?$KƆ>@_ 7 d2Iȉk ;K/$0K `ۨETS|i: 9Е֘9fXVJhv4*zFىF*~Vc3`ۋaA&-(Ɩ;7Fq2zٜ)&+#ڛ38cggÄE$RVB"Ҭ_7>妖Y$9 Jf$t^?6C`@tӨMZ=z22f$YiAk-5fLqҘMY)2rnږ3C)Ŗ1F`}lVjvgwo[m7 lduST~3@~Lud;=PT B!͖u)'"ZREնuJN׽nah{1}nS(̂eآrJB8ح/O`O]S=ܛ z46.}{w䯳\37W]nwчŏϷz?ӆp,Q_\wf\'a g>|~wq#0F?^r>ӕ "VJ908qՓ s;OI J3Fd]UR{3r1-Kqhc]^BqM}\}Ow#8F4s-`4HcJL݋7\ {/4 8ۤ;\d:5Ӷ t}G ?hkAn8_7Ǽ.~>t#hծ3nKY쥲Suܪ'OY! N{.EXNU|:a];V]|"ԝ 'ߍ˫'_ɘ>qȵ2aKtZ˫'$qcIthZ˫'7ɘ^BsnI#2ÿ48&8o> 읫ФY11 /{FD ;ʙVb `hR}^*Key;q| U˂ZrT S gZ>iUhiɻ%!SPz(;3K-ǂĺW˫e?a$gBcZ`'mZ֢eVl#Z\%Z8TtqM21qˀ+MFhTZDx|s-rȮJ7N78E8ZDٛhJu5AYd(8&u 7i~+#N1vfSLo~P+;i27sĜuFLVL(s}&5^iF]n+NVBA)#k;8]Gw^@(^vW_ićmzE^}_P%+/> Cē&qلr/LktySBK~9rssVgeSD͐"9j<{Dt VEdGRD='ܭT40QPehS\mP\=-.Q\GƐ"VѕEb[@pt0D:2&ͼOa5r8kq^`;P̤zΨA5uWC*3 y é< `Rrs%{AA gT8 'msBqqH׿{ݣ#KyƷ!0XZ׋d"OԌyx<"40v/W*{ANǩ*]DW?.?P?,-UUkBY^4JVi] 4-:!JTUMs0z'ZKHCW$@Va9$i4( :_5$T-Փ/dL/Q=l? fo?Bޟpm҆+s_9<”xcAm ɽjAœ'Kc&z:} a'woc˫'d 0RO׎KUdU9&?қqQU~Ovw)ޓ߯{ierk2C'E{nڅ4\lў$Di^3<9mj)Cp qs뵀,˗!! d2Iȉ/mwM3 J.OGÊvl98Sj9.+y3o6,nVd4Vѥei6Z3utx*zUn.iJG߸Fhؕ[U8o*g*Y>̀][s6+*a2qڙTg./-;\/@J6EQ(^$uL|pkpw,Ǽ6jt@bJsEn8.֚=;ľ.X͟#!ăMR4߻nf(*e'.Bܗ5 ~AEo~wXvT2lLd鸱+"Agci,# | մ!$O2^dBT-lܠ>{l1n)ɪѕv|qo6ܼ_(!gB2EY27sS5afJý7 Ǹxzck1z8Ǹ+2DՀN >Ջ xh؉T<ǃ|1HRVk'#kg)j_ >HRHv>t8oJnx'ŋ'Z69F.#knAUZ Fl_0Jk|~'y|TS6sMf +1 E4\2A ٚB͠+5SB~VO7nR<4s3qJui;CXs˅$MK˄*EV6\3&8kiY"CRKUeRA!s\b(Zp1p&bf["Z)q}Ej@usnېrpyhw᷾\7Gn6=/`ycvJ9}|EZ5?^Tb /ھ-L&4n<әýVv56ϋ@ؙ[<-\ӿWpg"WkОB-JT:Ґ?ҩTWٍВǜZb:-hv;"p[ێFnh\E+:#u܍ y>ωN4z=[#;5ͶwW/zv}?1W_5zjEdŘ 4hDTuh,UxV]*BJU7.<7f~pGyJvDuv*'$S,s[fRP*L'v*D Ak0u1G&W޺Zτ}$x{{N} dz[Hm.q]Ƴ~ETD/)[RPmqBILun5$K#f}f1uMkַ05`lڞic_쨾V/#0N0 q!1<2kdE!=O#ŧcA-y|&&pA­J>Oz}E>X-f:-lL[D/b)?_>yKHP"(gϯ-<mn"~mIԟglX2d;l,G1 %!8\|r^ƅGK"GV!=.'4Mn3Ŵھfr 7l0;W2_~.ւ1+6$PJtLICROšӌ!@dhQ9k?(}?IaFF`a<9*;3q8&*E+2.by_ l*hG,wJ,r)U :)ҐBq] xzp2X#F)P4aL &>C2(. ^Ayv}4 >#R^+TNMt4/}3+ֽkSId'HTxM]]2]P6lm:EX˱=Ր\}4NcLՋw}0ın|{ʏ ./d?)N|qVѠVcˏ Nܲ3Glp"·z75@}g[п+ X =!q΁٬'iB"&r{GƜuWayPNYAI]>ee~rCq[]5VcZRcA0鐉䋊 װurFEip8k_oLSvODjgd_?Sj6&Utq-31}66HvZ3…c4YƤSyyA~;1M$:MQSLv rUXE U*C0sͨdVc[BJLF3ǩ+N$Q^٘ J3;DR`469(@G1aXk(PFI m~Ѷpd c0h,_Ծ^~g&1jlyʟ,?Yd  r߈c&$6^>q`׵~H!Kk?iQ{L+nnuH{vg!-P'.gOF,J xFONf]4` E)y+5Nd CJ!E)SPA;s cL4iU 7:PwQbekLi Gpww~\a3T HÌ0X5\MK˄*EV1W38]Ε-|@Pv:߄3i UW]2ߐQ›>=*{e8(G~X ?oRi4<]: WmF Ӄ{(KϷ |><}szI"KB/gmM}ΙvhQ;TCWXt5L[nvܔ҃z,/2~⍊:j!]ߎ>DLΏ'gpbT DQLIխ.J1-^ԏyjyvǔ=*<(9J_(R%5AK1'=RȨE͖w7Q(v[7Ax,p/;${/{Cy,{u`V"— @p_xܣ_}OB\!*6 `,BR!2 !cD%/] *ѝj%TRz~ R !@(?A#RB3,BaG%fvkoL{]EAi[N!) Xbdfj`l*SMRF )3*+Fa D5zgA{-3ic'2i(p7*!j;"4 -8]9\ xp/O/C ewñ 1:JЄQ4ziB0 HZ_|Uylp*?y0y8G&o~ $ 'Fr=MZ+44yMw,QWA9y;9r;DX|=N`lɋ)'A n'=ihv%;) }iqUg8mG&Լ2H5iI(5e)22 sUf[ /rkv4RN=~й z!n4+.e)P;x9 ںaTRٚ V a:#f'e9PK1* !ͬB +%UZw0IPqMi1TK&*&Vxe)ږEnf*j^9=yo8 hx|{5~t/q4+}{4fVԇӽ/Ѯ?gs7)w7 we7~&:8xIٙE7rea4\g\ӿY ] \zm;p6$!_fɔ'slj7U-$v;5/2[D[2' vhoV[uTZھ i[e.~~.m䈘m]4Ga L*[=a;YPDI,=ܹ齍 /?>b*fnڎlŒ𫛊-KeШoeKz KDd19ć7-A`]3N0~ʊT Ti㝸(II1g>R m֌6͊8w9:A Z9hgIdKAXY!- ˩ܲm8Du@T1?x@j۾EvZ!}re]rN2z,)Jv,!*EYRYC_a.=Wm0V.5-ɎoC'w U3SvqIW;w}xsl vҖM%Ec|Ȕ{ &(*6"|ae(+f!v];={ンęݚ̘B ;92#8ȿXA09+ [ SPd:4!.c* v Tgͫ%`<}tI(cv]a'~!ZP![%zRrhԄK\ >^5ˡh ڤ/yE1m}-u.JHX{R/tM`<8(>hd AAA; . !A3Yl{tm_9 #lwG mi%>P 2-SWT*1Y.Uq/(Rt}מ:\蓉|Q9?<-7YJc/(,0]w"~1^ErL~cBͽI$B[m.vjeMbY kOEAm$h5pYE;X*Z_xьqj3%gǼ&tR;`oD΋zgacl&[)K\G`!2Ւ0vx- Anc҄YFV^39C眷*.SGSSz Όr5ё{ٍw(Z¿)65l,Ƣ0x"SBU-1zw.Z þi1;;"lz\ޤ3Y\ QO@VVj;t19Rwy^MeH% ٖF*²Î 6 H"nRx= 1imKn aʺsz1(-qzd.v 0d o*htEqvD5ff\n&"Z&oe{vG3p{uьof1M 1&½Y}9jg5KS,PWJDWF̂_z߼=D S_?1fZi&6L6I~*׏W\?^W-˨@!jEI0isxX3[*e4 egoПoz_ +n{-Z0YD}EK釰ކa?Xv/uD v.$Qi#\> 7nůd(˙m7<`L? iaH_=VS8FOǧշa {Ӈ &Z|~!vsߛCU_[{<,z2o3@4w>KNdf|*˨(DDE BU6[ o}~:5YۻrFVZL Izg{# ʙ⫵1"3f?B;rH[mδ9BrK,#KqŊ (8"K_Jjт)WW6"/.~E77npFd+^ȍ\S(VƠK{0`5 uZZb5LB@d |~+:Tv^Y!݈SZHOȈ{N{Ǵ~L{Ǵ~5j9 QQP(ZDpj|TZK'1%F g Z;P;$$W?{_x_82Ej+= pqS +#uv8Od( ( |M>Xk$ $ r)eiVb)F)cHCRWCepa $2ӥ' {)sNR$)vf9x?9x?+찂@ $(7bh!8FAa@-TJ742342?Xa5wXʫG)wlRsͳ#i Ks "R,979 bʗqÔă`E\600J<8qs,L4>[r$=NY2z UPF3&Urql%Z0l0ǒ@Rp2s"!BhVl s{$XV1ְDE }`DPF~pd$GP)&m͙M2ҼPNVK);d:9Q+IKCzTC{:6I\$+,yFMA=(&_{jOw0\7ѷ? abq0pvt5/^ яcwH$$M{; * B.9*cl!?PS-Ug["jy'W)'E/nRG\M G9{a_&Kj;kf8 injJC nsli\\..Xc6+͞>5הQ~hf ϴf{-=BQ$s@hZlsa'=ܹBA5sfn')oG\[įd4÷jR/abj(]WRwT O{<r(9ANIxES^:\UwG ݭB Ir&WmwW(Úⳬ?xyzv<{vY"$g@ <&$ h NZ{y98߾]Htef|`88氷 ξCj}tm^[hU9YSΫ M,Wpo^'{,%1 NoRD; /ޑHz|#u&UH/*OMW^{";[ʻۥc)-:̽;伞ϹwIs]éy=s2QueeK;y׸8GY2*Y'—-9Ӗw-Ҽq_ݩ/V /mIta*D!- #ZxF.c%gFY;'0t13׬6o^ИQO{TI;co vY'Gkњ=GKci"}԰v<އ ȗKa73 mT`|?<OS\_7NGw{@s 4PIzR7s|K4H劓%ɥp3}#g+  ~L>jr*Ϲ +>?TR[DI=brAHEt}c5yv`-R5o>ˍ  y0ӁB$ K%5rG.'+6.z)MQk(nNIxwv>Hq# NsmS\hw1>xy XJ| ff7+1NS%Z[mIHck7}Q>4ySk>О+ʖ%m\ԝO{)JMWbӻ{ɶ!B5f{tq z͖2YWkhee~/K PnQL\2Z Ewe9\EYInvg)&{%Qv<5L]ss`epb@V3^:Ǿ"o4e^ ]Zsͧjr_g ;/!ZjdC\ YL*\^_1T ubu[=oB>;\"¥@Zߠ1u]1 g#Fӛ.>BX)w_hv %5JT*?%WP(Vz8/?`LҼ;]n%L%#m=V,2X;dA2CHT*(O 9jBjƒVݍ{Kgŋ5:Hjt-7H0K&o$  qp6Dd{GJ>h!PDƚ>v"q6HaE1/)YP6z5O@$7m[=܅qj&.n$o_"Bg@~cHuHI_"snމlZaY_"X 6PYC<(*8n]耿H(cˀ * A0 GaE/~(rƆ1KmRFqFG󚷓8 ֎#/+tko[1`#cpVdWYqv7ֵ4]^B̟0E9˭͗F}FLKqT&[2$aER r F'2 Z'{)fF\&+R)'_Hɀ4XЇʹ0 ;dQSHh4&H3E"vqb&% yRwpvav7JDZ@  LDL( Ӄ$cL=[l"|qW} s`U|1m=(YASЧlW|)Q$MVdn0kMG2p7i &73V¿|eWe]6g@-e1>fGʟxFrTRnڟ-gOĊ1Ԗ hIjoܟ.$U]wDsʷv}D*;Aq%nM8eDbSDŽ 8!}AWa[ -,8ڑr$%VgΥ>+lH)ʶ+4J]m^r`xɁ񒃼l\f(VຐSfL\TH", TT"U}̓1{D47ehŋ`1e@1:O :ݦKUW0&OuB`[R!) $F!I3cR18`L 9AŞo[`+-(o~A:FUs^TL(3ɣ8!,#pƣp1&E$82m?:''WlgSlq8H(!8kDhL@L? SFyH8j[#\׷ ,ذ ń9xJT&q)%v1fI(+lt6W6kb""I&9V4(ܥ۶=nY:JH&117 lvO`HBP(M#ei ubLy h< D(*ġ n\)0\$<> y:SduVEf',JE 㲈V}S W[z㋧/x3M?utg̢ U1W{^p6Jd/1.%p:?v{ %W!xXk$*J1fYҋE9Ƙe%&ZjEñ襌Vt>Sc*djD*q_j%c;29}2Kh~ؐ7wih aM7;q MD/2sͯ.^}Fѫү̗[yUVkvO.B"ɥԶqXVZVrlqfl[M7@'. Xp)D=iO&#)'<ݞZD&[dNs9Z.^ 7EXft%}˴qB8 "K8N 6ū|]q: &_Hy3,%ƙF #8Th#ҋ2E&XX-> *˫>\^ lJz:=졺=:k1liq\%|Xz9\_28yzlmfZK+ Ϭ " 1X~_ҕׂ<8v hsB&έLm)?oS녛S檊O7U˧g`15UW(]{ndvxnu> MoJ$bƭ;R| L{Zan BرU[ёjM0H3gLcDP7).؀Wpj mA CdNn[5j01%SI'eVk0_)ev:Ӹ=9;a,7]b&*UVQ6Fl  MA[7%DsSvZ20a2}?AL٪gZZS0NkqV?۳0٤EqnMZMЃܸڮdѠJ-{,ɻ>E- !9 /:s^+ggWoo $ ]ʽ|WV}*d]UCBZ1ʼ B[u-2@(IG%HˉaOK7̝GC8#J咚VҦOjAD T| Ƭ|0-(bcQ $ux&񠆌*cqw)g-htB?,hzDzBV?J^S[C:E0m5YuAI2w w"ŕfLzr1Bk %^ )Ig%#Cjb |Rm28<`k/\t4~ŦCF)lx* Iߤ{I^ǥ n׼}N68nkdh[mdw/СN6v%2['vڳUM+X{e' fn, >aO#F9b4O#&ּiO}'IٲɉЂۣךbd:2^8yr;%Qgg%)a*kL?h:OB3 .bw&rFB3 qH\Hfۖ7޶ֆJqOhC+%O.fFw ⼺?w&<so9&}Ntuam"D& {8OE7*β?;sIGQԷ}>,(syaOq|,-R%(_ fI@ a\C#PXB_K@pZ8, v0rp(.bh~3?w5{yERz*~E'?o^ع&04B0 FLIo6q_Fh.q(Sc.|#\?)Z~_on޼|u,yE;hͽap>cgW?}o^|kW^כzs f^?ظIחdqe?q"&yxj]}g _{22*WksZ3 {?|y܍{$;ۻrr7 Gc.vie/>A/M8G3vQ2,7y7-|"TaY+7OTa_eȿ6(I>x%Y7z|cJ[ +ɿtHFù|˵5*jv,Pn% 0pb2HO•yi&)R˧9e(&/cu)KΉ˾5p}{c0}ӏP ,j&|d>^L<`Duth~?@TW'Uz!z߬\z?O`!ܫ_ae8|=vJ/T[}i4| Sy\݅\9`04郾Gq.(LAO]w))J;rd/Q\Y57ئf35MjE] }^&)>5{B78%y"t8y2nZ" m! B)U$Њ¢tTtӃ ]2dI$U:Godmg͙wycIĦ?]ȍE 7e-jME!p.W.BBx | '/C͟:s(H+9W$o9MèA,/ӇUyRoxǓbI%$h=>),a@ i,y!~G_J~J'ʹ6'J.%/4mUR:\y6hXy;r^Y{CIJpĔZ?f/>ˆ'<&FG1FB𡻶ˊDjNMZ,~7_K8A? ;p8ZO|ddb,4;| uABo][$$4BGWriE`AJ,PQ R"b2,0П(q߇w~qz^ݱaSE>M: 1LOݲ?&;;t5~Magi}J2PP֖uZ>':LɯT*7 5Zr";%G)쇩+uZgYf ,W?yo]nk&#Ÿ< #b<* E[£TX@ BB(46481EHjMAKL:sr|@@j.uL'h|%oY)v-50I6J]M@XE~%f 0JG EXM9Z#B1BN0"D">@qM i% QH)pP 9- u2.X&A:8mn?^a1, O 4 '!#ZahU$QP:= l,iܖl]|#,XS`@V# D2B5P1MXBgu$58f?=B%$ =Ih ̄'H-  0 9YY Ә 0j2V`h4R΅͡5a=g4)XkPX`4!tm1єZ@hBKKdBD>JBO"P`Ő VFZR8h!J2/[`~uf ?&Yt'f 5Lyy=\$A A^tW(,a ]MX W! ܕ"q䬂"U; u Z!~΄rgv`1(҄œLfQjCT,}W.n.i:P]mW Z xҐkGC5s!OQŊ0v,Kvgsgr3\e]31r.QOwSX Ou06iGvOQfڑB^ri|& `J"e)mx|e 2Zծ<PQ,MAW5RͰM Tw&ooP5ۗɃȌǫ KQ8\hh`F\Cح Dc|e޷I&Xhv%[8ߕk 1:_Ŭ~woҝ,(p\[zv1̧#;/KL&D!rh\GPXRGLB0nxPCNeElf$N>HK)cXWzX L Ł3AQ-kGvi(mnǦI9K84>R}pFխ3y/ЮYVk #NB@Uz Y*C܁Mގ{A%1p֠8i&,?-ӄjkD:-mG#t;ǤZi݊-~M_+J6=5܆?;HV4 *Dh Vw[cBr*വ.[j*Zn8lݝjz֍S!-$=ʬf R!剣lLQ˳&wٮ@/(n$]jrߛIZ1혷UlQV*561ƤňGkp3o7JRʗ<֮ոʗx/ n@He0O*' oLb+R = 8soG؇'{;vD!" 1|^dž_tj^kp<f*BLq^ٶŬWLcg^̾)Y*}qeHR)sDp!A$FR{ËD?L,/'UԒ(ٛF<" :q` օ-}YHwlY|ۂÑNdG*WqC;fKHӧ/kp >m8_8U~VޥW9%T!.xWjƍҖ(\=,$y~JIbdRqzl|oz)u8xZB%Hv."H]K>p!ҠhSvAyeH(_m@k:y mW\5qFuɺ\RG-:V#n"QxP{cl{*lIJhrԭjGhrMV1!ݫnߵT4޷ms-VQ^ܼjxvkP`B+;';3=3m`J!ov%okbJ*k1B$ 07+ zh"!Ec?ւ;)AycNH\X XXnrŰ[ocL ŁJ4 ~k0'~sנ@׃{i6`*>A>͙_c?AAǾq|]dDTXR "^uF(uźNb] &#&s90lɍ&(x"F!nV*\wZ*\wUX JTa%Pux6l6lb>[Q6b&*HͲ`*U"Υ%XRgssw;UTK~+My!)`2v>;~~]D+rq\`fqBW81RS#yJCw;pCs38$r0e5$TJ2dȂLX. JRcv/wnmȋ2>GB}g*X;j .<*_"Qa8Cp;nkPr\_LK|bow5ų(.o"09LlqcJ <.j쓵G\AkA0IDBN?PζpJO; o:#p^XMa_μZKhXk_.-_QV[x{J2RqkɆn=*Ic]`us/ qdQ;_>X浴2Lb)tJHEv)Ep.e7ZX0h:9b]itIl=YG.$:]7̗DL}H4H"gW2Ej$1\Qǜ3H0eezȉبGi'."[X nT댋ćF]+~vNHZS@H[.%D':oP:!t:ٖl@^_w/II+ߟ Cr9Bӓ8#. JSd96 %tJ^-eѓ6 ;xdV2< ?en˖8(*ewnURs OO \"ۜ䊾~{pwW(bvV 8ZoqM+$e Cc^RNO `q.,m,H1aX/@nie8* H-v_o!QM5ٮ5TEWkl~G+0/А %cVQی bp |HGJhUXibZ»õ kM/M2r*Twa>ۇQ|0*}(nIahn)6jr8/U*0򠸳*1>$o6pa:|; ~\(4/Q<Ax-` rkWO7Wo۱}<3oRy  dahABN2 Ab!rM3}Y6n%A}N\A87͂jD1np#,\INDZD{0W0!Z޹­2:i&@F"LaNZm!"z P0K3#,P0Ŵ(tuWziۚ$緸ݱa V(uyEv8` P< o_&T-` TRÇBbo nlZH0,Z_/HB9#n?l!b)@r%obBBn"QsPEh4k ƕ\t)Nxf˄j'}I=+pe5΁Gv ր8BLg(Q;;VHL`ke:P]Iw+5>=>>yVS&c=(:Ǯs=c+y^r<=??nADsw L(c8xzT6X}+gT ž"Ei\`!D/cj DoWpT׶J|+4f5[iY{6-??:]ˁOƢP<-./㣳<)H"X[Y%Na ٻ6$W,vfΌ,niF6%Q )w{7HJYEC,VE}Gfm|ϑEgV`L6SiUz#nF/FPʜƵ4&Εp>:%2"ѠUH$R\GZi`7²c>)L5mZ4 CnMNo2qM Nabm{A (Fzrz}:"ORB2+ UI G41_ұ-XF)dJ05L ^$[\V%rFh!G;Dᬼg[yd^,'9c>qB+eB$[ԩ &^%^t<KcD1huMɊ?Bv2_<1C9fQ 5ґ)j8zHh ƴF=!kOW< s&R;"dOD()g4+2X]ki +сV @j ` ^##N[A'X֊(&Nl`RR: 1qZ 0E$|&U]- 569KLHqiVxt*c3h+x!%;?T6Y\4.8yE@78|ur-NՂe~ImCqEġgk{."梆ꀿp{GQN4hyi[w[4BEӢ.V,ЫR}bJPou 2^MwU:C685Qᅨ+huvvUxeA&ɓu^s'52f&bJN]4H%Wa1}BD.m>8ɤ/b/q:N_>\W2_QEE_~$W$H ϹIPgwoU9/Yam0D_qيʴ: ԕS+X' /|s0#{qZL(FagMGa}!?^^Ċ}'zWCП BϪ/!]}I8kGat1\_g5p,0"*o)FZ32vDpG{4C:zQ;i7e^ANFF-\myáƾ`'&۰smC罔 ٟ=iHlM;Y]#<uif HB;RI vϻb:jW-ROWȮ".n S!*+_}vW_}XKI$m>óJT[M~mK]\@;-gYEewj-ϛr>pB^UkyvN): LE>VyY)ӲsF+m:g˻om=yR϶9kؤDs*l;2[z8Tگ3R0 Z/0jk)PK ww -.mTJ;1,O.Rfy,F*tXjV󹝻g@T._fxHsņ\Z(j}xkNX"r {1xFB|zQǛQOp_r0R(N.&)8 KT ޗKz;)S lAX_ǠCwA^AKz/;MZ8v P<} )RGltloծ 0f0% -Vc&8 ʈְOui9^ J@堧ērR9%`/qxe=ΗqXuSL8hOC!&{=v\}v0B}~bd"n3wd!YPJgu!*4;6Fhh\+o>`MxGs5a{'?^\r ࢋ ~|so\0!/o ̭:˭+z㣏-a=ܮe罳Ԯ!Ù|*z^^w&9vQi"tPk7m]:"IҌ7Kp]ss=$y{0۷o>kh^BnAh.!! !' c7#5'1둥i|;!4EdސFg?{ƴ9g7Fi}J}XnTS IkJHO|~& x\ B[E0ªDbz,=$skϹޝ4uPYۆG+$<H`2j# 87ҦޝN0aL+!`\6N7|D/ƠJ=Rޡs,X&y,RQ|`I*Ahv&˔< y|=U9k,`%VrYJ 92Tsƭ$5 HAN($ZjʦD$l0w J<1\N'x$蓖8(x#R;B?h># 0,'~#-_Bz$n%>0Rq9{+x6DF bȥFI6@h m⌷Hl ELAh]S=%$8Dʃ T\Q UT+ e^{_Y'f"DFK_"KR 4x Mh쑃㍽ʜθqEMOˏ"|@˻& =?˴~7ˬ,29l6t{(`GgEE<"3#!j}.;dϤ.v#Tȓo@R9ur;/K5dޮ@TT/zvԁ2Rn?0XuHW2^XaX9>g ief`ל@BL⯠ȝŏ\a`ԉHD8#dyh7H"{8cԂAԯ4Iks^ C&6ڙ~,+].zMBz#RGVuD:54Q٫߼_F±:aU<; gz0}_:WAۍ5Gh B1Zxn S956.RI*g"iNHpy7;12/\dלs:\N:4TE^VO?_׻e(s?h%78[(>SyZɱ@cȋe0T51hߣ> KFn=TZM.cYEXJRrԧ8rv$F2Ujq̶vSL#jNgn+x\0[-?z#ԛ7a툰?u?^:+UK/~k;ua;V?]-H/n% ./~s7]󦊣h (TVʨzBDV>%'TE]QgPaޡ)yUUZ$p\uuk9ءo7MFP63׆. h6Qv@|zr*Y»Q\Ej3V.~pu;fuJ-5gٲ^ip+zCzڝ]⯺Eὣ.Cnݩ4/]'ii?aF&; Z|;ij^%crzqNy9w[_Eg"`ڠTH #9Z"a!%sTTeaJNFnu^n'c;Q9zξ/[״}ybvIf.b9eZ[ںShϮ v Zg!JuT1)S$HܒJUpSFjWpK7ʘԁ$8@e[s6 =EcDeZv6 hЁ H+;n:MV̨T_15/1-qWj4t?Qyʻ4QfN;01d+v۞qwFaûmDl"y.#Ɩzw&w_cbǎ'H1UWnG}to;{ԵҐ. Z1A$3(Y t j^QU8ā#ȵ@k=[2Ղf(u,e$8BpZ@5 b@e^NՖndW(%6Uʣ4#L B畅F[-DjL5KjrYs[DQ$,IS YԮ~m,|GV7Mv&UXO;B(VBMpS埬$NXA(fa]\U r]>އٕ'xNn^C52? Qn LIK L#;I%RǒvG!Bw!՘QwfT"/[xdt RpTW[5ָ T1(d4ngn{c)<30u?$N]\0$s"CTv?W#yIӴ>>AWH&txJFט|MW3p$0g_sl0W>~9-mHFC2xJ萊EF Nb^*;W#7@_Ytq[3 }tЍ:Q. 6 z]EZ~V=&q󵂾o=R(BgQX]MmFDf۝^ssAS.TȊM OV7 ·("IC^M53!/:*4//. l%]/B!jTC=\KR/.ݨ™B=Jy C!h/:'+ (l 2;K ~^0iS^4Ci_ӠB  f

AmgUv!mxΜO u~_aH$W))!KdWQ%5n "rH4QFMkS78[: >BB#>\bMd30 dK"bغ"qL"k4c8dQ3JS?k^<ϭ=3 qm-4[O_3K9+ )nYegLrY*$ 8M34*qޔm&RLGzݜCNyAxWㅣ-atCK_Cv!Fn.[)SU;{qW ..|.1s:JNUͲ "O@yqFsEXbEm@>mǛ\Ԧ2K$z? N\XnhJ ;1~$1Y f]/bKu[Ŋ.Iygj<9g]bOKsvHՉ1/Q7u._ڜګ\DlPtL`F |$3\,S B(џS.y{UH]eC;"YJeCnIyx%Rb9 ' Hlpwoo^A_>zݛ>՛ɫ )I@} I8*|T=cax7>=#9 DL%"sYpysK Og;ͼ<~&yVYor|`1N\V-г4 6ڀrc=v=?Yl$7#gvNż\ȇ܊@.{kARTu]}3>WѢ0G!@Z&ŸBKT1QnDĴ+]*85vp-׀@T=Mhg,.G33p]d\>gr}oW;x'+pT ) +p 1OJq[{y @ ԇMnynfqR(C:UG}>}xf~j(+/fL;&fK[G~CyCq{{{HFEǘ7Wwrm}Zt| > idY nƶ3Cp4jton Gq%w v\=UP=<hm\}DƤ $c.LL2x8bJB8iu̘|1`|3"eSjgД 8aBRAyg F0WΪVѢ-mH2Đ&^gF:"-儰n TQ!tB1:S.\lUBKiYE/FF z^ek)#{)`D9**cV#`VrKtH(+cHVaSb$N&)eFy8kpLj'eԖx3BduEr=0uF!{#ml'®ئ&8+}%FXО&h`tނĦVx\E39ضK%ѤFBpFh-T҈%o j* ^ʿ[?vEJOUo~x2-UĀ(\S.M,vgw{y+PhO^ Ջ9?= V>m :'6SHPsb']|sx6 ZADi"4۪B70bv!\wQ `>\GUTx)ʼn&0xo9ኹPeMڦ ԃQ8ԦD=s(t*UsEUęÇX?{FH}6y9g'A'Hr`RVd]O5%K#r(p8X =_UץPx&*NjN )Õ0EN(ͬlԖD~‡K4Im eՂ,8.oAhS,'B_4/|E벇l*5FǮA-zJ*vRL+"Sm.g竩2Z8|_w" RǏ*.~$hu;V#hBSuw<24Sfy)5vU Mތ@V(V5[p5YgȷgP#s g*{Sh 8:Wm7{󅆱mכ'{W{{V wׂWۭY9Q{JbJNrAiea1kU$+AViTxwTShlOq\|ve$QZ /ӝx,81gMJ1ZZbXbr Ar!Lk(t@F"^"A9C6A85Y˺v"[Q,SmzRmN)M4ɽ?u>"\vOaͺmcPjt>=ܾWwrX4_Y}d͒O6׽| 6m،&,k"+V[Ɏ]pI2Es T2(@6Bp,2j=kM6`"M1bZ,f#@fwGbɬA6:DP)ZAƞa%QUX{U I[C ]+{҉ZUuKϢ&9r&|l?ۘhvQb #*(rsIh^$e+](YoMI)ssҵ{F(): b RE-xh'S[R8A&cs+3|>CZPS")r`IIN*J% = EVfu׈_稉M|J&UN [͍ћk 6cm'b n8vReo]9KR+DrFq;Tj6IJi(xV+[2 5BZVdz2k @5Ęuxl$V5ؘ~kP- Z-֔nV@,tV pv9NFWe|<1 YHOb"2k: 䵴 DNtx۬EͮJQ[-C"H4;30V)fhA8S_g!:EQٿ8{'p xlfdg0ciʟ[/zٯ?Ve:/>>yAoXo|v[_ϜˊMP%%$BdguP:*45pRt.JyW'A_f|wD쌆<tf+a&v=!^FRIH¼P}H+7yO>5%ϯg6 4M{a 9x{qR/ x7Jy\;Ry# fbAPup|E VvQnU݊K4fˀ;hqʹޢF+ tfzB:X ͥ4HY04̮PMºcHB -&2{vUNܜ\ޜ/_º74/z?=n#ng|X/NSGJ3e >J֕dWFQhpb ~W:xw)}@³9LI9")j V22XX%->3B@2D־'xo #lb9 A͚{|?CܨbU&񰘲OVbʣ 6l݋i'_ Ac/9lOR,mL8HFMnl*M(!5S|o ]|NU1/>._YO~e<\t볤?zWs)/R6M@Q"va41+ZC/hߟ+CQPrR"BIi0 ;OmJ5 u1_SIQsuRU6Ѯt.)̶ ϗ@XV:ݪ7;@_/:smOl_Od/) a+d"e|)BN*2d Rv Z ڌ\yZXj\KZ b/cقQ)YT 433 ;Algz xM C{}n^iZ:/xn_Οf>NgӯUBn+MqpjgAɐ˒zJ:_Ur2')E9eݽ&N]' }wв_7CL9$u@}quPc5jf'n.:d_*:io/fhoxè$|;`uf׿ҝ6Wa.~ [l[ *HlJɝ:Ysc^sÅܡC]A!mv;o;2!vG LreX!eD5k|鲝 *(:0]B3{aa~+j+j/+Lw֮wή7 RиrYq~{co8R`+;> ]{"9Y٘#UlAZ$R8_ײhR)DrtnNoN5?䴢`F҆< IQ@gJc-E*hZͩvR2׆Il 2uO2)8^>3\>#)hTj%F( \g9?:ZM7 6\ZUhiuiAc'_=iwvnAv7J(y2nasq&lN?ݺk}?9{ ٙJҩ8Fٝ 0=Y9ff6?ڷty>kO_k*+ȴ~y\%L[ /6&Q:G'xJ ĕs_gsv ` WCJ=SE$LJU2j @l0o7Ӭyl'$G.@ׄo:Dj-$e}-P|I8V.4]P)^L6(.LW9 :i2J"V͚BN:$,K}P;,Oյ2u(Wː 0q)G05knmNn{B:Uٻ6$r{G>  Mov?D?mʒ@Ro5II×3=3LD3տGwunV*@qNs3O4WZc,mA7eF*d_"_yi^ή!~ul<a|'c]!%sd0D!b&BȂ$۱7K'4F } ԪED*YGR*fH",%BD__N4~.W{9zШ#yip|c+_R_RJcRD~wbe"^by>I'$U^T+/D6իwzL(Wx?YnLYuGeҨIn&c?Q .+?큰.X R*'F1aƉt35XFA.VXᾀg%7j =RbͱZ6ʳ:|k)ˈ {%ݙbÇ ]kSJ9f.$S>֒1C!"q(XaĹdnܵ*rHXEdH Yp0Zg,Gp/Synp#,u{%*VV_XYxZn. DXi{<\Jkg?^_4+=|N`Aךƴ#}bޜvΕGN_V6Η*Yy÷k!-\fa\C9/s.)k1xCƘ0֡nX2S%O9G)Lʾr\Z8p[MիG8I37swгJýea U}a){6gRa9i ͙̝!}>N8&%G;D+*ISf$M4©+F5>Z *xtZ68@#+q" -86IUwֺĨQj4IP^Rwm d%jEEO7o22|p}BHITպLPlM$xe OI4&B{ܘyBx+IuI1gߌX!_jňlҧ"jj ao=n)}~he} ոt~y-6`'j0ˈ@nCJ(~xZsŌT`-VD2FK LA Ձ( N-ik O7E ${'E@ȓ-_ۿ{BuV8Bu!*|OK![[!u-8~≻T.21wss3.eM-e3J 敽ɧiI4w f6peF>6+wQ|fԝvVWF>IW7W2.(Zȅ>>>&U%<(wïQ~ڔG3~"mj qOKMzqRs12"8wu^@TߠLȁ{)ӫ%gLo+͗e6`ωPc3`Ȁ/i9WW#h7r,sH#@ gލ4%J> 9:~$攷-ߞ"S ӸKQ-UUTIrF yvW)_$PZN/gR/7"t{2Iƅ ųCVq"XtBk4uVG ViG(f.:C0';q lr3vgxcr+s=ZeUfꋗ&\r.ֿ6{9߿^^&D/ pFԑ[# D]j;.`D!$a)E~!3Œoi!.zfuP < r9h>A#Mtr@&@rf 'YHl8AEJ̥u0OQ"& !\sxOuy1AR|KJARѹb:`3AB %HQr{pG[ |^9Ly CP$+IT4DDE|4ts zP.ꂢ8ZyIrMUc:O4V1O [X!GrX 6Ԕ$Vكa XMڐ$2J{&sad< py䙥'd捔%1E$j(L}zeD] ,jMEۨN[Zɛ1#n6F4,r0 +!(F"(!Nqa WtS]wHP" #!8"=)AKBG!n"@E M~ܵas .fb6 %+5ſ]ܯ I3˖t7&^1Eqw-څrxGf9_ ߨ*\eF7g܁ j`ȳ`[|sB: T\wizkM,cА+6fY0n1Dtš}4~CGC|g׿_^U lWU{ֿ_Jzղ+mʥV r Z ө#iPָNhznfMx&L$Ҫތ쭁- ZYᵘ9Ø1wo2 XJÐXIÐz,0%_άpqyi;Cu<(QAf ;RsZi,Ot)E b"Z F>`4Rx덈 ^*Q:[% U}I4s|1ɰܫͼtۢF-R bBC nsss=ThF#KF)Fr"&3x2 aZ"9g7T *jQ3D3Zk9Ϳ/r,xTijoTqq8el~)vӇ&$5#GjÅ\ DuݑT ~&IwxHc;pi`->|+ r+Z)Y$'ǔƚ:;l9w&UXRʚ9:vj׆4+R"<,# 1! t,޹%!ݾ&Lkfwt ^\/usa٠MOU/Lu{qϖM^0McʢΥnɬsjɢ3s𻻽y̗啞_1x|?Ҁ/?cIypf?0GO+-n'M|Zˍ5Z6(~5x'iJX0r}N|nz50:v)=KQJ^χ0r[Ӓ|"Z"SҔx:y[UPJ1wnǴz[kV+#S!!߸6)-hϖr@vpu} q7& 3 L1 dcg &bqF]yT{:5caeM/`& цrX(x$m*M* V˚qwgIJ#Qg.i1W(=EAA]E߳X;* 2[2(SA0lĩ5`JKZ8o6I5'N> VJ%(qIQS6c JqʰB;!HX ЫVQdP e=JG,lcp@RX1j98= ,(,V Qujݰ<W':yPJiHja1Q6z""5#HJ}Dz}O<ܸ/ ͏6E+Gb5z) !\"FR#?K޾y()מ^{>Ke~^]D8Qm"{~ pt j;no>C?}{1Rj"XҾk3:Tԗmlՠ{XjY3v:\Ӻ+ueM(9RA%r?%)`hRPv<5P{~к7~ԜӃGͥk TӁX-A@U>ˠ'U-JT63 EFaT(l4HD"hj0 (3/C!T[mL5PRڮ9ݰTiNU5ĺT4$Pw؋A&ms5+Y÷\Vΰ,k[yeޮgFv|{1hIw}QV=b>]/v<|83<+y+CDê%37/ؗ%4wYZWyH+8re}p5k+&k}7Y^RA:b#20n=sE7ŭW[ {,C.;wRZZV{gË۸_CRHIvjET% :b>{)9¥﨓~m>xGnmH7.%2޷lwK1wn;NI<ո7V~Jڐo\DKdTk n[J2[)9S:xǮSYԵv+OlnmH7.E2E򷵛ndwݎ<Ց(EZo\DdJSKeJ`{4=6W Zz$PM.̤dsVAN-eJW{߬Ǎ|ç~12zz0ʧowY0>:+q 64']D͡ZGy}J%G+vdc8 & *iXJgA@.W ?+XؐbHhs4=$-h]TplIts(N/v8ŽS@0~b P;V 嬂/OlYz(-) M_Pb3K+q ǗRiqd.1Fgj܎De|LV14Sity:L[P6z[8w~{ɫUaȞxaWK{)wCC탩p\7gUyӇw?&fH%< qVTn'+?.V-Y ېc3AhZ:N!"1+eh3?"^D!MͯiC*Kb2\CJb6X~ } .{Kڴ֌DY!pEOOz~W ,hoKl?٠7o o't'7`XsIWBG.W#t.؄R]u=Y T-z6+M V=\ ն}]PpsJڂm,m|vfKUwtB6/'=qՊe(.xS4>/\о~pToXqg\bf&cjW yWox-uY' Y*Pn30pfyM޵x}8Bhv**(Z *Q(B WS~8b 2[GWPNf3{p6!2ڻ QZd5xuZT1%ۥEr;"3UjucxdfTO11ƪ"~ ._*zή' \ޅ$CQ, ɺ5TrV/)K /f%cPABQCTE&w@+1jʰĢճ`ƧGwN5α^nF<{8쵒Bwj䂸DS^%_2߉P(qəuK 4;^/~ F1F0YmF cAn+7vd_P>4kBNuO;mqD/$mOb/ʮEU/ŤIKwy~#l ' ֈupjfЎLRqS)X  Nku!LFe aTr!xVݑٲF|ym1,sp%P){BEq(DRI$p򌳡㳽+QLPYcxY]pрN 'Pg:)0 c]Cr4#?zA_=_1 .0]qǕ߅+Z⽾T:%L0xn0Z ;^\Ë-`(WBk$>-A&C-5kI1+&:yq!Lx{ >j *3,`51kY7*v29r)BėmmvB'Q[sļ|4}o6"D888G1,aW1:|6 Q)4=bo)I+Ѽ]Q9G'R{3ҠN6vv4ҸՂ$DGRv]s*hSi} i#+3s_}2pMRI/1/{c8`"G܊u05xl2Ž3BzpHVpq)9!z#!DyǑق"uBr.œ(1 G)b=aiVW$R^b5Qè)j@5Xʣ\4'p&RTk 6bnA$# h:4hQ:-1hEJG<;EPu~{&RKr39F#f*ga*) RrfĶU D)e1# 2L2)%i9R`cޱ@|,MS!RR`j ﮊMݩvJqs?n~z߯¸ ./,)_/_x'/3@hVG;ӯo/FaE*)˓*bJhzhC<'xHQh82[D[%D.\_8b'FJ:OEDD88*NJf`ɛz'+^ B@"AxjRvI#*Ds :&cii@d+ Hb-|2v _$?]yW~UwU[ƷY2D \ydƓLZV܁+x T_|\fѿ_f aAre(xXez V2Pݎ(q+h7y>)LQdL 愑1n<L"iE3M$F9&dܘ*p*h+ %sJ Lv4S a(EފDlNS+TN6%Ci"QKsV)9R6Ns\IlϭҚ !%ra pB#9S>mdL&(!$3R0U\5rsn &T`l!_Έ\ i$SpC^#4mE $9CX2mёKHKf'#$F2kB9 2JⰛ#*[62(bSRhLwj˜ZՄߎ_-F{ښ@ZYHiz3R浒I@3R ȽA#~V+/`,Wx݊W;9rgr?\k:&чxq ^8Ó ,:bazP{gf5q%(]B">Eܱ6w_݌;ck{3KD Q]@wh7j|Gz0FR"O&_6Yl} /"v j@>!fkF4@/8`Ƹ͇4@^n5GhׅyӶ$ #ݼw т'N9B&GץyS"A~hCSC4'i)N'Nk;ĘD6@g 0cM?(ElWMzf,$Rq_ിORLbIAVT¯:aKGtmnj Sf* T뜅%YWmfafG04gaT c2 %B_!vх{d_w9zw'YMq8*{79zw[Mq8}a`4GWOUqpVh}Ϥ9Jޞ'uqP[/\Kp8 V|ԃϚ:Dcnz*%$R#BrM(JZ*V6u ƨupz>8hu+ˡ}pRlT||>$>#ŪTz4-ե,NW =Ѫ|5gt jMC$K O^]d=RSkrfaԭ'.iqZ%P{$ij{ E3piP"ŕ N$TӲ ͐dGɫKY\9=K_lsq4NC? '!(O؍Fhl9.FMY͘G߽OR'dЊ[a?7aJ6x+[/glO06ό+$OˎPXa3{cd֋7O]alݾ~ ,XP l ITy0J=26wAPBsR*i X-reo(J<Гr[E96@& V3"w`X4W.mZ? qP~,d$Aa{T *"ޙԥ,9pl-\@k3^J/,sܠ%$-n\K..FsR]nv)dL$,iXA3\Hֵ~-+x|m; /Ip;vK>$.ذGۡ$]rIl=Ԏ\:xt)^@8ffEJWj(!<נhETVz";]-\zXw~Pj{<=pW2L>ܕyӧ{EdwցOVgdB̽ES+^>d;Yݝ~ت)5(kσ {}Hs{TqIyZ;S0$XNAΝ%NHKOiK'k%m h2wezRNEv15gӳT^74kD؍TBu6_*@}.jt5NP{k%h}۹xu~Q6 ˵=uTH,̝wp"wY3wF*M*T}kK6$f&a9Fw4-ZF0ȇe81,s.56i h,f4ʵ?AXowyezy1*9N֕vl8L׌Ej&7ah*%,MJy=dbR](lZWӍ-8UTJXDj]%n"a,h*%,wqm߼J{1{e]N]VvW]eGۍ.t{e&.Su:vԋSuK ã;gNj(aՀpLE'| ~Mc-o01pK8fXA_l6nX3$?|ߘ`Qz 푮„8~֓x &Y%d/_:ӓ@,Hu̟"TX54OGg(*锷m g4^P~5 T#~Sl>"S7B*N3TڍBTSOB\@buX͐\"a8^Jbjqjcx+fNY8qk^<:Jhh^;x_8ll_fY!ƁBɄDYEJp"CVvRp~X']^li!} zdEI? ۍGkL/k.N\_nm壻1jszq-MS)BtB8OcB h$&ioLrKix*=jm0$5Ei[H>XsDP6!fl/F_#r4FԤR^Izv_W]L/nNl3Ɨ4&|+I2?w(LіN1aR\/d0͵1ȩfht_\kbH)wZsk2i3'U3llF$r/`.&J݈.VEq m:/RB>9oiDkG? tuc2y+4 j.&8}1Di6_$FkdFkd[,eJ8xaPH'D"1,I#BĂT2C) 1C:š N|ȱB_ nL3OO:;`gm%c]~JӘ{VkXmbGJGF*U:U-Iqg*1H΀Ɍ`F"ZPx(KӔ'θF*A1, 9 @^;|HGaC5碾<:؝,-׮"_aINp??o6Ճd<ωX;H(K€cI)t,i <.?| 痓qt /<&/5O' @62/ zzw_]T9Ā)毢ֻ2.Z3d.|jU}n,M\ ]В)`KCx<¡)KQLUy`#H! 1c-bВZZ-5 -!.%(D:Eq1$ph0KY,cMBн4TiGR&2vDK:."A*uD ~HMO %T &a0`Ѣz( _aƱ`)$)KDa9)XTBRTr` C? $.Y9)h'#Bkj$dR$(.b)0g&!R^hQ7K8WfBƱd4{8q:aƯ"d98_oa10bNpۥyZL}>7y gdC( ү.o.i~7U*шª{+:9Ѡ RdUJuk:[q^_ކ[]XKf!7O_`S/j-T *0  Do/xX }G EXHMۃ7Ą롖!*pv</>>q1ƂmO(#SD SxX$|cKާ9Ў9EK-{)>ۧcH<6L5r]gsn³9X+IxvojI2ZL z ٳ7HWJ*u9>&ϴ:{ן4~qkNzi;vZf!cMt$-: &MSk y ^up ~'8{ӬTolMj@}1LRc]] @E4[/S,!V{v^S8R^bl΁١Ns?^ 37( ZWTLU>/1 zۯ3.+Rl2x$dNʢ~G5`&g_~ٕ#өIjUJx(U'*4}BYX;:<$a@2EOy3p9 z&[ ?hvrpmW.κ8_<;N,312r>_" )Q`F7=OAcE[6%Nko uR% /MOIKBhY˼G¯D+lQۥ'v ;G#ΑNb2,LƄ%T\8pL?xg4&R1LpD0:rn;W%jǞUGi).:ǥ/[39&GCt㩰pKǙ1!QpB 4'q0<A2JcMIƒ{F BMupRUh<0wAhIgX?~WO38 tk+q;P;bgHV;zicKAe(6lHD$y@Iw'YI +'`*d{DRͫv?vEhָse\%bbg*$k([k 5Y:8(ࡤ!΂,J(~~+tyf+BR2?=Ĝ=c| Ң9L'<2" (frf ^UpuePN^U FKQ]@Q駬`ed_\xMyT@bcWK)3e̒K%4JeeN2jΌ ` ͍ `/)uR50q;RBowbwu qD8 Dł1MvT&qg qSGqյ;s@}Fy֯:5=ݾjwA2!]ĥuw24mXSg1Z;B`P}JW]TQT NySSK3&Zʁqq߂_ iV7i/`4M69vqW"b8%) gpOVOfaŇIUUbDV:N.7@5ˬiř\"tl5Œ$Aƹ6 h`g)epvy~(؂#:_.HF;,4^]8bHa9W%R?{ܶ K/s@*=J69k휼$Ś͵DjIʉ"A E" b_tMvl+F'ܼl3(F|W[d7 J:M"Z#CnD@e1Ge(2̬W$qǜwye5[PIž%n]%h}֠RH<%2YM\#0> 2a:I%C{ x{4ě-c+{;dkly,|HծNЃ9izQCV4,]$:"D dև7}ݏ}̾еRs^3^#F%REkF]'ԝ<59ƅMo27 k:.R >JIHx׼+5"i_ ^fOD%:`&ܭ{@Wvu:[|D:HZoc%)K]A ʹl̋u" 7ZwwpDIM|:`D]v*xe!&>ʼnycnJ (ϘJ2RLFĸ'Y͛߳};QʎnvkԞX35wXt;Q?Z v2/ozPbAؤ^3[8nFI,5]39)'Wӣ&-]mڬ5vgӪPmw8?".Q!bHrխw9OWI;.=?:R'*7(RKgEEERt_w :de6;{c}oqx/| YﯫχYwMC/3Ud[ u n!-dP Ǚ4p s3 %^R YkMpQDE$H0cSuv%\ϖc WiR( 1y̢M{1E$gV7 1|o4} V%L™gLz)bcr"ag+n py6?f ӻYnh|77ÇK=s1(᳻Reqz@z>`)`ށY7}6FÜA0(AdK9A9=2B9 ,z@N-I=]UK1|@|?4HC?4HCbPQM\ I>hіl@o8PndԢL$h*Ee n:P# n:%2LvQe0>{M1$N5`I9HI9(&m-uA[lÀԑGEVFs0zD8jiDpfKD-Ɗ QRlD%FM*2ƋbP\̪wsbB ډ޾tAi|],\t˿~|}vQ)eg_2 NY=ALX.:pUԵo.n_\_JJUku iaUi@_>YFYv+( +:&Y(qE$(0<;A5v5=j~ k*`U*T!)P H{iN!wx-$Fb4*a%0Wa:zU%G%cY 't` ` S`> iQXQ=fƸcU Nao9VC0?yw_cei 7\>A_ zoy戒U<!L,ce71)6WmUQT_K%k=<}ZJ2QՒha5F!^)qHK`•(ƒ X=EyKnzyvwD˳*jety(!RWYD2P.X*d^. Pzl'T+͘&ԅ.XrS0,<66ߖi\ 뫟RԠLi,i˔=+PcJ9%? ,Ĝ8s x@_S@4 L~fV7#DewhhV|]Nh+˿enKBWI*eBLMfMP&vJt3g-Kf$Q&SqZJ]ϟ c/NpɺgH!Bp`.[Ƹ{j z= blG'=GQ$hǻ'<7>$;T|ڻB+L"MLV#gp)ZNkeDa5΄4K"c)]lQ"%(E[q2EpRY ňNYo9!Q!q30FGUx6KmŁ` U.(/qe s12+%$; ʄ[=qY`¼ b+%=B9#XZ#˰ ԃ`Lp%$s.lc#4:pàpnjݨ]4hxäBi!aKY>N =fXPp,U~կan2獱?Q/)j/V+;QnBN"/T0,4$~FL-')Х&N' nTW/qxF c} JB2_*B>.P/4h?|0m$5H5H5H5(n2g2',i$#O RQZd)2hQF!)3(/yp_/zW)#X ][o$r+¾$@x,V 9q6>X<+ 3ڵׁ{3#Mϝ=d,ZfbH+m2szshy~&y63@n|q3-򮘐*[gn$mSQ&At72P ee9W ߀2QN<$ncq;pTm>A`jͭ^YUc'Sh"Y7+J8&Kmd<G5jp˂8ᷚ1`;ϺYbG6˜&dkkHZWɟoU:*^"h;/ir8B"`#&,I0if|q78>/R巎i:o.^.LSVeۧw^b Zj=]\TE^QN~Q.ijOWce-W8+IU+kR*+!.6Z9v3Aџ(BZxr Lנf@hz@91)9z֞yaf-, D8# Z*xޯ`t:)<ЃT[qfrq.OE` Z_Pn=kQCOX@R%9^ #$FPPbhR-&QjPYTft,/"z>PG i' w0` QAT3Xac "*NoVᓥư-߸I'MFvڈMqBDI3Di-*D)P3F~n)33&40PJH$%F ŭt$b[XoAtڂ~G0(7F"T[LY `\HE;c"p_9jLe$-Ϸkcc45W_sO`x!WUxl#5YOyc%Bϒ|!=˦kqtVb^m\[Ϡ-M 1t **l&"~FB>eB-\ŨEEh )ήScW̿.z7jf'dǂK,mBoIK;zL;^ʤ@s& ̆6{|MA^ o i׺h8 ꑴRNô&*g5rVQDdR ̕IzrIgKnf0 Ց6"oP `n%{{?ȦO Rn%L՝o3e[ۡ'n;;N9mH {pD_T5F'fˀzh"@4XGU$ "@Kk#C 19r c1.;uHSA[b@xw5BsdZf5jKj<#wI)9EK]rS=\:5,䅛hMqD3Jbc:mtn֎Pޭޭp-)FGt+;ޭ)mw;1Vҩwޭp-)؆wc(w+ tJηיt*nڷw+hwa!/DO)'`kAz:"Xz] 0nEc`)eJ/J@Gߤw#/I))zo` Om'7mG̵>)Zfb(Zl V;B E`-~s"q͞6[伣u" /|`cD"]̈́xb+V#Wֽacz4.^+L/zxd0R.`>f_X<=w1 V>CY{'w@R.ԨF,IqB)ոTxRzِJ;[z ѷR(iݗ_\6G)*ueNrǨ.í.N* 4̾>ṵr۹;Y ic̐kq>lM>tjfZ]Ҟ 귳zw5&?| V__|OSjtڡܹԵ*SL퀊.*H t| (*ܷRMj'C r1 II92iڗJjWPRލ] 酕:ߡ QulΕi8_%w}]RYh@ rTK|V~mkFm1gyU#&WrTdgU I5b}͔8`Z +PHSFZ8 nuMX;?'E~ooT/,lʙmVҽ\ZŔNKVII.|_Z QMɟ~pӎwŝ bj>-iIL>Va0~&~MGAӻ |仇Ez sl;C `84irLc.fOvغŸ~( nYpߪ;Z)C!}r,U/ *==^\p!jO^fG9Low~{G@U [<*?e2*iН2y3.kԸp컾R\g-/l;2(ָ6s##XhW0# ]qޓo)*!sK q%E#v)(4Y)a?"ŽS[NHTs\XQ뼷0%yRArZY4;&OE1L1r x`k.LPKSM]u~o+x]aC>n _N}kG);N"Af ͷOPmf'RڨhWK+x1B k1%8aG_7Mex] F[>,ͳE$tԯ0w͓7}sL/'7re4̪ylQy采:s'"Opuvs*ziE=*Y`%-z +NUߞ.Ci <ˊRw.Cs/:7ԍɍ O?66Bz~~t[? 5. Gl]31Y!V9'if1g|uO9;ك31 (֎R :`2Ê3!v+oᜎA@L&oZ:Sf P"MPj *gcz8iuDzCL?ccg4@;,"ŒZ:)jkƕ Vi X!֒P0gQЏSaf%o |rbTr`!hʣW<ADВKi u!ra0/": OXvz76Ri/N kffo~Ì;ZubH*6qL)h~.ʃ+N `*v6jp:6jh`"ZvVVzh%ʉ6A9 0Ұn"K,kb@>_cbB9:,m(bs 45?ṈfO']9h=MM rgY`:S#X ,1bK$ u0d&ƌ9Ť7'w׳,1yOl/K"˼֒2}DVV>u1z$ DzzDxzpijUv vtemI}2x}kJ%mLm3Xܖ ItcbL*,W>A(OI8^ruG(wq+:2߮ n |caBX~jCjPk^?#q[)9Xmt!n;ߩG)WxfհnE6ڗsۻ)gJ116xm&$21w+Pn5,䅛hMqܲ@jբ ~_#AgL1n sBH`J1*_%n(aD:ccʄ?|; .wR {KƑ\#Ш(AJlb]mIUFǚwJ۝';$]F릴QL5V1uӪ"%R9LC5}8!}[3if_\\"ԇ P ,E z7P*E0:?$K {x\gE & ҿܐ'![f/`(7 dU}WFW0t.>?2"C %x+5=zQTwmFW~FJ 2iw_ ذا[Qkd4ߗ,IT]̬Zݕ A2t7]zÅiG!:웿ROԻ_q*iU5pWExZpɖkӓ%׶uQ1riTF xmrA,Zf*RUkFjtrLuKW݅}b]2-x^Ke\qw#`F]oQ[Ðq :ږn}딳lI mZ&jcT"H .Pƥլhӱn8Y~(c(.e!zc 傡ы}XS,D}%*5/m 3FBNo1Gڏʡ8ĎAsG3* #D8U%C|-%dS0Ycq_a#уǔHmk$ G.eWSݫnQҜEtȘ(P(8'q,ڠUy5D*f&'Yˊ־h H ѴbrՁD#+!+3VӚ\ؑkak~#a|t )nq""hy/2nO#.0@P绻w; x&? w-FK "6N0lyǫ \ܔ4=h *(sHF,Rع9끓s[yX#g:_[R͖w<_FG'e\'' ({%4^-/ksƴTr@a2[&J}wDsK= av<-p>+Hϟ?JQO5D0]uqӓ+pӇ%sɾ=[]?4c*`H9ڃrN?SsNRAS*hZs4xW&p$ㅜVO 7I⨽xPCFC$_@#%@pΙBT6E:b  hKk`1X&( ^D$"I>3]~8ZšNgD]S0t8k,h(Bgei :uG,o[r@9צ0kHH`N@Y `B$HkU(8hN̰a5f`vGưȻKѡcgJI 5f>!@Dh3'cG>YA c]bvf?,^R4PHdvBR)xKB<ٳs*==}7D|9<#_ R'??y`lpW_bzїe1 0t9ҟګ??~HB9cO;vq_O C =ܤ糀d v+ٗ< ȥVc_F cZȊG$_P3Utʎ#6$oUSuH ;>r9\h-ѢW%KC)fGտ?|)?qôTָ{\5B8x6i-xu΅BK_( 9r?"콳<Y)Tƶ͓Ђ룁S`Uh@ap-);g?_^ |{/p/,`kP7:@Hyzm3,Ze 1m31 Ug׾* ]ÍQ{̈..nM97_ 40`7<#161٩5=4-,V[q_>X2P}mn~=pջ`]<"%-\L VZN`|fcƴͲ,MG>d|" %%IHQ [ #ZYbDcΧJr]d%bbNܱn_a:I"% zE@5_Q)`LdJ2](%6p4Ť6GfdߝUK[hcg D.0pE>FJ )Z%H6(}4BX }pBl^{+Ҏ?~8[C74KG?~8#-*O*69\L$ǕLKƞY/yR@d&X# -rU%r,ݾ®BF%:e~jJf|ܣdƋm8W|$(rGcs1i%ͳ OLε,Gzs_fj!PF#sBF$[0N䝊`$SܦI2mR@u"\{1WpFC% t_'4P^t2E.=G@%d4r?g_~_wrfDUh&6[.5c{Sv]T[whbAvP7XntG1Wӓ+pCqkT$0-j<(>A$i\h 4ikѷ-XnG RisV`9l8|6fgks=rf|5!rvW  E FUWbPjT9z^n3L 3ysi$IP^xùP$9 'lD ӺgDFæ`dnnT'ߟ 2͓dcɩ8ɐn1XeԮ =8Ԥ # RSU1?*֝RLG,[ZAIMw)[z S#deKo ˖1BS fj |#ZZ JN;X#QLfݪ;Zj݆!o|4qل% B̅rddBBڐv9ra0 )tP(iNc2<.X쎥.Qi f;h9I [Jkq`h%1=S־B+ --ViRBU$ wD+ёVa)&ZZS E*JZ'ի:ʧ6HΊɒcH$B <ȴIh1gFroN V&؈:4'U k+ -Wi/ V{dgdЕN8)%aDzAqCȸdn?nZ DܧR,󜑨 %ڈJJ*_}T$nz J 'h ٍ ]’-΀bR6x? i&?Zk}~<8/֪kuQ4{!&GЭ٭"/ƒȚs*Td5KVYE#&9bf>'NMn? AɓIXqk!)r A8hm˦X>\-q˄iZ&}rI{k"V#%rLwHA |O:3"qV#{ jIfۄCf)(T!In`1 lS<.:9jgKѩNaT{hKpg69WaՏ&$8FEഅ-m:|pW7=ٓ/ 4_͏F޽l"7?uvPZZ/۰W}䣝=Ja=ϗ  |ٽ={4&i/E>Z:Om~ }Sê>udTr CalPF՘YùP*i\r+AUs.diSCU#>18YF1 3$c.%tCvl` 8p =0ذJlbu}Z:CZ!u5SA\W듆f=CħHҦK,0a]ۑQw}ϾJpb45Pn+u^d1^]I7JIj`$*s^b쇋.D.J3 ai=3%5-֬moֆT$q\=E}E%>ςP{XNH+ȏGG$ˏ)t ө-hMvCNMV Unx)H_h(w+W/كL} ytۋ謁 jNvy")>`CI0۞B41׏RhIA !yQQdI5~yINxAYmU#N\6U2g>܌9nЭ4ѹ*ynUaܘ9!< /nz`p:]! 96(j$m-s͏Ng66$©%7f꨽ F^VXg[}HM3zm>]j#Gqވ+??;׀C-{9pv25 v33P.z_-8ʲv(iaK1)%ZcӧB eE:樆dj P9R#(j8[G-*MCjcm.󕴓U@p;k"1PV{ t'z٭ZK]to&V !3GnIP2!=+#&HԬ%G% Fjpix^ٻ7r#W9rŀXCKvn>\Mںg&Ɖ/~ŖFjͰEve0v=VwSbY8nB>GUjy 8*x%1d+`:`P`sT!WR&9 mYߢ@M}[]`yߢфiD9ڈȅ }[*vMR2tfV"WXA 7,B(XrB9?4S,vC)quwEaR6#}qG!*x3c4){xrfl>ԍ,6"57^|06&T?Oؐ=GwD)OfzY~j6Nͯ^/}U_n&ί~[pg>@p5?~~sm7.ߥi3F;a'JCTD%ht;Һl˸LmwrܝB ZC8+g rAH}ᣁCDIaձ c%_UZJ_OU8UL-ڀIćNdMv}E&~)LVS<~>L\\ݻQF#MVyP5l!I.Rr?C[bFs@trlc7 sAx8nv%*n_tq $ F!/)dd!Z Zsj7(Q9/=Y|#:NfGٮ .;Y$j$:CM*mI͑k7ts^Ecق'}ymOn齋wOwb3|gwpS_1[~v>Կ}BBͫ;?9P 6!88캊B Y ? FaՊ|=.g_b5(|a= SBv! M4˦ C]|nq5A>w;݂bikOJX7n961"w#[.)6t נS"Do-k7ѻoDsl2<]&>-}FvЩj%ڻWc$zhMI~n?җ5Mɑ@!8!I|/| ϤhIrȶfp?wKLg ^qD$ph~ ҁM \/ %Z*ܕ^BMΒ/q6ت+/Aa%ʛ\" 8zCԥ 4g_PQ#Wf26{AiyaT/0^"Mnt*7sA;K_rhBTctQTp^/t%KTĝ.FJHg3v៷nQIm0&~Stk>KbIҁъ;ͰHOtXŲ>0A[6o϶CK4ڟZ aǑ%1J+` Jwք`Vp"Z8&/Af%}=.3%u0{6 /"pc)*=h+LRn . i׿p]rȀ*ƺ<KȔ#.>DY?)+>OB=Qe)kս+pOitJYbe_m=TW(B K[;󇕶s t#C*vm<N_R)&U5 >9Х TEyEmP s+ l8G0|XU3⌄ĵx 4/VXd/iF0Y)USoe޼WhoHj\z>9;z lVjq .6dڴ$}ZIӴ5eJ`0m(Wn$ \I)*% b e@Qt\l- v7XC[8U %Fj Lp 'Lx|ԠD$m5R0UR$BRtD%´Nh6r\4!XK| x1m>H*5N H:2rJV`R inuLZ-dpfޚ! K<h%~'ꔉz'=}KX7nY6ulƛ [.)6tk!QJݲ_4ջoD)lDAwۙ=4*f[FcX[O(pD31o,{t-֖f6iT꡷$w=hܛ'=. ŬD3h!b/ǽ;~zԉwAc7l$7˺F 1SKȲD\w@X sDhZ+0 } 36.T`L!q}91DW5yaD6~_cɆk (AHDm4'$hKp &Ǿ+aA\1(iA}ڣ {3wz_Q]T&DZßG]G]G]G]7նNƈj6fH $(J mp *P8}] j.n[季-b[Q~[*mP^҅{HI!)Ҹ(vH E%Ay(b耤jgN;JVTՉG_9*D9=ֆ`&ZK⬄%Y=wТh+yK(L~Ea(l` +$fprp:* {# T#xr;bWi)y$ZZc+d|`<0  gUWZX^h!yq2j$ -7ϐ-kE]nqf3NC305Ώ1g|{I%oW/_fR\؏{xa0U'2t6$YޯaUP}.ֹ b-8Gf[PUR5ƚz:Cu_-y`c2Z^=>cZϒo&WT1pd8~y Y}7ݿ~c[@?n>~l|mcwM%=wRru{ܘeǍAYX{?yE\x%?=½P\h.i"8ՕDwMcz#Qܺfػh--qAx\!^+z;E$k;ZW0摖ⶦq?C)vh#DŽSKYSrl/Rce2+Ә@׺'ώ쓙} Kҙv1d\/8BRr'e6Gow[)6\yR5]q3eJe@Wu—ƭ0&v=2`Ƚ*l`c7rnC!6RbZVR/Ep~3h}]z_Go1#]%+hPr)6I!b#y11 f&Cj)$E: WSAШbɂ@˓YAvrZ WC^lV|QTH;8#*3TJ2V_ɏJ rz {zA- :~QɘwN8aHY T*8iw 13~ ϐZs;c=z}vϕ-J?)V0^X-[2wOZ* ֻϳjH5n| KI{y%ȳ,_w˦2$;y&)'aFS+5W "Ȓ$N>ZՉVur}WLJRP7**9CL&Plg3?&&j 8Qw+ !|i,^YNGR .[k@ JšfPTvj9p!]$#d}P.%Oˮ0"J)%)Ȃ Jq:Š6* (l2Gn}?\e}n]e}n9(iAG 3ėz-UDyicrߏ(}D;VkZIJP.tFn郏+^oG>)>}'(,Ɉ۪$n-j4l., 7wxf)MB`SDLa%?̯U=|q Gt=o"!}qeyCsz>=,>=O?8,mi8W[OOLOϠ-NP5 xUO?@+kHu{N 5|L }zr1TF@9Ͼ'"L<#%а$FP<:0 H4dea\&T2Rzf1a0qT*p5{79+no?G1%=;QSRLOsU $*ޠa1Zvou$rA|:+M)OpiSf Z2ݩupop݊ŀP~p灌b[aJApS05"gR3/%k.tÒR %UΝbCҾlnsX{"*;\ QZiLd^-Lg46.fbe-*vz' k,Jɗç9NE?1 -EEj8Dr.!{5\ykQFbX5&[:\{wtrS^{k@FA$>$HHYb'Lj". &_bXQF/uz&WzxEHƗmI.'Dⓓ};n/no>G[v{`4g77ws_\? W킾dzC1Is_?k4mo*}uݥi.|ݐFksHM+m휄B}.@ ?>; pS?lrnpw78m<3 J|DA똱P }>e^!Nls) .-cPuBˀ_Zi=D-YD͓bs,mB)sr[lĔZQ^>i GQLd}`{>xXK`RF%S+[Mͷ>,Ζk)Eͪ ^x~2`F@t#So6 v_ß[+{]yqooM\6 "*8NLW6zMPQMXm<캍SdzϮal:oQ/ngTmI`y:һٻ--Q BYsbsP.au O?܁E7g_g3}sAaF״wBJP%t20ZJ  T2#=wXe J% N ؍VZEx\ Kr6 &sF/(Ơ&0Ym_'נP<~kh K#ލq+J9Yq!UhXD;w\gU6|H&[5`CUq8tCt)_ zp3QRpr.#L&:( oU`[NzayJuEIHcR0'yT Y1ݽXLwJaͳ~>/a:qH9.!#e<ζۋdI8~bј*_-I|cs՗FIo_d@ `fϾL+ҽϾLz/r-r8dDL,zܒ -ӓq1jkT$d*ā9V x <~lJUіV.I[=A~\}LQnF Zt(VOԝt5*8|F^ׂq1wx(Hd@bM*-3΃@L*"ZBS2K4 k0ku$+ Ch)bIcO(AOM${jvHS3-Ym9;Gf;1U[7Z쵱dM DSupa|;: wb>4YSQ8BrBYX*^Y>$P6>R-QY2؄kf@^l}>w"b~P߸'Al1dL3 sdG +m!YA}| ^w'0E.K} "KC_!ieyRmb\2\U Kr6 FìCʅL4u0!7jL9Sx&/oE@Iv*@C qk61ʣ`LjM1i1 &I(J M,oԲ{8r]]R\M4#a$6U䈾${dȵT NBx|[|k.8+2gj{ɍ_.@$HO&r8,._=ydzE=Ŗl~Uk t*Sc."& gÓ0qP註b:ӓ+Y5J BOĎ6xle<2n-PK[Oq[iv-ה*<UaãAUg0'GxGx;`j) o6K%q t<=rb08qوQb<Gz?&x"?&)L뚴 ť:V(יhP j-ڕ؋ZV8e*CŦ݈A@磌nvʨU B=?ḖN;Fg3S\x j\JMJNeԸrJRsxXƨWȢZ P057FMK6E9YRs+F4FFDz_Pu/mN[#vAV'xJ"_+'œ[ct:;ϊ?^ 3?܅ 2Mk%͕߾.~~ǴypKCZfڧWo.nӗ9Z+J]ngP1_|^jTDBq.g<]JO9F@+EwV_#قlh , Bkyӆ}_ ZMz$l5\%wW LG#,O38o6Pύ9q) .?q1qP^6:FT\wc::Xz :==ъiQ*ΈJ*唴 xHڳ[F>-~\+Nޢ>-6D=Κdžsg\ ?^nٳHP6V 5KB zHDs-㸜IK]0[ F*G/&*1N!9,UIՠ-VZ3  6qGh>6`R69|HYr9ꉕ9'`^6gBpkObKO/w=fj.݇_zh]=s4V?^_5l*0Np `) ,JdpP!捰=Q3S.021^H  E!5gYmw3mthvS˻{ܙZ먤Τ2:{wMV7M]`Pl^^bqc=_!{ h]3Ӓ%j?CsK(t7Boxz jÎ R))弮)rnA%T A6ySOP5tc9LU#}2p4:b q;Tŝ%'KV{(p/@ABc_|#(W:BqɶªBaUnp-4iϪ:V W7NaGZ~VRsZiMfũ&,kyO5^0~yoOyEV5wHJ^ӗtY0*9c_wW9 8asVzM>g~:)a@[A9xQӠ05%oe_(j"ۯ~+xi. (䎣QsGWbb!>'j;%JF_FBN;IXbRZfƌ}G}7D-wݞa2bqY cjrley &7b*9 $&+F Yx4Pwd+\D7;A^d8{#jl`_9'IC %k`&f)#XPszBdP>9Qz!-E x9A݈*"2 1AA%T 4zTbbG"U—tPU )h‰J#<8DXc|#&xI RgD)ɹg5T>*{P9-T4~9Ï$pp%d5S3Z 56V5Ruf#B-r3\ǎx{w=Іjյ4*?nCU[901|:kfpEu8hvA[qEI+:`xdž=1s7ܦW.-eHf)U,M=9_fyu#.gj&L͇NK3`/Rk_)OeFעg9y)F 1?֗ @'VbJcsc~'+G_щ)|,Ǵ 93bOȀԩkr[Eh2">WU322p}AJYKANeYliҮ]gK[Z|%i祖K*%櫸rhёJ@&CPG~hߏw6UŤG%/f-PcyZ RK޲:]t§%/+:ZN? ̀7 v+vFDo24:||=A@{*vJhSAtB-+xj;d ~h G`]9"xl(CI(9<"984Vi>Wp*&2li"F2l^7 E}o0xRsG'VY<|Zb #k&JI+ѺMd>嗇{?ofT:ݻ/"uS|툪 :@}im\~9|m=酅V[Y;7>6kUBsMx7Q[_ JL7x%߅ɼ[B6seS\@9xNon;:ۻ?/nCX;7^6Uu?}@»Št~w;aD`SRz!,䝛hwTH1]ד]"<24 ҹ 8'\C=Z]FaN p)5RT͏ť"-Ǡa+uneSׁb5"BY5[ӡm DVsQ`Sm. wE |[enehHkŷ(5ߒ7c-\b9XuQ06ѝGS(YY!\BTb.V_",DLxobQKe$t,: -NVYyFj\3rcFIĒP)"QHAK?@EWpk&&n\ aeD K7ITZe#5.DKh[-0с:}Vl]}nԶoymrz.n mWԗsYRzjKR\2+[jF}{U5˔wJRȂs*|U`q4ȗ#B#ZP7z4?V"\he . lБqB0hQ+yԸ89Q3J6O&i\(x!lƸr֓$/*i9菫:ͱ.Fr*VVjN*폋T3MXo}X ǭ>^>mJտٻ%,+x<=`TD=jGnjC jQ%QdUsdHDbIѶ 9ܼQ)}ugH}kd@4;89ZoQBUc9ZO~s#WN=sv|e @;LIB艑&z @HC@BPl$6sP@貨RXj33qtj5楥53nl'FSxJsZ4BR coxPpOiTE̔9ZU\@D_Ң4_RZ'˝CzIG %wUG >|yA%O)& c}^CAw{o3Pӆ9z~H(7}4gp.Tc;ӳZ.zP^z^,S6)ړͭO+?:'hH] 1}4y.b=,͖us!~qY 4{c_]~"at{vQ]hVrϽKxY棹(\<)T MLU0 }uf\/Yn]~S,9xҜBKv @t+{''܇[shB9'P[L8~zôQkڪ%N7>ܴ|F,oB0P#Š,img&s$eYl2=;IdӭSoFCd)F9gg-YSf2\հB)kS1BUL}qOiJU/Q-# ̞1U 9o5,e* ㅿיpYjM-(yMa2T)*$t8)D0sC9,W~FRneTNmM~Rr]Vd5xr8!3qgtmݺ Nlmv[=ʻh>$JR ָx$bO؏8,(uͤڼ>~! ]]G!7/Y$.ϯ-Fm}'ڸ$JpSRu+N*])ER*]jF]w0 J+(7J] $eObs^t0锼\n9(| $ rQWԪphQuohFk>~.FBW_9S39G\8_om!q"9cIH*_V/`W)*n2]~ǁK Rns)A;;ZB(VAvg?Y wN@cNxNmB?FYf 5M$'9sr , |a2̚`ݼ,Pg3-EZ"4ҷd1LoVA3 d}*DۏGN66 ؼ#CGo%!bVm:v("[-`u~|3Wt C)r0]#ܡzlpƹ>+왮l>%+y7K؝ W) cLdz-=FAd@sW= هZvuZVw̶ޔU9vk֭>[ a=JNri+r´^WRڕ%e3JA!Ph!$0_p?`߷!O%`-IW^*IƷǦ3DrҞ&Y{7)2ry䨬8y{ڠ8y)=vٰZRhn4w17+#JG7ϱf {?(g.$Kp&D scx8s!oϙ mkZҶ%ȃK8ShH`¾!7tJBW]] ZXN/q\r.EW,JWw@p\p%!ew@? \BK#rnz1,.e#zo7^C) ۋ fK`c Hh"eq*%t F,ب @@fWpIB)(`H/jjGeX3@=HG*a;E|Se\ yv/Rω;6b9wޯ<, Bc0I~yBQ1R,P Cs Lut Fq 𔁌!.[d3QW-æ@ZIyEǀGXE,@H )?R2%,_{ƕ:Ih]EsE|lp3=Px:>1][&>ZD*LC˹ _ɽ;wMfjEy~ۿ{Og=Z;>%0Oޛc}(,*wn/Hlm {_-7S066H.hcg휹w5S'!YO1CL!c>:go}އ++{8Ufrabh~$~^IP#GƁ"*s[lvπڴmf>/8\Kiv^~]Eq|~Ys Sz/QgKfbK'Bn֔5ezkIWAo+= Җ.ߖq[Nۚ}wƝ\ԝ%Vc{p-Rqӷo^=͠[y᭷i{m?>8ɹ;GhK' ] 2s`}aW8b +DE8BQD`K5f¸zj#C]_W/K1d2+[i;w08\{ؚtb=s҉!ݨU6b6mr&CH"ȑHE-%NX3oz/<&cKAeĈBU]w@BX@%h<ǁa`TTr9JPʹoY<4Fo%4jf7`*cـٔfEhO2W_7ux+@h6_x]Tˋy^Cn9獁yir' &] zc-/H ?0&5^DC=YKM5&A }#YD?C'șM]1F `=Ћ IX͌ ӷiY'* >"$ FHN;ͼPy~6 U1ĔɘQ#Nۣ QX(@eHSD( cHo{~Al.KEL/ܟLOI 5k&%m^tΐuwDĊ /O8 IG(0?F^lLpƦLD]LM"I\@?i.R7Šs. mL8{hSUꥨS,Iiq6j[h~Q pMF^zv*gk)),JaJo'Wugg)W#zr3R#Uyg]gj $A*=R!QS{O mrA$(t{{U|7˅L$L~lLzN8e{Ae_m?~:l#F+-')\u#i eW$/':xu< y$k߷Nצ.x2 v<Ȧ cz㸕_r=&Y )Q lV"KFAgFRUawO[#MOX_XWKC@ެB&,o([UC\Ga*˵AՁX/U-.T2`aWn0u( RkԒ5änӟ7ՎMLo8 jIǔǔ\(z@8;iOp@tT}U``5RKJ,4_lI>*T\XJ4R8aZ# )Lu5=iaE ~OR=lJVbƗw"k'Jr0\۟.N(&6 ;n⤮ԠbRjc"BZEi<6 Ur94MS6`5 '?.ץFd+{@;HF URE ) \QK4S=|(ׄ}bǼHELH?uJnwKR m,%Oz^܅W/&N 6GiyIC~U3ͥߥ sjOz4-Hai=,+ԕDKҊ,+7"8ylkޭT9SFvyg-[qGs[Mes)3bzš9oS_CbESYuHўZVMN4 uy4FEs̫F5T@+/io}ǡX:#ZCp}sx(!(f 3MQ_&o6Su*}N6D@ܩU9b (W(eS [0wҞBmd= [Tw-[Tj| 2@h9Y@2c-]qKR{l}KlTwu  [LCE%R8DJjokh/r(OZߒMC2f۰Gj4k$G>ˁk"g |~#9ᓈ mO2 PbG$m;EO1 h"K*Uw ׊Ɨ\ nt\K%훭.tDg]||_Q=(M5&4"k, h!nAU{xx[I9mM\+WYn+@x(ymæѲEmwoaZ=y/7]8|FӔL G8Z54.&ʻ( ?RxtMs׊ ͯ"5VϿY&mc,/J|@H09lRA_1U?0ks?&WĺLն (B]iɗ|U ʜvB9`k;qoOCK]>yϞxSN?(LvKfb,spF"xH9 %ne:#9eoY䚋LklG6S:W0u{& ||4_vy";9}ϳZMaYDnc缫"ojX~҆͗6~k+A`/09IVpao;AajPIc轏5^6ޠkqzW=*mDz,p͉st˱aH#Yhd.-󌋔uIMshu u4u[: tkЭ~ok{ם݅ n;yvBwva"ں^+gV nC'WzR5 DiaK2Yz&KڝUh+\y.O *Sr9\y= =j\-h' Kk LQvL{iެ0z0c}0vCNr-tAЁ246x;#*ap0?5:|mdϘYʉR 5GVOwԠp0⵮p{ߘXήKMrq}|cw%Vck]Q\~8\F68\HZ93)Z(ރZiF@6r|uSARuRM3:%٠l.\\Xcc/~dZJ(]&׋=[wpA\v.S;CׄXʋWuV5.氯;cڻouB2z[6'm~o3.Ezڏ}ъք:Eu}-Ѭ9S;B0T `V b1&,_JmqXͅFH\gY04;hSK ܓ@&D kRuci-xLnlRkk>NJ5lY>X~FF;߻i/rI\ɕA3=eЧd*vwV|Jv:U{zjEz)?.ckZy.1HnmJ% h2ݾrJiPR[D;<~n>\>:[N=\ad>HKyƻ%e>KvI*+'Scq 61ɻ ׹i= xF.~w<;>;8NV[;%Ѯ6Z A;n}QmDurPUv?rq0uXưu>çMI@-AV6:<ف"a \W"UsZ*Z4h? !/7eB偏QfV薎+<=(#we@f;@iJǴe| 0;ߢ9+_.# =R_VV@qD.VA5Y倅R XJI7o4,tCArnש|wzMkcE U$;Jg]>bwq s_ߦ]ijү =SXKiU{x f|bP.*'|É٥ Jp(Z9T9Z$ZYrS0< r=Ctv3z?G1'$_N~PvN4!S|!Q_̱M3d\ښP״h JFtj4h7 Dܼ!3I*SX‹-LixnHTNPy#n͌V/qSsWCډ#b5bmXtMhZϋIT>VنdE6RmLf3ԖI>|&A(ґj* *4&Id^bDLY`028ҍ71jMS2:!$"eLRPkRYXh;D$ uzpn i^!e*8e ԭΠ1);iU) _pR3yD˂V7N4ƣ3Ri5yF͢t@k-S\4K 5(䧩-=k%1QG䍭o5}xad4j1zIYV!'O{%k{)rKD\J+952+X춪tNͅȴqKVZCup gŔ%nƙ 6j}.ƨ lgc|.c-a\ J6Ò=n,IX€k# VJO )ٶ/<+Ej| RϸbsOm-t;j!Ɓ⵰:h$j#PR< r81YıaZ>%ZrRx}i?{G 3b $YlyJ`Vgas2JnbEcA/ce>-~OrO}Bڢ8xM uzss<^ ϫxvWo1x{{k jtIhe^bd,ۋ[}M@X>d̂^\%x× LV>wSzyi)r>޾gސ#,hBiǙY0reT "Y0bBEr%|@IIٖE6-܌UmSbm hەGT3NS(:hL%#5@xiJ14P-eh49t"IBI/W 3,.-q9 $HbВ$-!PxTȣHS,F|rg%yPE!HׯM7@x>e[?$/]mђJn~;?o;qaט,0 ũc,u΀a71g;˒YNqC>ȥ4v0(U/1 dŋlzɆ N[-Z}- !Cه{|}Y?w}7^YOIxVGǃcij|bT sH5IgZ?Ԕ}hx~uW?$tCe;Qǚ^*fuCX1Y(:4͸\ۺֲ`y[ڡX95Ӷ&)L2YR&)M<(.) hR3|24n=ʮ2o>zOztXyү+zUsC u5Zx~n* g!KY.O.pZP"Ef>JTPY >&Yz @vfj. ?"n[dvPfnܯy7Ag>Жg'c%µ=hV^$;Wʂ/Ԇ 3콒k*f>dÅTTo{r ' PG zpJydBʡ}јXEg ᪌К yXY*tޙXa@I(ǭQ+% D\Ѡ" ƦYUBLZJ%c,BY5y~nsFSd P@`x;!y4փU,'D_ -;*-!=*( XZ"E$#w)O@ױ5Pe= E-`<N-QNz%Q}"by{PQ1BBJׂhՏjɗN18^=8:Q*zb.yij0%gir7E }\ݞ'fn/~ݎM=bM'5yez!]ۀh0jJ;a3, q DԿDW:nݡWj?>c uL/ HBP҄vGGqU:^yW2bh˗K axЕ#wsU2QR p | BzŒ$oVVLd1|UzHc)9D$J_Ek{̄cyjAf%"ΐfVnCKe:J뙳= 敧:Xrov=6B6%σmMR;d5ngjrDxA,Z&<(n٣ ٨I\FyǍZ yFmp52<4K6@񐧖8$hմLDwNXLʠc j:V]\D 0\C̀ۙ~֥kysRj\ESJ <ʚ&KmfưsT,53o[KrudT^XKQgJx@TRɴ!'^E&@'~0N3 7F|T:-B9i!A)N~'7qs;P'o-!'t>{r.ݼ"<5BPtaCn!%TЎ/'Jd::>jJR("/CD( ȋZKIS ~^JEq_vR=| ߂XSҒtURV0ײo s-ya +lbPS%`%Pud#,0Ed  (lbBVv%ƖԘ]Q@Y4k."oV0]j LK*UhҧTy翅<5m"'f&Ybՠՙвj@k$׀n\F3'k@H53D (sw*f,sI:,)`| NMa:(ΟA,4hh:jNr!g US͌=J1zSͨc%zQ @FMv\A.ƌT:u{K֝9KyPmFU l6ejtROi hN)uu3iyP:ߨn^gZfR6*ڢSB?otiFʃiFu1FS`;ukǍ.nCh;:U2b.[%o8R q\aFX?~4~;onXC:wΈ Niv"6?*v$3/T=p/çuu%싄22Hz?2NCI$4d%AE"dH@g|\dH! t6\cMBbNgAx $n<f3 ɿZ}J ~@2Y/̗0ff/R`wA#ǵjYFO4ń(I6/TݪgC9# 0hJ&JpAz!KX0w^N/H4(ǒ!ҙ 8BNqB HMmd*bzcdd2D@7~pX2T'yҋUčZDL"!H)#SXcxE#d6:H*Tǧ9ދUs,K~⚴I]5*,"؍N#Xꌕwq21#Zp$;t1Z `!X8%'k2nl^X3؍QR؍j9n e :ARQt)%0S:8KP7юKwTF l1XЍD q5h*Ÿ,o,^Ac+`)0&CocF*FG6SB7X%7ܘ)xX|4 Qx $ J|%Ӄln-ka5 'N `M k5̑Aqe'j\Ó!8H 6&'a2Bp(ѭj>׈8GBp_1#xa_!bxIݠQ#xhLDNyX C_W3 J ͢ovS[1L>1E{6qZ2f(9y[njnNcm+pVQ4O_6lmHBx pfeX<^K9b}ْu3[͖v]7Vuնj !?).Roiq~ m, jDg;hy_O>vVZHO.cdJ\ǹLm, jDg;hs"Ȋn+[ EtC*SqIO%]z>Ʊf=mT"ofTކm/VL"`"&b6-,H4BHr9K҆L5mTd@Hb: pu`^0H+D.|xooey%Kw+7wwFܸōGI({'Mإ5yz=eaq{u*na͟Oㄇ&P$eG=BH}bl]읛% e ?'SAb-3s kQ,SDXvQnS+C}2̌yn>Ysn??H_H/-5?sܺ7sҵK}^,޳ْ_sɴFSgIէU~;.v:qOe9_`d1_$7~%S_{n$W9vk1qHY`o[{2˅]>v5Ob\Go} &ܯ^z±xە߇͇4pv{"/T,܏9eU\x;U]C~n>TT=\[CI 0merN܃NLQ:Sǟ<܎3Jv=ڇS6ꢶѷ=7OBm/O{9w3'!`NJwY{K$Y/=ܛrıtٞw|;>zglb:^XʶK9)~L'iFixJў48q\lO4 #e؉[(%G IeuPdS+ ,r%"ƔTY$ M1hGdu,{ϰ~Lڼw1x ^ˣSVU}bE{q{ OwF^TtOF6'-n4:_~qbkVc}8m&wH”Ifp;$Z(1-]JExdi>^ r=1EƟZ-N%gY3TQbC 5q}]x6)~jɣKfo;Oԙ[@ʒ!<69)UxpYzr>_q 7WZ<#4EotO[t$iu;\f<^ʉskv~)vmvm__\VAyͳNY,pQB B)EFoFJ.g|FIؗ-dݥ)zy^cثa$Zu6)kNFT[tJ:LqP d!ժLEbQ(]^sY1i9nguw-L8yۙ#'m&eXPNxN3:oSn}Z @'GX7uoM`Q% #_>+Cbiw7n+~٣KO[uVm& =)8}rJaYa+o00KFޢ5+ۢC1Xb xq}BBF,5+D*/vfF* ׸.cu=_X*݅w,r0W'O]]ɯ3\4V< >]~\}\^_.޼~-@ϊ;4o2'w|ߍيj>}ݥ*n3/clz ɻD_+N[ +vPIC9MHP*G@gR딋AsywZ|Yμ:kB:^1Jb\̕]cTj %mjT(|'ùj;S60*Ҷ2!FSi?"2PRWa2 e ɥ<]%F]])mNesbβY?eSn=dT=\eI0Ja{!irOǶI#RvO/\< 9*n'uÆ &y4)4Cpf2 HB6жBn(e7beD؍AގaWB:~r7׫T~ Ț_1%pmkƎ -KeKgQ8jیBw(RzFO^pm8A '1qoj~bd.@Xq1‡.2zhgm= brۮIiM$5:҆$Qdhg.p`5X$<`Xž+lSԋj8nȾ%ʶ8B&0-)nwD 8;nڍ?}}ܮ{S3Y)sƫ /c&ZWvOQBjI/=O?vճKp!.')Үk <'Up1I$E"\b Ԏ@iAt1>n+ ^Il*yW%'RJdTQ!@*.^JLoZіR&¤( LAhP;#tIM'KaA[,#Iqg1ŠRrSk&Ys6\~PhrAϊ@^s^tmߍu=;]6)iH`@go"E[M ĆY^'>ZދAG x%xt fH6&r*ՄHP2ی!5fGwz5q^3/YB"P;AeJ=@1> 'Ik@Ze{vknbώ]"]ʓ!r"W++@aJ'buz:ިڒ5Jp8Q)0S_w:-'mKir꒿!ܵVOjr $NJZSR :ypor2 ZV(;;ףGh`-Ђߛv}ﷂg+حزl`s]s 0b 㸂O;xP&{m?j@Apkz;w05˨sA#iMm<7BH }HJ"2+i簸 R;+L70]aw|₋j8dT^\|6"{ʔ wUV5 ?bU^8Ncg{QX$h^qy_%=3z_C@.փv {w˙'R74d)g [ب~/f(K@X<2BR/PNVvT ` ms7f$WuVV(y# 7wY(bq$%gh,ndr Y?Se\>=ym[{-> CLV9ٽ.|ĺbֱ=59 Wj WgzFWy9go*1ej2ğx_B َD>Qa3,/]PhFf}WFڐkx;CX{liVۆgveÝ4K6آ1ds r\읊OI̐{E(^,_+N7a:hzl ۿZƗRLVWף`}V+mbeHvrhENARdrzEDzq.w"@4jĚ 2L)a}'oXobg.Ȼlk2jGF6] t-=B7L.39BIr09)e[?AC{):Axd}pȓȦİF˜ HGd34PSJ 6D Ijn߰*HX;X)\I?P1ֻ8#)#︽7w.RIL$ ,q2iϲ)Gʽ8 r6Yqxԥd5b%j1i'NSAN+='YNp[?zXY|ĕC4;17q 2: )#CR"HOdMNb:cx, ^*CMfF&nn 8d3-|xЏi{Vc,7qo~]" K"ex91zM+_mௐ$I=xDf+o-\\_زmcѵb(Fbwg}6!bIhQ$Bh)Bb!?fQ;roVfj/>_rC 5܆SdZnevGڱFquid\^5 ~m1Vb?v`rlڿ1Bݾ,)I+RwLVqqmolB;-tno濭jy=ګ ջJ8Ê\5+Ie=ݷoAZ챑G'K@Tڞl3,7H$O7a<G1l}ߖ7}uÖYF1tIT nvnp.8kȲgt1q@ޓW7IC~)??o_N@8po7q?=QyHYRB2eBRbh=M"Ʀz8Q (zKL F7"TfSjʽTN7j^IEJӐ` {_kv^\{@$qk>QxN5[(#9| LʑHLm.:>xh^ ~@0/6Ԯ*$1'L[5`dըg@8l(G4riq93nG~.˞<,xv\ݎ6 o'm66[Zaݒ1W]cN6DuΘ4Dd=NXtnj2>e''` \g]]\Գ4X=*s3 +@ z}@ӽ,u;(58Ib̋I] Ӛfpx9|^&⿐{o5]ASBtU1⶯0' 咛:vQ+|@; .S$$q-T9 ?8yنy5 ݀Y ^!wM=0>zMs@`(1/yw+V!YI DzDo 㶠>;%"uj%mQ- yIʛdtۑvIgVD/g'zdIZX|R\T-nLنu<* |vqG2Qu@YigQ VBiI~w6?_] Mvy[qѶZY]}!\9׏2Ϸ]U:|W]uyB\_~-/?E Ыd!E(t"PMdJ6?VŶ\~L`S_b 4ru.ț!BevCJw y/;ؼ6M,j/#Zcނ Fe .gd#B(Zh'E }PRRNnS*$wc9pmXieIuAjqʝõlZ-i+Wh 6\{<!><""#چk߈O`[BV:mBv ׾ӗ {|A<چF|B:k-Q&nn;g;:wٕ_" J aR(ld %oCM` Ox.AH3}Jae?09>yJ]cvNnGdym/޶PG0KSHc^'YSJ.kTGQ3KD}R#{z>g( U Xnp']^?|KWN=by|UxB$񆡹x훖ƶ˭}[2f33흨ۼ=r"N.ING?,!]N9>9bF>.iyV~}Th1{INP#sh% H5 QST[Gm,ыN\ƛO,Fi{Bj,OUTvΒŰW#KݱYF#f]r<~踝9~Q>s3`F^ h]dK 8;{ A;{LgZdg@X6Յ0))=J v /~TNg+רt&>ߤpmo[oFʬ RZiq%Ij g[:BN$=r(wov?S>%&rR#1%,Z@;b4e*Ҡ4A,J9ȯ< / v%$WSkle {dpp.۰[@z-|BS/T{DPC"wR*Y|% p"E<PkX66E 1gS$g%W؆k!YZٟ?vfLR3(9.掋)E)Z cBA9'Z#H\JY;p0٣[y2 Qb+BB|5"]cX<9gku,49FjplX"uP (DvuG j ٔX vΧ:%wҹ~ˡtz3cn#H *teNTBp c7C aȉu'<>[[)4i"jܾ ';ǗUcuKh[CRacQACĚ +S ULQ+ù7sHAN6Q gOzp-e2K W"sBak<ݛAA1H^1BhA[ cSZ_8erf狠$+u Nb3RC [B[&+iN#R!KH"0‘–TF  :6ʐ T2ųjW%oegz%{{6ZڐLA{}m$K֕ e,?u*Շ{tcu=rר~w@̝{ފUWG[xT;umhp{+ΗvSZNM|#c,x-G{-?Q#_0QrY6V2;a9,#/<2ZXP8 -GKhr+ΖYp$s*.s2 ''>-Xm ‹z`8=KA!XNRd*ݒLPNffR']˖W.j{=IƜx޻LuHG꙼ L#Rr&͎^bx>s;_'!,d=+P_]SFSc9:5[eQ\;8-0mq1K%ې{Vl^zE`#u]K36s > svf(ڰ=-2CyᇽWxS]iyӃ=G6,#zr䈨A?E}(]4 wBqx0T\ u Dz#g=$0džh;.T'a:{ rrx,c-[ 3G9[("ek0`eŦ?_O.|~S v57z{/Uq?My7mq7M#a5L^c6zOZM J22IIQ:JRJq2bJȠtHp2kY6\]"B(ΐy2j-*׮5!;d6sCЀCZ442J'͍1V!^+lp{[IU3IQ|Z hn$IHrMv(; hcއJ²%ubl5F;;kT; / ~o%Lmɘ㧟&Na`}*_aPEt =MkqKOS$ 4 )4 )i}ތh-'A=3h~jA#{$#ꗏ8~q#EzLa\ Dvf7߷8#[=#DL= XIXU,֯8qKs>1TjP(ZV!R++<`:k"r^}o$z佞5"*["hWfcF[79P78 6\)z7L<3s(4#vh5&^f 0oy'A& m.Fę}'t̗a }95ޤ\2n\4Ar"rf%IeY`W|Tbm6U?U$־|1_W҆=kc2U(P',+תB@De 9!D`f;/;h?#{ڕSL8¡e 6aj7b^l~\oåS/dL4BJH {$XG?}< ٮѮTΙ%iZ^j]{[[o!;tY:}M>a{N8)0a}6)! ɯ`;uQ \7Ɓ=tROKCə ulF蕨&PY"_ 8tu(qG뽘z`J BWr2Ad,cpE ppE ġCsF}9k!wNoµdR)Fs1'lYmV̽f ⇎oYtSn "7P2O\ /p;MK:HA{zqT /:"࢚Qt/e%tq3|"^I)«hp0%HatԋLqپn'xʇjofYlZ3Cۛ؄tNkt}4g}]wz Cӡ \v!9 (ۢF#l%B/ W?d~wo}sq5Cor} $m.e|ӎ`63nj=-m%%IouʬR諓*a;#Ij24;s#(;ɢ#Q:^OdT}&wdVp'w(ܱqN @x2gVP:|2C[b䎬Tg笓ZZc&wڦFOi_fOsc*w#קauJ X=9r ӜoINܚ"MJLM9d Kufqns@_&ߩU[(%)UY"u5JKf?ݵ`^Kf͏s%BSt%h'wjTڀZLl@zžy>Jc '8U#y6ct5i-M{z^jgW D['%PcLV?L#wN8uvXG0c7b?DᗊN49enI)E.:+RL{a*'cTI_3C.=-W>5tƠ`X)VJY>P(Gݼ4(Hcl@V2FziF.$![+xDFjF:OSڭQixۻ}eXY *m5.r9]~sOm]?.y=gK$9U% SXj|i](85D3*!wsn sԃ+./1HzU=O؃:4qFPDޤ/S}G7rp~JJNmRD.YߘE^TA'^`)DS75C-x Z6AQ:DcdNEŁп+yó:o:!z5yk`'/yeaKJ^I`$Wo5Ētz4~zH:,HС`8UD!Zge*yxVu69?ܥ'bFM*eSQW6|cǶ*jC ù l0N;țvK^-M1=܇s`XByN~[þ Zr6tjB>yхɭW@-yg۳Z>vvI޳WyI4cU{ߺkE9%*J*JqJ|C;.>JԿcc4 fZ hȨHP?+]kR7ֈ'_@6-.gpm6 ,mPNWbM{g5%q)صEdbEZZu&P@_E{b(1BtӶ^pM:vdsQkgUg%%ȶMׂ2M PRʂ_ -P:ٖ"$!|\T8ݷJA!nbK.'2,<ړhϺ=e7KT_RMB(sf8nUQb;U<F:CS{:UXUră_Ay}QVF9 8mѡ=|.S2; bb>V$bE F"́BE(ٝݻF]m'h=k~{*I 7H?%o0PWQlZCO['~HoH>mJo@<3Ǜ6c/6'In۬]i+ּ ^=fLW6{I*mxJW|w4 -<NRm%?| ]"KCp2!Vϡ,M6蒣{~tb+UC}z#i ʼnV0A^CErZ7jNq`wPmB 2$RtȞTTS>jA v.F͔@%Zop h#YH;-"% qp0@,kK2%a_[ ew#"BKT_S-VP~ơܽE\7ZL> XY[i;?})eXeVZ5+=k+-kmR1V Pf[R2+\_ZEXKT_S%⟻B+dz:_%@3{Uо8J^ذB_EZIsCr2Fv$UէFl,(oQ0h/Udu }F Jmd*PJ54psnͷBAzsKSV[<7*Gg5v?<ڷkh +ǼѧשE[0 ,R,d:C mLrjkIs![=G|ݕ e}?M)T_&ߗ|3u}[xG|ǧ/R?Aʗ:Zj^jA~ۋm^h9/ }3īBF\|G@> jFkLi#Wq(p)CS1M Ly<3\5ϼ KlpԼ6FrQ?sg6lC~uEZ }%^ O#Q NK+F1qaU B`pk% xpGVQN%;?|6~fT\1!mfHϡ*k QgɽHL'\t^F(Md  *@.tsݏTԖ% BX+E&9m!-Cл#j"+иb)zk c "n)D\;,x)RBqF\:)Q+q/dC}&?KZy2XA!SS)R΅e3^O[PV= T|/h``-+p"[fGЈcXal}]cɠY57J8|}D'5ϼ>DkmּO$lG}GZ"N6l0B:Z%ZgNUJ|4]E/k|f)!,CNUbG OTYiiNFYٖ)I?Z-J=.GX>B&Yf,"ӂk- @:!d#AAYK{)=ݺ\=W9Hr}3^a%V.^օUaõ}Kv(eg};,];)GsfO8W.[#,\sxTbLzM?}~Xo;#|0TZ*ma>B,F*$ƒT@6MV*$ysaO:sS@~\q1bT rF7e"RZ~ %,)ҩ\BPgp=8LLQNIJZKVN~9>s:7WuZֹ©ULʞ|lABH&X1TGʂ,UMr𠃑NV禖K`<ˇY&닫?>[Oίf>puC//gszezo./ΆJ{7Z"f1m?ڏwC lv|Psv5?6 scK>kysM.S`w?t)?C]gs;wqNy;5+lUJd"7xJ)Fvp #/.j:~KO.G-on/k5贪efL:׌I;]6`x;im;T/t5paۨxHBhFݥrx\78#$iH.Rh ҈r9}8Q*s(CQ0Y&@Upp cc_.4[x[ +r.$$fbTF |cc͝4Yθh0 ׀;fkFCx08I$46%8Gؽ5#!QylecQQ5 0*)׌0֌ny<{Jy liƢkZPR[)_bkƭ>l2'k(xe7 _3Knǎ{Ii)364gIY^)5vwP̖6.8GI[ [_Hu5#<#0u 35HE|3m:݄:v-l<QZ>:~ZxjRmgh&G=TSMEY6bX!v˦E]Ej]r!2>o:̴.%n/MJĭ꺴{?p[eoSs]!8GvI~kG/cCEz^\N8Z'b5G  /KOquzkC +KC+/i)>!ʘ6o ,|~YrӞA9iءJa7}s63wno˰Tod1nj*dLlT 9̮Xq{VoɊL bK&=.voN`qV3sOTqLyeF0eO,DUncc6۵-#r˼aB'A`ʘfZ-z SFk 3}uj_&ʗN^w򻛇}-)l&v"tvq?6>)e?wœO~W<]ۓW#&S ("2SQe C)`*$A&NWGe|4IwMSޔ^x9wNgy-hζhoAeS7ivPUNN'$E_]2u;lq5X ܀fߝx3((|9CRC7U萶lVT͞;@ʵ3V#zRB(.JVڜ@9&9c Ҫ)D|{LǻM!YE/ѫ^ӓxF3 Pofȕ_V뗟ȅ8ߤŏu\͉`^6/ |eo/.(=;%Z@{_lAF{2pVԓ+ANi G񩢺Z BR?^^XE+п̣D9#RB,ĠK#Z}RN)jV.QjK+>jٵb;9N'8_JE?M&Mn׷7'Os*(ӲYr99Dʘ O,J/)C}?% ob5WW7 owe ۜ,Zכ/ =3OSQ_D9مilQ}JՕ<ζ*M;vUA΀WmGթT)(KloxωQncyW}hS>nC|zG5!_I Unnvvxz55m2yYWs> cY@O0k&gMxv]<"l;;mXϙraS-jYh< bd xRBr^++1J?LI2m!;e'E4ȹ^ N-}Ԝr%G!|msKA+l42Jt bAZ`ւHKstP"e ]"%]Z<ӴydbPsBkdQrt,Xs#VuPknZM<SZQGT$:ҵqJ!p;T#xOA'ђ&V|A2RV>qa]* Ua(IQ(584[6s.a=TUGtLV&1LNQ*x"Qu9~bv{HqCM]((;ee=c kYV@%ҖWU ONI觀/![.ً㨂3KJ=I=rgknT{挑ZʗVQxMu(nf=1H1Pj&莘,0?Gx5 va% *34pV +1Z,3/&GZ)QKkzc)ډ 0KΧv _҃ 1{bdbd34QuDګJ q!*Y*#4SQ L$f}}j 4B$:hb$oɐ1iI4܋KYElTh{zF!u+:t6 wK; +x# 1zpԺ~3.k2z3y{o_F],) >Ib' w.9=p2RIJo˽e⺐?n֡[?:*`1 hW~DM"WҕWh_/n>6?<۾֞D*Nt UQr) @3|{~\\\Ϋͧ/ .bpyg1u ǎ.zXn/*EQhm0,p}rpm-Z)V-!b0`W#֩t$ z6LzZ'#DNXf?.YBK0񨹤EL]@]Bk +hZݶ[[<ѿm裘6G)@V 'F-iOWAHʌU4tZ_ P_j񘸅jAoHHjB15B'h47jryxZAr"[_E(rX-Cԗy>TmcN/)ξ pegnx{~ϖo`30<[^Α?_]Ql}~߶ b~bF;op.JJTꪮ/Y4dg c(YAL t, w!C1~vZSV%mm5J)Yz{.5MayP )'j]OulhID/_}݇xw/WzNl1[_=/܇rڠ㞣~sq]fdžϷlG0sF)~{9 $уO旎#fxG=A 2b1(r)]mhH)zzX*t) gFrօ&ݺ~zt-&օjܜ GuZ%`PZ,][OFi-&iBPeoFσ rDž3z=@tȔ%4f*31qK֞ɼLߣK^+w"EC H&~,MR< eh3aІi2d<# 81ۈ;SҢl TS"h! &H4DC~Wc 0t+51klfAH#"h_lD}D)f_ qiPQ}}L@2 AccxTPH1b8H6;\*Itͽ+wvBopd݉sSxT; T-nl!ٰy{_׫oZkg{}{Y;CE}^[ B}߯|k~{L7[6||:Z-4) uaVo(rK#{HL 4?_[ڔB"Z\פd!/Dluqiw3л tr&LQ YC{|HևpͲEҩfލ΂'n21ox wiօ-4LwB^T$21JϨSqNżu*\@^W_[g׋w #=:kl{ QTfbP¾JT r*?yUݕJab_-Q5Zjhr[qe6ƌv#뽺dݴFO2 7: 'ReT^p`J30klH҉4?tBHxfJEaMzHhHg ~I38r<9 ,% ¯J fC9JAV2qBs5Ȁ$%Re2 ('E^c$h; b6\hg$$Ў&!i"R>$e$C;(j؃dtK ~4EB1:㘳;>c}f<#@pFFC\==>G䊲S:E8='(ftb<̢4 TG K iGc] 3,ED Oa%O..k* *M8)ӠAta]B>3>Y2_R#*j (mTY×JofTs흑d4,E lISsغF$K!e鋰-"l Jj;|w%㒎@9foyߢ2|J.3B@Ăo JL!RboE !('fȀEy:x)ECRZha|wz{rz_Sm8tl#bسTSpժ=!D bs91]@G9FbNa!/D3lJTߞwӠ)x\ĘN7Rt TP6w˟nY6gލ%N"w˕A~#ŻM`[!C{)ёޭ y&cSxa, YxaqWż^l)c ac4T%/b$h-p`RT7=vg\*@b{vQӟC[A; yd<'%h{"=0Z%]q 9"GOzy(f)w8Ab'U .;.RmG;A[Q+;Jӣk0=UrhTJ?NceVrŁ*Sg 0y*f7Q8FMW5 R )J`ʧj{w+AN|;}W'<耾wd(81WyTpd埍8ѝPڑl#/9V'kXkuˣL95SQs.34u<[0("vkM4˦>ȗ݀&n21oxuþNx-(,һa!/Dsl8 y߻,ϔATUn7 ǧN[HMgc1`!/D3lJ3ib' ==?T #<F }Jfp-ܲ;si?U?P%ƫ{=UMbߜuŅBu]`X wv}]=\n\\Pfo2b]R=//Vlf]9\8 Yo5J6CȊF/݋w=džmpopRM4qVM?0#C0 WyCOny{Z{e҅.ߴS? +C[I)a|߇kӺN{*õu餧QɊy u-,.rAl}7r7WM~q e%p]u3D lتt*BnӤ4dsƹmߝH|%Q UAi MZ3Wy$^h@0՗Wk)t^|f0C3=P=%QݾbN9u1 E9eVIK3ZXNj+O)@dfZiپMKbep[=oJAւ)bJY<lxCqs38BĶqpw/.ٯtc=Fe }arxk#jonPr`RQ^%Y >u (|XgϘ{Fs"L@dHB%}:.huN.,h9R0q@-*ϩC|?jd ׶fWL Ezf% xpFߝ1m;K}z%Hm]wM: M3E/{ŧzXe#ZT釣;#(udnd%wЄGZ=ƿ ;Eoպk硦ecϏٻ޸%WOw}tWW}X`?cNJHso~65h&8aSdUs:]T ㊆RՂl[ыߐl㘯'cB=`?kyѻr r_%\or$#^g@'%qm̅΄6(L0m#|1xhf¾ܞVpzU>x|.<[o .r3õ#`c&ԃ[e"]XdaGL)h*4@Mړm]jEU+aK@U"nJKWAX=Apux ;}A>)Χ|`!F԰_cZSy%ӃuLfGӨ|9o_R1MCΓsv9]?ZWxH6#p"o&DGga<|jq?l'bEduVW)Ah%As#f^tŒ p;(f(]qi z79Ҷ('u1E_,# Bw,cLF'lv(& 8nJTb ~%t31o8)=c#/`11 f) Xd۪h] VǣK6jZCǖ+4b=I۵P:AWmn ش,S?>r%|h_R|{f/n_ﻣy)*祩=nynpLC`K %r竾^WD.*&vrW*UUORWi3_j79g{D->k+_,~.~?3Xsה}Vߊ)VJoYӏi͸EcZ3جތLY4D݀5oV: ?qq#Vf<< س?8Z4Jf -΢šqaM8=8,s@"#4(g-S8Cױ5JXv-B(7".fUL$r앹'a"l&j- P|VPw%i+wPi߭kgWMVٹؓG 9(4^o}sQmp+RNWϪLhl7Ћ>nڇ݇54Th?ː2~P[LL~uf&CE a"$˦3˥Q:njڂLF WSLNL6Ǣ#כD,"H$ ʫNNFUqF_O˟*>oe8~&mF846}fKE\ympBĸ ްV&YSD&uBa&kJ;>6-w/P*R`njtk UGMJ*[[;s]Wq?Ƃ!6HT,B#r2Em 0M׀:672DNvRԄT޺Q+[ڮ7 ,“"}J%w@ O^ e?w|Ϊ˼ #6֜>>5=ة~@ײ {t$ '_*hFHN=GImTe CrmX R6lk-2ux0JVvNPSd9FC' ge9B@F+Ks6KXݡ C2մ 3a\X Oܞ"} 2%DIW,g/_rWHk/b̠k^l%+zF=<@02$]c|V?8zȤ]LWz,1se 3ӪryU"Rpr 9qiD( j W]q7Fl~;W}Y8fd4z3{- _˛MKq>OOR宾2XotVcM@PӍ(n+IYqC# =XiY!u>;zΒBM 8*f %ʬMrŠSہ5no?~ϟ&ghLo_O241CCD^` 3~E#yJQT(Q:~~s$!DRM7m) 8ß1;ՀTEU{,Pmm ;ް -ƻw7w~w1oxG:wVQ9W^hܲF Jq5F%ISBTV;vq2kPQh:Xbl^X̤yUk|i1xϞ?EO/i3rUo\I8Aܵf]yefuʋkO/} ,=HogV \c99k2g 7VŽg2`欧+~y=Er2~)>GGݓ #[h;&libY^s=%l4HDe|ޅ1eT:qP ?MVFai/3 rR *mZ j6yXps_]j&j28fՅ…V^T@ 3pj181FY! X5fQH8$;Yl:<~+!J((-f3ފjM. hn<@JK l+QmܖY [++n]Tg,0 >mQ#ǖg38`,x'%aduvzh`ZhUD1m9FF۪je@UrŮSv'IR!'_4Z6-Z3,E4uh[I߿j6I;kY-v&]lJk #VT0o5>q@>N A9e54`elCbGTfX%8&`1qmƄ[tvUn cϖ'Zc4ҕ Mۮ/`B[3=R+#W|?8zOAPլ/[^x Np+ssaI>ipo)6-J@1l.IOZ%Mdډ6Kw>H s'{79bry񎡠e`޽?wmke>gOOUʉ=4 )`0Q>ޕ;Hxd77GN&C` .ON2QcSlfdHMWi )F?rA*%Q۾56vZ狲3 ;)C)'$|= Eg&EUrD<<@}SGw h J8k\p H6 w[̌%˳ Y~um08Cxgς`ܕۂ"L 2 :DPe QS3diuIxBb|"#8i1Up(BckԒwԴR6BUWvnHҞHe;OTWށ5i2ĭZA4lWR酜o|SWC]CPUSc5u6RBF 1*U:?gL ixeoWVwsVA噣]NWL˷[ ľXn&p$cVCk[c\/޽|mB p,iQƸu#%'9)Z8uGsO/i3rqU1f-\$t5;r8xIjb[oѵD"g+-8W@̱3J$1i44ooR" ?( ][@Sjy޻o'{797n$&=XIʼ_2Xg5&DT&%1qfm\ٹ N|D{A^3nWL[4l44ľD0ü؍ԁ3HS)P MrRQ7ĘrkI=tOg6UX/K. SUo AbJ~jxGEo72/ĭYdY Ou0#XQn/d{Wwm6<^IՑP7J$r=6=C"3u*08H`,֗wY Rƺq֍Ĵk-][o8r+ $ d/E#C6Nf(Qc;NfC۶fK>=ŹXG*W,JiaeJJ-(u #g[WG R0U+a )JxYn *CʢR+b k@Swzwu>'t '"PoON! 0鋽R#1qJX!v#n 8<_ϪligY;AϾG$ykCUledt?y9&<*z?-g>Y,.9)o>ܼv(&Ow>z7pzʈ빖W/ldHAwy9zxa\urS`g hNT47od AL(hXwaP[7ovy%#BlhֱH,0`!'O7$%+hMs2 >"<7ᖬN!*Z:Chyc]RE*UM9*`)t)N( P@J-.~iM%MIGԫT#&.?FKF)!;>kjp%FW/^WᨇcbtR:FQÁAF0I9&`(. qH@=^ ei*ÆHOQ§ j\]Dj Et#ڽ)G N1xRzSh=VSTWt Svsؖ -o {)ZRDڠE2tkWUb J2 @tXI5A8$ffLI{gAÓ%h@*ݽW y IxEJd"d&Jg͠BK*'( Jw/ݳGI>EfQ'h CGYY'BE1<gH?@|& =HM?V~+0 0_]8V8蛩㛏1F11b" Ti4Gl, "siu?к/zD c0Hyi a%Mb1my^\(eڕ +ڼ*y`Mn*dXT LTK!>yVUb3Nr{]$rBRaI?RK6jԆ5IPK]8 (|,G/?XyK烟Af*rç\s-$7,t=flݿ?Ɩ$Ugp8; O*cޝ/5^S$Lx.FJ]5s̊ǹ_77H% ϝ(-F2eBp70xи`eG/ MWo~@X\?l4&Yf.aTi;=m661t P~ 0>w'lF(l ~X 4qrL6 $jKX]Im7a >叙Wye,$~:}p_7^)\ a{+Uts^j"Yn*:UL:*a<W j>CʵKȓf&]i}fcAXj q,]=KX<WXc//nXd!Ǿw:8yۛE_ Nߨ[o VV}IGs6-Y@Oƻ2 I`#-!,Q9= RJ?cԦђn}1cBáo4tw9Z8:ѥisW3]zՒ!{$hzwK [ݒLP*[a$>W}icp7OCûjN1㼴'o})@.T$V,޷RF:O/k٥ACxgηCTKTF7$Mw?6:ɀw7#~ks͇U;$i M?RFQI[ 5f%KTl5dw~}tsfh<$(LHSdW:eTU9p-Uqs^ɸDAu xPJ_u7@9Bo.'Zo5` :~ʷdJciHﴦT}V9-` 2D'hAc@)3bH=bQC@(H!:"4z^{ݴ&;LBҶU ΁1^8)TI9REKomV|m?~i.l7b|[ Ql[}n]ΊdI0 y|-nܕ71SBR$Uѫ>2;2^ p3^`>,~JRm ed2?sr#QY/N]5Bz zt{Se.cXxľ 5~Kўo|lzijZ~ak:O7 iyHC@NH$!+]VXTjÞӦ'4"ѐ\m2&KZM7 Fa,"`T[SP#r}68N^ޕxu &q]!x̖`drʴRVT;^'VZsCqWc`oOMBItb'rwT-Fbr64'{W-y3Pܖ~%*\o@LU8IWE9/>Х֡ԡ˲˗~_%Ge]DRk@ hWRƩyn`uSh LK)s ` @6.]qܒ*aVKYu]S.!Z`v'L 8 yb;üKdS.AT [J+cEFFz0d2)}4zwNt҂ 6qJTNo}j$@ˏ]Z΋BdB T9=Nk*OpmlqGn~BiUAЌ/z@h*z@Z8vؗSߓ7Ef&wBlWph+fXv p,ER(%͇"}v?Q0n)(ik<]WŸW U ˤY%4{\WB}n˼O?+,}1yzO ɘ\n>χcE-&K$~맇SHRc©awtŠ (5ywAp*G!6$zb#ž(QKA:^ɲR'qGbp8քAy"V/?Gy-ѯDלw}!LkƪJa<:#(C}0 [H~hnp&Ȭj=I^;;fWC }:FnKRGؒ) ǣ0Iۼ*Rs½L¿LqX]hk9d]4zb$M$7aՒc= ]<vhmG` JFCDǤ >8M}hԣpk`gQ0T8nP-O XH6ٍ~%8c992FNF/؉P &9w$)2 :q8b8߀UHNoT8E:PJE^9ˤNm]PJUSJFZg!',$ +D.Ta+nDܢdJ.Pk`d E)*QmJ#!wwR)R{X`PRY\^'M*4 ,C_.i"FxB'Qt!UHȭ)rD-Wz)01v T#NJSf e41 RҔK :~[{%U?C$~ #ILW i+w\ F{M8Z]-nSh{LƄp,(UQأ*;:0~D 1&>.bZ~F={M|0Y;>8~b#D`4n+SpaRc©r>]c+D At^dTTo;-@&m"Z)גtɮ\4j=)o9Vkw&/a\EBO I_4]'&O knrVF?wHr]!oà;{8`zg>ݡ7NvGN";MY"$#KzJdX:W=V x:<TH*-%A҇rTptwv*.Q hĖPb/zy M Io?"qQt y\.8QpťS@']q/L44`r*'s:I3B&*Re =izIԙ:Nq0s=Q3$B;Pq}u#)4R'0ws.r!grsZ[<ǟNpnsv?q<rk,z^x9М<0d!v2{ W%}M,jNH;%3y]c1uvOEiީqhpx]==,(#n +'XXJ̸sh;o,%s8&@zl ,hǹ""#[a UcVf`(sLdcep0/Ә\y&N JcNCBόB'pp}9ƌ7=Y{'.@_n8CȠ9u d Gd8hF" ={1'N:N)Rڊ`YM/~1ҬW4e!U.b6KO_.)+> 1ۿ/& BX<*zksF2:`ˍ3FfvGr6pu:K%+H:}W,UvD&1v9Қf?Wgܺ|A4sa -,2[Tܭ啋Lj} Q|rE{ΖoLݸGUNw ܷnJWn=Gf H}wYx:y۰{:csZ~1[H:@^ο7;G]N;կ*)zB" Iy2<rf RK44I272OQiXPCQVqθwDi1 gXHst3I3``Q̰U%D *R=ɠa¢?6\VPH)4UR** !eyɂv"nASϷ{S%HSԪ%NOy@bo|:}< wYR?obf߳y=[:q|tX}?7oͪ@u4Bi¤{yl,?0_ c: ww{u IYOMM=X藳% 8uՁt}Fm8T`[灺n}X;7N6,z&av6ېj guwn۔/ nbPYӧIJ빿]}~.#F /ʣPw˫dy5WG_H׎e/E1F*kcөfy]d ǥޗޓVT:Z,y@c/GNev޹ᨲ 7RwkՖrԪjz+O2pJJ4-s˷<1Ĩ:Y:)J͟d!`ZP1U?s8J{W#`0oVU7V*fSƨqV|Qj׈D7n!>+BN9w- i;m[|[/tqSqqUmTut% c/K#G҄k빽?fm,){)om.Zsi8ie%M8&4m3%8cDyT])l./WkI)hH4{3a9ÉE]7l7t}p@Q͢Bt]GAkxD#>(C 7!7g4~+>"KW\+d4&h(4b{;=&LptQO>X>_G vf2` Yk=IYO&٧"펻"MI讝ꨇ)+4_ 쁚BW &GO-x_qW2xQjJy&zU&S<}IWW2q(&iMpɒ~{¶*PN!==3\*Y=ѯ ;{[nڅ>>Ջ&|V"|zƒfw_oEmdSFj^򓝾k=7LZGIGeޘ禫Ͽ~1֐VٷMa!81$揲SsD'HjY{!9/!@܂ ( 98}Ғf3+N'%avgY&dϊHLhG*^zl\~ܤc\ףd[I~*w++Dm_jy%~}ڽ珅}\N^drV"-v[#:o%_/3~XdOpcrP".Vbs"L@oƾܚAJe]l51JA,v%F(M:qy3m7S$A4" ` EH<7ErekHRM OqQ8+`$ 1EN$:M OW .)%*'_L!eTEY ΋.zThh+ T@6Ai߮a$. RG$WS a>l4b/=0(r+j_3Eco`D4q<:6Fj`CJWҧ̏RGܫ?u W9uN6C:6l&Bq|7usupk+D̕^+E# A/(QE;ÁcΖ81]n$_^ y 2(`<2؛v^lL&9i$HwU%)&y/)2:'8>:Ɋ_yc5D 5 '5D:ұf5 (y߫yh'm-ޫmo^ikA|v3Dq֚cNFeTڵQh93tּ#ҁ["bv@VU L++YiV]f^Z6 (V܌ jk YQ!mC3 ɪVn͗gK rͨ[0Voxsy'WߧPΖa/OƩY0aءarvR V犭G0H$@)|׎jX\@WB{Sy->TzGXď# [M 4}g2\S? "L r] J~?$ 8\_˸;lؚs7֤鏇"om.J=t}SmX9 3=4$.@84N jT}C=o+NzJU{ $ C[@& Zg2Jr?=škh+;|ޅ$tOr^NbE{fl' ։<IO;J5Zi?RO`8mEO,wNVgj{U@";#JѬw% bu }U RY ,K iBJyI-%[ȐkRwwܹ~0s0YH8*%3R;je(˖RTsSQ5ٽg߱mێt;Y,}Qq Bv1*8Όҋlf̀# l!5 df?r}|:^KcF\˕q'AHY<7>2J%6Ht j HFB5p3|HVxiJ^{Rk9S65 &i7$dғRi^!)'᥯I}'ƙJ촽4V|.M4T4/]I-I쥧襐:PvCҔw4e@_ ۍ`0hXD>ut&{R[y^_ 7%<԰bivfM#9Ib Fk} )TM N$'h͸!4 S!iEr"T=" EN|sܔړd%)Τ7n1&닒#?~.0 1YRiE 5RBeP)Ԅ웱5Nj[&܋!%bNκ fU-Q5#sU ԫqŦԫՀ UӫRL3^pxӳ^-N+Vz=vdp6"j;bUs[_B; ,Sʪ8rljSR΀*@*LI[I$]6 kv@KJBn !rp;p&~5>/wdäy_L#$d&°5ϘGAoFj( 1/{Oiͫ n^F{ck@9\u p 9>Iǣ=H+j^EGy$s$ S/[eD N,APldG2:etuSR밈 :'4UiuҲR/s#e#m?Ԋzw#-HE>߷S~+}OlrARϤ")F_Rƒzdrj VXyA|etA|)L))} R$}Rv40!HAҒL|E!c{3:6sePL-D\+Աz2@_އaAԱAާ\A@d0u<ʼ^ !+n.=Q>cvD~M]P1wD>$Q4^noy\>kB/XB+o稙rd-+mx~8 fa޾*ھ$|K-K?5v+ %{j{loe 0AHxF=^!JBB!Z #gW)DqO{2oY-A BK6PŒQ3R=i4FhQ?9MWڜ~$(j5=5h1xܰa7-&9\GFL~z0mj4>Lz8#ZN:g!O7`Dl =D3,#ʝ{EծrLi**HξZ>;uvi!d;{K 9j<&6_./U*@C^'Ywgw_V?v@LJ]y!<¨*}dzϛO6ϛoM 4T)R_Ji1LR 4 K_bGj$ͧ\y)Ke֘^uUq, ͼV!-Ă$T8CBbtƠE*hWJ:Otc X)F:|&AL:)LFOiJJ 'CtpNI ɰH,Q1T Jp'] ,G |{E$ U@Uke{ ;aBv98k~jnp x˖cH8 ĺq+6^e[ 6~czyĢ<%V`*__1ȍ3,Ȳzd!k^Z%U876֬Ʀ|IT)c75PQ#T]G8c xxRL)vцO^FyJ$Rq°x͎*%A4WBbZR%9s'p s(5wU=]e;.j4U4sٿԯ .retC2*[gێ4QLTDUQ47=R֔M`mzD;f^צ6;;hcE"D 1U)(T@5ӥ0X9l 4=MW.Njm]?F EDS9UA`#di)a~krFL.]?)M7Iî6 Q-4!=L<# ⳵v|Cf{A#CNH.R}baʛr(X% uBs7LZ* "d%I9z9lÊs^edn::fݵ ~yC(x>(n?\"mW*YymaC$-mGI w7Kܿ@v>Zw\CzG"BXkM2,_}J0+őo< L+,"ŀǨq ^-ܛowߋ] }"  ˶۟c2{{CH@'󬖍 <%c<_5C츪c ɶǮ 1fDתxlcuubQ "D2Ez*5"=%y' )ns|bh*_V*f*dEiM)7 vFE4,ex_CO*9=Ba%@a"J֩D*eH&&rD pHTҐF{Ӌx r%S 0(}}Tʳs{\QTJ<1qI@.FpXV>.~xř;YRd{7!Ec`7Ld]U5)):,!j&Y0 =ߖf搗u^\Pjܽb~Aŝ\5O i`|}uxUq{sȼV+=`6tMP(I@p;m.civ YmOMhisSOV9hWR veS kպd|<4<2,op~{RӃO4yGqW#,y}{Z< m A:3h4N'.2^k;kuc<!k6Tᨧ\n~i ;ǷvRsAHnǭOK *9jRѻ f/.5yܶԜ %S{R~"!A\%"~5~Ӑiՙr#闹颽}gםbRv՝=}Wc毭mO织s@o4 8/nW!%zXNfV *o7@Xe {/m[nmH:|ρv5(ׇhyw6>L>Rna%ʦ0AF6^15*{vk_8u!Elʻ5`ƹ;QBk1s6H -ZH:;>Xa!4t(mNG)o8o 'aj-[J^JE'hE[9zt$d$on}{Z^ۯ.e{ua/v8iuٟGJ[3 >$T!qcVCV6JRƆļ(:@BS=uqU.xXM}ݥ"TKzUw3Ԫ`Bx_ZZ(WB;O|qV ^%AtCi9@ó$p7ODADɂh@ Y-ixn< 5<cd2 >0a ) .3t9j"{8?*vt?4uZQGD|Ih=/)cRU|R)Ykm#G/@>rs ٙ30XYIv-nɒn˖a&WU,VTA#$$ZƱdCwZ.xQc/,'bL1KFI*2)l*8M`*H?vw, WI{wBFICrL|#҈M2V$3 A2DC9Ib1Qa,6F0,sgd~Lv ̟:NFYZ.nMH斪]:պ%U0ɦ&5d^`'csE׆Ry~]і .d5qri0ZH]u`ݪ+|~Ň*J[3qmYB*mY3O!\1F%W2eEmH"XI16JLkdpB V"l[&B: !5jnһx[Wr=3WKUcjpt K^||`0UtG+o! iK 2 }](EP3 H_4RD@{KCFn:>k]([£Qlk A%[u|ym/%4B0n4b&jbTesd44N45i V9Gfݻ̃$EC_t3P_ulMٿ}9J_^+P(3:}4C>“b&t^ ]uPˁXзN3LNH 6$FqIkDx 3i7:L}.$C.<}eD%3"E R@0Xn1'Kt ypd1te o@k"̲٨8m]l,6RWLVkUR+Xofun0ʳa!Ji5`gR':umyQPn+~GbҚH~8v{wgqj_A ˯Sbػ%<&Z)BDX/; gxJ".@T}9yJQN$ a:W y 9GłVn/1_aYPi'm~XuZpi̳m36KT_Zp_~_ (:9;?0ZbM[#3A0ikQ9߭ot*znu%Ѣ@{]êka[G4jVQ.v kF5O&YZed&bݵTW7Ix/˛AP &GϝǢ%I [PWM4,4zw o2z5 JҐW,gͥsu,1_dզ^woZQMNk $I%^OowiѝVZknyߨCΜExBwgڍcv v̚q Ln5pșhOq{vs,:-^}QuZNy{[x:gYk%&V>'>2)Tj/A*lL!CΜECxvAuBQFuh8KZnu=[r,SB_-=BuQJuvpڞv xj3g]" RE%К?Χ?}F8I< q&Z:s%&o\\ӆdifhgv};e/Q[?„ϟXSIlg;sd=ٛ $m_j9[-]JԯWwU 1D7BuN^`i'F/3lp?hE4.Ȣ% cv)wW[$)h '$<܈HELkM|#o&,VZYdڀA^uKsC+&?o<[-~u]ƧxˠIStl8NuL^gV/0XQ:uVt%ׯ,dTfԽ =Ue/bΧc^\\}dQReGhp$5Ȉ)$gFGޱ5U=^wl*!adU i`S +\|v|C$hgp2zO4қ!5NDb&, H2SCab]3:f~p7#҇xґ b:>؍!TqY8&C*"tb=+$f*keiQf%X%unf}09 ! y ! !%2cBl.H=CJ)嚼Um_ SH4];x,ok\]hLD맷}lmw}篧QRrwBmUX䵋faC -XꥈȥǕɭOrsiO^ ]]0=Ng, v,i3LMk7γvq=ilCo5h3}hgݏ$4.Wt?8}NGy0V *(\'>:':Tl/S]qyGcOi)(PzEG!|zu؞^ݶR=q(r>:&6>n3ポm3h<~xEKRG5JN)жQzpi4F#/V{P~L<JD),&S2K46kILIf4NRYRqJ,(WfBW8uDWEĎ\(dGzXNs﵀}Ϭ8{+ؤB%s i%0 @KnJhyϡdGn:a"]Z͇ibY`is?Cyz& nOYw4E&UStq<1]yFH,.Lcc,djle`4_"itbMƙGvܝk]" :˼:[<1M8d#3=Şń11M ]Ô161vގji"k8@-Zsٯљ%C1C/uC^shcF+BWGq mta%Dם5//+so8k"}رm[粩^sH HqY'^,aҊ%]rhK*ڕ5uO;#5/""6G{%n}y8chѨ ګ*Ͱ 7Tbou 틧^ѣuIXRC[Llڋ b#u -W%KPl}d=¶`6*d@͠I/)P%h6%gIk+xI,E iyG)\3|Ө/E"{Fo zt8v;*0Y`2)SG@rh$Y'H&w0W 0jg_sV|^%s-J ^3l υ*^S$b!.,b(y{Mb&&%,KE F(qbOL5O*]Tc+U`.E[-7*e#3f\$1tS "DbJIm [:b5u&rqbO6=7OzcvXi;;<涶cl6x~nmtogР̩-O$d(B홛ۭ5"5+a4Yպ_ ޞľ„ϟ~j0Fbks@}ɵ4 ~Y5`A>j.G =οYZ#n\_cCd)3xy RO '`ze]WppsYLluDuYܺn+]m PS$;SbLJpQnDR""ɬ%HGLY 2Wm@6(lscHWyƆV*P=cXR,c|~0{꺰ZBXvQ4{nss8cc-@&D e@…` dR f7[JDl] ?{6rOTb;8/ƀW[hH궝`VIĢXEEE~Թשrw.wDڮ@}Lp\Y;'ϳr]h'3E0*!L=01 Qpl_1xA yAƘaN'&ewυlfl`58?&$}W gsxFHhS_@ήWqT%T1I-))W*M@clF] m分<șE "[,NQ" !EF$ `&*Ϡ8 - L%eT0i@Eh(JIHMi&3)<g2E,jQC]0RxNf2  ^-ܥze$U넊 [<2i=oW-. 0 I&:q\!b,'qjh%b@:m3 chobqWK$cYƁ Q&AA(S0!Pxs \Mq5FT xt867~T bhH0C 1|_ȤI||*!XB@$A'*D1&To",b h W l## Qn[LGf(b$1N#*vhLJ -*'A~?] c"`?JtqW*i8SB͗o.%\5'BIEHBC8ԭ8"R7{ۻۧpy;DGj nMp{xQȕ1'~Ys g'4E(qOj Iea@p^9ZjY jVUY n| /  $e 3uum>g.o>+!l8y>J/gY#'BշՂKpeGM/QTg \Pg7|(r]ZN2䳍ڄs߂g P}$[7lo6.1/IhEG7'ofCQ.?Ӓ,&iJ7sU=q/xX%te V*iW*R[PE<ڢ"Jg*3|=Y{ƹ \ql|M+0L2 # .0'# _q":`DAU# 1-0Uvc ,i6"aI=bw}{UQXBS ˵Pj%&KNQ"!JMeD-6v8!;س\XwS|>Š>x4ֱi.ߛ )>~k;CIPL֍v l~{(߮tbW/fV,IA\py>wc/'^j|ZSZiqo\‘:'="!}ϊQOju = cg>n_zcU؊X$'`^ǔ`~][SY|yFΥݜ z/VyX[s Ydʳ]!xˬ "筿W1Q3˩f2?پjG8C;FvD Id2b!#!u̞JugAo1 2eV2 3$JXoTT&7.<M6|YsǮBϕ92TT*'4  qʐ0gՕ^ i؄$[O09A9ze5_?I,;g$oPO'Q0OUe٪ܹRJ'dS܀} X tUv,C `āw l>e0W;] 6.C<Ad=mG-3Cs'h%s|].XPj &'LBQr As?tE%'&(c@ gvv/ArV&P[IimV :tG!+ Z_M]u5CYܥ+7t4F[B(' ȥʯؙŐ3^ ]GK\l9=/@܋Xʚ5y~sXO]uۄ5܋uk%Ubk~iJB)-ի>^ov+B2 v*>߮W㌡{ÛIkl ڐjǀOBқujc5ԥ:a‚zpFg+ĉaNβy.9\ǟEAn;osC}Hzb̀e{;cb^s*'Ϝz[k:+|e|M(3ڰ,}\[wњ{5. Q3f&{6Ɠ` L}qzԎgw;E\8t(dl>t`(n\&3Hl~onƌS0_aPjz0fsae\h];x;@&|awO say(x>F\Ot V,?6뗋(JՏFQ8GYPb5CmȐj4Ɛ7#tˎ"Y|f?.ݍ mw!MaD: ȿK 81'hZy|z[.nS@a $D F eJ.).J _J쳄&O=V‹Mrd*HUwCZH(w<4ҜӮ3DSD%ffkx ͑4chA?zLp"=*ۢX)2Y:1ʖ!#:A)*^/"?dA D[Rg! &w9DS@@Y>] pi5Kd*Fzl} i1s_w7T  $e!V Yq~hp‡c:1mjQ|y.WV\*IX)'7dP?݌llE+%ԏ c/oFpa?~*`(B ׷#uAw ۤ"gJvIN 9PA^L3@Zݿ^A8~>cjRBk/uT_{XW`6*"tc.2 W (YsDDQ ( Ƞ1PQp"~i_uR \^p'Td,UnZP,83A$C7$uk:QnP߃޿A<69UP#ΐcU4OySɁj/CoɈ1Ew1 }af ru bm c5klmSLF )^pk /ζau44-V?7+= ⲥ{Das+%X>*z()DD4k1EB9iLWt_N6H]&C L In}#͖Ո UoGV;IgI8}NmqQ6@Xe$E(rND BĂF"KrTNMn|V-*ń3^գʡF]06;t"V`+*_8Z Cv8;f@aof& 3U{RChzQINoz[U`Z\p l=Ά-G?N< BFB C *ߵ~(Hg#D -b/4Bx9e:uBSQ#]~ě"&aDϫ*νk@_8Qd]ކ{>J}J1[bPжsnw(#i?{WƑ /#}Ї@kbx81[MRMAb9~?囃՝ )h2sb,L@3o:]Q895'8:D-MTWk[ m^j-NX_^o'ZaěpWFQ0eLi5h6kI8YSVdQGTlئ."Es@<?l_ŧnc詉5ÑV]yZA&AO(|Rb?g#g>>Cz _iFʯnr10%c=x1'E^u%խX.1Ɯ&6d˰ؐ!+>l]<7pq}BfsV#m՛~X&1WQmMb(oYAh땮V"ŚZ혮&(&_,Cmo>v*/=Ơv@ 4xɦ]AڒR3*ZnO`0i|q(hݲR$pMǼx3MSS8mru/:zdӱGКXm5|rn+-y*oԲ]TJ e6!ale%T[Gޙyf7flb%X7^d.ɽɯ늀B&?3ko 4%D:(7%VAQq |5x9I{ Wީܢ;Kc is%X.-"y@K㠴SRj>0BXInHȵܛί=gߠo:oF#q7p},ѩLӬc2{fgg`U~X0[hFEEؕC_TH-Pc '1 cihːCZq;29Ypy рp4؈]AztՇܣ~Xrޘ sp" "-'arw0PRvc`$N"з ,mg~ h?ʯ `$ȬU{a\97F7z0=iIowtrO֛kL?F*M뛛~gbWD^zlUa:;we!ܗ* Kx1͚f;4*#| F? \!2&4L!TJ@3]2m aiX0m.X" hWC,u"|^[q.\Ź$sIVKRjik5R(' jagRp#=5vE@!8[808(J^B f /ކX9EZ  Wo [TK˫]' W7oں⪸[mU9 31fKƠ'd8E)uSS* T#5V|ױrkaF#*w8x*mNԕcPs6LȥPHA.)Ky鉳 4B|\{lsIZ9"TBNO>TWյ"+kuX@405gzI Xp  H@$`88q&IiP3g-\$+Љ5 ղݤOu~UxO>o)o;}3|R`GdI;0Hqw{|'Kn*@ &!6Xvb?Ovlq.'b(oɩa\JBGKFU#geQP*<}ol*A4::apenɖv=$!SMH\϶Ă( 2?DKφPTaҠkz؜~lu-iC?^K+!ܨnJ3`PK#^׬>|ߓyLesj66ð&43,yS_)Z4kEM"*%uS: 3| yf-F67FٴuY@f-p!Q|,Y{|$4mΙ!-c8I m+q aEchdO{Ni38Y{,en)@(3G{Jaf3g0%tV )&CJ J LEF~yEXku\Sk8'%t؞xV򌉐7)lhv(NuI"e4rN\hNUeXT9Zʓ#'~3QyڎBӝX]H/^*pb܌ؙp' iJK]uct\<)hV%'_0М86..޽=PYwM܅;Whʯf7eV6 Yݯ%HmM(?yIƿ²pap0`AmI)B煗6I.jbOPn.KiDm^b*֧^l@@KV9Gemњ;I*D53^"q*U{ L%0x"k lD[bD kfD;#]}_JVH3r+.e! 扲+)l i RX1p$𐣀AH9̹@Np=Ju\pj.5 2N,(Fmfhp!HmR3+PkwѼeeJKH:rE=AET;BE4>;$&0[tB;shC3hG0XqЅaYmzEAq saymr+QZ,SRJ!9'juq#迈n(~?z%|rX(WA!%y2'bzb QW?[vOK *bRŰkg8:_u?=g?Nz8 "άDyTg&*`^m]{1guϮx {4)q.CIK2e*#b{SP\j$ıJ)& WL+6KMFֆbty{ȩB;A <7!O_ڧ6͚x -ųX{ptV>{X?`a߁i>zH4Hmތz˘_ڕ`Wx9LQ{?DU:}ջ{އLN)j 4a嬄%IƟ\ٸj^ref^M㺂RT>[R/]FL5A9J1XU~$X-ΠB^)| 3dJi)ncoqa "iVXZ 0zр}B(rBU̡>/MRSaόԩ֨KIsեQtGqPA\r.癞(Wc&u;|$IʆnUSXӶ]p%ndX'Jptv!zv.o<2=X:a2s,D,7' F4SF\0 ƾ裉_4́rbV0ɤFzj"@abͭc) ܣu;Ճ b=rhλ5<+($֣^VQBHQ; +)ny+ri.+& %VZ*U 0hp8=;|;OCo0+NorXWXUj\]EBwμ݉O{wBy $ʜblcD\.:#S㔍uLЭ)oOG]$cb$JQD#sp #; 5x`TeLrY >?KE܅AXhAcx/;W)3DPv[*_v"P_5SI'RkÌXK֒Wa LԉHAce d§Evgvg܊DZWo0ı: B2k2l+#Pi<6I,*TL1Zbas`b0"Qb %ǒ%؎-H7]O 3vb oMߠd :Ӹ8X Y [JZL.I痄!gy9#϶π24ΒQNǷ7/;ˏ.-Lތ#I"LZ^!.bE"+j %{ H1ηןP3չٵJ:oPgfU.W~_X`L6{uwq.Ž^Byȴ.I=4JϭVXb82W,fk.Tv~y?}y?7I_d2'efЍS#ӪliK%wO/3U.WeZeJG/˛Jovj3~Yɫ˜UUlu m-|2 9TKHr>mQ F3@ W[Z:RPK E}`@:/H.5QGײci9 7Ă|rB=c\RfY, BZCLK ̃9ѥ.t)K9+xET8%o0<98\-S0VR&[.gy޸{IE29GE |o y#$e<.bQnN0F|) ɗFzKKN'ڷɗ(d*:\N|oFWɗT$Hui|X >5z7ue9b7fIYƤ 'ØԴ._LR0ɽP!-!Z"Xx<,5𬔁l"hMOֵk^9fNC_0HfB!1T7(`*۴ ._^?Q[T.Wfii\> ,!}LvV[&ektjcUw `]ͧH4 i%>/[yϕҬQU큄+WxUd mSkUGyud" ÄھɅ,xLS㾗 ͷz@XOp(VnC\ʬ黬W;'M:2feBeTxPHfi6U! K9IK'yj:9XoB5K:Epw^2\3MXJdJ9bWbV{~~j;1/޾i >)KqQ(cDsй sUg&EHbBh{M"Wt͊ByHyTJbΣ-:ΣP$,HDMx&e6vO, V 蓁~#RrRgw%a0vGa`u22~3]ʕmiGP03+ %掔cNhs ( xC3LR-FOa!}:A3ut)HS&~FvBI0J:Ğsff! *Tp 7R!C㞗J]EKQ4ה^ԔyA]!ȕ$!]Y'b0wRޠMAuSVn Ч/Yp'}o3Ҝh ʴwRm~{Tk1*{l;~c1jyG ;1cTf1Ho\.`> *nqPަ~'ցqg+@ݨ>RpS#sؒ=HTUQ4Ԯ'FFeFӃߌ0cAS8(6DN#\e~,B[¼3yJ~Nuru6ۋ8&i~dz`D:?׽z2Cl>dn}KTP%;Rokt@ٴi-L+2EϬ>|rV`݋TǼzo~F>lm("v@̅5e~DqV6a"XzD$đ/vrœp)O]Ƿr*aɽ'A€{>GbLyI` !j|9*r?,Z3UtXj1(s=#UtNu + QF-Z ঳3JJT\PG"`>%?ٔ{{p\5F$^7ۀ%}M}HcDf{V}Yv<(c̐q \Ey5݀qL13PG֗<D]"N.zU% Ͱ zxY4z@cq8:XLdKCoBgK 0ڛ+{aq՗{N?'q>3sF)ttӏE{SP(Ar`ݤ>5ߝ7K- U:86Y?TZnZ嗯gn {(bJs^D@> qsh⇧_7#b$ }e1Ax ]!Nk1YʥX8R9?!)ob&OFTޮix]+ hU( ; 4P"PvL;M\h:ޜF#[D72mM =R2[T{U<6KMޒ@lKc;e;;)CSF~1R:=?} @ g]k0'&b`5ho?s'OPQ֜pd=*L P*$|rWmdH9auHB3>JkX5߾-4q~iTv:PyUik:HEQްHBs:aqZlx\3b!BT5Trd«n@%CG!|q>=Ԥ~9ҍx4m_wo}7M!xn̆߶*_yNJϱAa,n(n;ͬ:ʿ ÷0e[Zqw|u,2/VʼT[Vo#MFBAR\qaɭeA>'Nx1^0F> 🷳M7Bnfm z;X5YxfeHW9O*bu5[օ7n,W7x8E1x0"ml o11BG{b?y +nXwŠᓂHE; dfw[cSېJsWks j#6hm LIV a|}Η}kaYwA Usbÿ,}0#=%q6xһL!N-Ϋ}}\ߘu} `xic:|h1ظH"# lX^}@c E. x^(Fsr nsgAC_^[;ñ}ʱ#N~Ϫ:ky?7)̳r}J#؏7Jk[+S`&HqSǣyy;/rQ~[?U+rf˕6pArx4gZNBp:>Uk[PS7nǜJ-Iqm/I/Ǽr|%rz;\λ f,  #.7\)cEp4vιR0ic׶O) @)zs(PpgfEhEQ Ȅ B 0 NShfbWj^>Ko^:| Y̍1ZXOL΁Vz%ZG Ub4LQ>zm[Pٷm[#/Ų# y!+ R[E,u o]lan2$,a(^S;?>YaN{tas)5RK[&;aB3APXc &Pu+8;xnAmma0/i+_ݿ櫥)g۪?{ݟ do^qKCG$Kzxx}yb0?Z4?gr aIPkP+,oT EUfS)vm@w`T,4C遄GFaI)z TCۤu>^qKRrp bZ0ԯj"`$'l꩖ ̳x`(*׫\*Z\ }_kT;\qdR FNGateB^hF/$ڮ&j;P!fgSNPOr%:ڡѺ ޹ёJT-71]r)|43V->U䧧47H%4ݠ$ M$vY.)TZ)-J߷`5"CE,̠Nq,WQ{ВǷXHsa.N/haM5$sVr9yIB #GQcMΝs>W( ʵTkWH Y + .hQts( zboj}jafa5 ,|JvjזIEv5;P:nCXdX`K\e~o^fLZ!&3,a Č )-2=gJ6"Bu.c l, AO*%KjXD((M~;.P2 Wd,4%HhF"2!ͅXǍ&gLLI*PR a 29$o!\u!s"TFȇ Z'(>X+-Rp[Řk+5ʹ#Ms3B:0Υ˲pZk0ni3/p\rn qe!Ɍ9PĐZxQ̭sei#4jĬG cL2sdg9pV+Nro 9h@q;h %ȳL4[4K'!ƂCuN ))ɥ%m ('G)ERJXNjjJaqSLoHo- ::.M{=PQeX%sIQ*<&P%o0:L% !r"p["T( nSph|4 yUP)x}e!W!٩׼kX=_cm eiu{c>FWn.` rB™i-XAVz}fLuDjZުJnX88$ѠϺhqGy v\ e7`0б=PtJa86Va;vDQYCY Gh;5 ,K#c)h8T͵gԨ EI*'1tJ `DH:Gg=чm0TW0 c@0$u48J( ?UVLK~Y=y3nW.-@-5dA$-z'g)jHicu(msq>sh$_lttdީC%Mԅ@ԅZNHSLdӪJFr!ɝ|B1 SZ}# R:‚^j!m6)zWFU>ҴQE"ou;RGS zmoKQ#aH.jLvF qWTWrQ ZcyL@YEE .`'C_I.ƊMIjtވXPSb_|[/͉DVgQ&}vw'mƕHosdi0R׆K<^e3dMv(TvC0TA\ͫC-QA'ފMjeO?6-]umΤO7>7s2ካ'i2Gh0E!L |9D433s33{q=m`XWB>Į;N.8g 8KK,:J*Nh|Tr'L7\17sMcض϶;Ի2#w5FoqNrEۍ^rѓIwuE)#IO\4`lĶ$| ilW;Xx qQkH(uvÛVFޜR<e"44e=#.mKSv6@钨~*Pˣ$LU/ H ccn o!G@ ۜtj>u#cW kj<^ʙwhѢ48HM1#$=ά~jlF%kOD蟷(!\YгBey:@2}'ﮯ)Sc2(pn<֟x:x P^< Zia>`ZPqT >֔2%# gR& JLI^R>í#78x5xqm/܈ިϠKRL͸S^8%?@ߕYMBQ3a0>Tp')FgZiFSq F*24>L!wvnTǞk I\І߄gΆf KKTOm*k$t 0&.Z_kIsv `ڜ z9QHr\+Ff)bP-3zC$;jH4E嚤NόR)XeIi!Hf2.4:j@_L.26''jʚ_Ρ+*Z?Orκ>CX `gpJJU;94A;~cuIG‰hQz_&>-=G=1Dǣ?>cWߌm/{O66玖l[w~ۊ'B]EYVw'dc *8G5b}O;?-,.MS#Öߌfw *؋µP^h ?TǟѷS; 1<o<2[^|<-&G`)CM-&DK#E؀pV!u/)~;~(G>\5Gv iFjwHh-ȎN}hT$IҭB0aHE~Wcb}MxߤL=KJ%V Z/rWGz|ۅUgǻMWdqUɆ}<ٺ_ca E(P<=M'7߇.e}lr2DZw.O8K qP8q^ߺ姯Iyۛk-ikrp\vi[90 d̨ӱyuF=iʐo0G}k _Q "z*/(c]&PҜ wbdCQ6KOZRe=!1"L9RHsSrnu,͉T4Bl0id C@rJb< u3x;5!00R.mD-`dlgaǍ⿴^EPJzҾD7YiTeB[R!U+u#w ;.: N)U׉-ӛAu(s)PyEg!Ikl|UPRB7bf(3<}p̌͗!3 e1LK:"%\cNεTCrmRbG BR?Fm Jn "lf5:X.ĩ0|%6}nS86 zcPO& |jj 6j?Hu$l+^~1 8RɑAơ?^,@m2zrj`G)]hlb*{=?Qĥd.kb(g0~aІçMrfX۶rZDnh^ڵ*t%pQ9,l}0vCkw92WpaZ Es5Èz}B@@ kY.x%%p覗v"YGDMLzhTB3 \I?](]/܇{JހA?L&4Ç؄ed^냝\28BmADŽvWxl|貦F%+i@9-dg:Ds Ù8{FISe?^gI$ߥ_jÁ>#d]aϾ-B ƮDVT^O|6yHn'm' gI3 B BB >JiJ2 PS)8uOv= b챊Y~evm^t=;kѾ y`0 V-꡼D&GrDw%_}RAt ?0C$;R`Tϣ`&o:WY ]sFWPrK :^՝R]lgKRfF"|_@QHJ $˥MuO?] }jSW_+,ɳѯ30'U_$/j_{0_.C܌z wᭉM_z]=!֗;/t #$Xonp_6F St$/m'Kk!=*uRtI6>Ѕq)O*$YG'.O1u=>V#՚Q%T[ZNDsQJGC*BV͕'B}oF|B *du~wEE[wL7N'(uq a-;5wePsA]ܪp#1 $zxM4RmV+ {*ă6X; 3Cr]fhbG0`,0W?=s8IhBZ- bK (|CT_lQ-pt3) s5n|§A07Á__S#p#CAM d̯(/[Ct5As0oXuFpLv~F&bv /Oat0*aTǔJdC1!TJ!b_EAG9BN[R}PX+Y;kGiP!zF J`b3J æs'S10U `< qʱ&0&ceՁ@ilhCR,I)RڀECr"ŰަA aS' ;-ʡE "6ʖh_tQ/>@X{AX P ]ұ0vl ~9{mk7ovq{$MTJ 74f^XT^v+p0sWb9Ώz"7=˙DvWy`,F7>0q" ]B` @p8hynxwTS9K0M=JnDD+)LF3D$$VN,!2kX)5"RK 5' 3\u}C .ۻ,a1cAyG03p@1q1+o b?-czO/d?wf~壗}gr2;h @,f-%0 oK7zӮF9DDAD~ Q g[MݺGf P{#,E X?J[\YX8lø‚yLzJTRͬ.-ќ 6p1Bm;M5">L2gIp/|@PQ&fw"LtEI.(< #z*F]6!Eu(heJkY^VFA}3ZG}7@0~FP5\*Xlc*?V7>LȅyCM/Ȗ;?Ye)F^bDkIftarTqZՔ~b]z~6<|iBA x\Y[y`NWts=Id򐀭OwhwQ#3^“Tf&FǗ9{9Mn /Pʟ/C8ͺ<5Q=$@,$=w;h~KLetKjګ|/;WY)UҺ< $˞;]Ǐ1)hL('m۷h|AC, &btMc\R$G HPѯIpI& L , 뺶\E5?c,ɓm ]CU}*&#7t=J˸GCuiddR?oVDX.xs>rq8՜ڴ{EY#'; E_?:#K dy[4]l$ e! h-}fZlm hp5]޸soGWP}6CT!RTׅӇ2=Uj ;t(At/@zDcZRH㥪B*O]\Xcp z.ÁɇelzMS78-Y %P4k½җuh%N2ҚjΨ&+\ A-i91X: Ȏ5c8֠RI]ݪ6/CbQ@/iwЯ5b-Ao_hg=_`pp~z8FB[`SVװgh&zQ8SZ3Ce[^Vu.Wxu[k\uq& #HX*NUӥ"_= hk,$>FY,4RBK 47b0+ 2;AvOT@R!yزƚ+1a/f {|;iX%xǭ{'>t(-$}c'ɀ0{'**e( ]H A.NA#x &^ =}rEygϯ+@^JIZ#ѓzs2Y_)}Rt9薂wSt{ N_љFZJT`ퟴ (oӁM;]eq*:SeqAC#ɜ*Pb!hwK2*wǩF3*Wb -Ş2SᾱGmxo=!g\ӝϯd4 G7̷#͟|J&UC'EkRHSd96GIotj5#9n2eOĦ)J*x*0K ړTң#G$$>R8 !Ōtqk1%.Q%ܤ(yATY7mmv KظuPkyis8%t)lJP+u^0jRmW1o{vUJpt.Rz܍H gNpr-/)(OέuŴV 7oV,rSГo1eHdU:%=c3lE;J[]=^XCn Fx1lEcy{JoLϗDy:df2X^/Ձ?{cH3*$!l8zpb$f*ᷔP})'Hߖnmj0YcP*Iy)*-N/U3ƻwUT4aWA_|8<\C%+g. rmZcрۑ7 vl8ݦ?YF. >؂,Ž-l R_ O#s eGNyYhNHh20>?Q.m?(a\JXuv,ؙ.!R?>l3D/,2 y Ċy3$y4"T&T)FrN1&?l=S :L/2,L5OlL"U3<>Mj#cFH}H#TCP7yƔjUJٜW6)yn@3ê/E4KZsݻלhFAO̊Mðb oW ZՇG(̿zݍYfkr6_d/ #͕TfBL)3 Zpsv7'kvi6Dulay|x`B(>|)NH пz0Zd:‹nY+ Aٞ~~r;(dBVQ'9;nHKD=$.eyLrVx/Y~ @,{KQ!>1jVsu A BqI\"-|4T #7,9 ` NiuJ9e+0`A- \0OaG6L1LaVX5/$Wώ(bb@/ZBbKJp SMMlח7w!Gz <ŗ=vq}꾝ǫ3T ~\Oak "\~!DjV#~ߦFo^EѡW /d*;O=)%B$3`5 +/Lǫ# Gǻ~'!ͥTp^>yGC5&XrA,$ΦQXf-Y|nAI=L)fa>]yݶYt"cĥq -؞Lly|t2LЎSX{=,Vb{2k6[.ćiκS%gyȬk\hvô4ոQ||S*C䟞Dǝ}mzs- azzzwn=ܢNel?Ow=v9$%⼿ - {gq2ZPsWz^K.\pX,Pͥ2"$u\9e uPEBa[Pps-Q.8*AIGJzǖѱX^eAkz )8$=7S(NۛNgZ}w]"!vЫ:ýcG"S ɱ`ЦR{S F.&Z4ϔ8xsHCE,X:Qrd0n&cۚS-n箴3 #p*." o5;D g ąҌkdh,Bx`Yd߾튦w4RW{Luh?,m!$t;(9dӝ{/Y7Ur7?/Gf|ޭv1CvR٧"&1d7cLOa wvZC6FV0>$[$/Yo5&Ui_9VMB6Y+V2GYȓk줰BhfaE)\d:I1ɞ=5>~)\"R%S5P#,(9gX wE==-srv OLXMvY5ݩA{V^Hz 3-J`w̶HǤ*eSY=oN)ӒUԕ!|o~HFI5\gζ35YRE$˴$`Ksy,P9/bU_Gz?ىE_\J`g8qzDXԊkĨӅKm c _<=rk $<k:Ä#> $] .4}EvXk\|ܰLJUscc8ˆ ?PB(ݚM(̍JIvo<&^&lO)&vl<`AP1fL(E,A=6((?ʍGc*=E/Ɠ@֝D!oni>߬~`(%|o75I~yuiTͻDiN짐o0=cJoqɍ3!IHAPcQځR1crEHHaDKcw h n=q8>S䲐 {4Va 7R4E616ʓ0vy5r٣YF,AD0*]EHR2J@H}.'`> v !X${pBO*kD/͖-gaHnovSp / HրD$GtTȍ.ք91-E9dYA݃ᅵhUr (%6sM]U2Mn^޹Ε+yGv:9}ё1Rv knē bBW`m4&LQLwKpW>m˺*kp̀G媀Syfaf{K' 6e`KV.r8L+FrQ94֏Y(d)ЅZ2N\ِ?l^G4g€YK8A!Mrh*Be!P xT7><*/~y ]Ͻ1msEV|qR6anTfBPVHׂ+TdVy R~5z1tFW!ם'^xwy-Nk tHV 3ꣀ:'C3  Y]9 [R1y?~ciMYPdx)tAz 46QSԐ: ףCVR}X,gwA}HJeK@/- )MN:,\A?A z8WG iEm<䴲cv*qJUT\G-a`_ UyL'G{S\j vythZ;0ڭ E4A+Ֆ&WqS-T8K@&UXp_ VYT"?O%LIJc+04,q ~谧1bTJ*` A5! d)LCCâ*02q?LHH";}m;λc@]c۬ks8”]չJℒD2tI{ Ͼ"0VN2KHG7N_v!XcNx@eu:f*?n0)Š]|ѹ-BfsI&rI0BU1hRx VlSS$nQx."x{4 2չA O6iM/XQ=tl dǢ?Hi*[{W㶑0.+M,0Xl_p[؈|J`tM.I'iW-i4iI0p2SUUT@>aBhqXEsVWwxS۔ax cH&,G1a-bZ Gg6-Lj S%C}D{ a`JiȽyN5~i5.v6AY%dV4:3canBQ&p"SGVHI7sF1aFMBPhRpseϱ̰@%&qN%c*l-'L x1KU DRN BhOh*b&b7:x.2e$"wpi<)2d8fsC@$q r QHIq6OV (befcyU;fߕ`GD}}|/_w4%Fb6E b3ͦ~vG34Te 1Ѣ Z#B / BXHNa ,)"kubbZAXpSp91,Lu< m0INcȔ(V Y7nR䴀@[+HXjrG~XwkOouu\ {_G=& [m ЛYz7an2`EZ{Nګz; ?Nb)OcfF!<u!e&I׫ڏ]?#ͣY yhQ`\v. Dɞ4G]<SA#7c秾[n* }{ QuԾK$+>N(q6ケsf s{|Hl댌YkFG4ԾزlO#wC˥j q]J>^oaoatqY G?=lhKt=.kkK#9펔_Lʱs-)Gb(wmPRT()G 1Nu4tĖv"B\#'{.\RtH-a.7?S4_|\,UDd<"DqHd8U"Z}{߾#6A 7N~Epx 8|d>M%uDTAJFY4VJR Թ ϋ:@@ \'54 r: >s!$3ASI*A'=TV&*/hk VT^T4.QK3Hj6sS4V ND2F FB  WMlc˹=W6cX$5HʞpR K-!9Ƣ׵R+0'uS_j%g# i#2-[8uA}w,5,S> $S:)!|ƌ)y&YzE>b)mNGJyC4jLQԺI -*|QשκokZ&F+Qc*];-M+AЉ}Fv3q6[ITukb!5,=a@F1!ov&P}q' `jدљ r"/nrZ8a Q_!Nș'hLW'hRw&3N4G6 CL`-rC0 M Z!2 CLPuO'h)8a \\r' qB3d8#0 N1&:, F$XT,wzKWb1DEYD߾mKw21;37wo E13|?yB c \ӝ۱ɃYKGo>Yw[3_\TQlUzi] uT)pMEЩ0=z R =jB<(X}krVG%l_FJ0ZM-(aFsͰn L!>DѮSz=1Qf.GTFo|5<|yIwP|ܶ51-o;~:?9LI\1ҎQ%bD,e{SBu$*E 6^O.<;_=̜bO#ϹS,MMz\]_'Q"[rrFtź@h͕%vl~F1 ^jx!q/n5;b_~0g.6-$c2C"1)l͏cZiXk\5H//wq\,~%"$!)v V-oj[HGȖ}ѐGWBr)m8q)Jekǫ`V~I} %d#bSsm~^ ޛGAuڰiTh"E 0$iK8lCdUӦ`H{ hY6!JЖJqIΘ&EBp|5I.jz}hԗi\sb5|mfLd:DŽ'"R$ucg3CX)Ox(]2ͮ~W/7{n;__c4!XND]1FS J|sesWRS3m9L_`nL;IZ$Ր[qq?jL*$S*VY!S-èǚp$T̬ځCzBR$5RBYa)F:EScR#_pJS!6I BYYRNhhI8cVbEbe"ha @ Nv_>87ʲ׷ta XbW( %y$'xqUI̮rՌ>LF"ÊäqrLdf-!AJ;eNV=ʯ%ʰd(uY#ְz9iʄ>`Bg8;s,()mvoaR7`̩eTHm ED$Y0:8S&yŒA)MKS,ec/Z!;dzR_ -m6O׷{Pnn+F-2ݼYEA#.qBf{oditڿ>:cEa/D1#.WT9X, 6)ǒk )ld ְ6==KS/$a 3;~U'.!d4r܆xߍ= Kk9n rI ,Ina/&.֛/Y@ìƳA>yA[ZZH'i!}TrҚolKXQL hv+T Q, $>yxn<!|ƌ)Jx}n gTnsfNs۶n)֭ʇhԘKgzu^`::L68@jJz>>>\zDg:4fòu?;Gbk} A* t*DFnXoOe֣DV1"R{u== 99L$%TPmo1C1('w@|L)an\>j9Zw1Oԥ"!Dv)2(=2v]΢xyP 38*Gp4?arhBHr.*"+JX^>9UۨT5N9M1c0EKrRBr5 !(l 0Hݝ ?^23Wyq:|__4|+$H!%ϸ@/3pW ^FR} RIaCE6Tu^]22-, RAY7KQ`s7jW\;۶X~@?Q k>[ܰ T$ma Mⷿ# a7=ȀIyq4W0.tzRNYNMA]O?++T8WZTB!Hb%m._a)C2O0Ryf :$IjUBneʌJ;œ`-;I[fuIT>a'" 2&.%)ʜAd,V1) eJEXf{JEjCEJԫ*+LTmԊ8#J Mq^1B$)I5% au \4hDO;;+|›3~9ZۿٸpN+kTpv}pʑVA?/W8K$J)Xa% IƳԊ$I`F+ɱ`)/7SO=v*!Q>f^t-ru :Ĉs}XhC[w.?\ ZZl l74 sm:}_.d%1LټI(r|b_k..q||s7(r-pp!_xwuZ,RЩiXZ ׃p:\Z\S&X,OQ^~t]Xm޹Xq@@@D/Y ҆i" LI@$Iz8#+V|T{Obv_oZZ@a=@<&tdG`]Ja^h a3n[8r]T(if~kg+-A0 i F e=p1zlNJxpv5 #OG 6^sޘd1glAb>]fa>&[lKbFTğEbp Y;Fx5W[]؄UlE@ Ԕ>kہb5yo>']3h.co^n FRCiZ)9̧xwT$$AUiZ$SA;CuXJ%ГIBTN-0P ]f&5 hgPH!b1ѻBY!ey+a= e ?҇VИ'cۜg'o3}߳smt{\E: [YFK Z{3ܩP I >ʽm Bz~Ѯ]IR yl0&;KZ0ܒ`֌tHa抂5cq3!sf?򹏮ҥ:uX{pգdnbo=ژCƤNbZ6/{ ̹Q)[xD7Η%ay{ty\KiFő`A'TjE)AZbsh&zfIEbM#>eLMD(a1DH@"/T}` L*Tj8H*@(ghB:Ae~<!:ϯwHv&CE8%8F 'A'\YOqyXi҉ UqrZ^Yxe+E vL3.cg=)e{dV6a{)pm.j1"9lʽʑI{ DCuaQڴ*Q{"#Z -1 8Ww.]ӊ0fF.VKZ 2t jq\+n\SV u)}4ǩ>G൸QEʼn[yֺu".rw*h!x_-)$L -S&|ʛRӹ V8mn)yd=R!Mt `BFtuc?KD28 =qa9cޞ1^kOa2vVDR kx]hsm8RvI=P^G{ vLDyo"+fa΅68VDJPQcO(TQM ? DFĒFTbVfh¢PmCM\8tSO2bwӁeguQVuoPe]O %% P_HN1>)"# QKnB%  $ 5ġ 1_iYAΓ6}*s#mܥVAs 1Z8 n'Q.n>wHtBZan1pjQKR)2Y\{w)Xw$?8k.(l*qHbxS8B}z$r፦#@R`)VUrBVX:,{-i {bgcǞ V\!X {\zؙ8:pԁ[aY[8wPSL\q ȥ00=RM,;˟qd8DNן̯aV//9ԂQ; & d`aְaZ,׷Ա&:] BnǑ;KP%dV#ɋ5'G%sZEqHQc\ 3Fqxgن Tm:[*CC 7_{KX^$F)E]KMH> Ǵ=2&fZ< "qLpQVCm(Ugs$D ]gD:LNړP3}0#0--l59ɺ5+IJTutɪs<3jѺu ڳþ@- $Awc/p?H'Y;HoLD\KkzחSX)0/` '+]+/ga-Z ļѡX^:\uG2}.gڰ/MDŊj 9S&@firٛE|Ѕ EȔ}93BM! J᷷v7~!Ċ׼@.pIYQ,l؆\oս^L=YN*Μ:L{HRD0Jo3ԅRgJ #<5RqPC}RiLt8LA T'H,r& <;6ԢhDcsLƲz&\QPZwp9r)wLqO2,8.U5> $l$B&A1aqnj14'ڒk;{PcX\ukK;[Qt:K5f[*b8Y 'qS{6 ̅6Xd]ǡDL30<fe 3VO]g?cNhLpx>I䳊~͠]P*PxgpL1io?\n΃݊¦ݡ~nR͋ٻ4z2ۗ+8/_o-<1٩#/?1,3c//|A2OgC0.ߧ/7Q7=/). N0?cӿ5~n~ZWydK`>6sI׃7[fȉ*t+i 0Iܑt >)4߶l E>⻓^nn=נ&v j].?}?Uƚ3?iNv>L5{foj;7XQބQ1<.jj0ž[m[}@Wu7/-ږDM߾YVE]z9M9fO?h0NG74;>2ffvvU LT; `z?p0.](Yۿ_o^@23Q6ӫ%&?O~$g*ًLo&y_~sp8. aSJs6YLCÔ<)cR NFShgz4L!pڭ&|ϧd93_ W}??'Pz~7S,'[<ԿYwxcw'h7L!|}&3sCf2J6YF,)a-yЯVP<,B$^kKѴrq&?=Lۥ09^WK$߿N"Q1jhiӷUDԱmgmDRna}K[(2c۷ok;۾!%mg#:Q0wPc/&s6Q`17.Ÿu"C=A0gzXJ!ul(a iL qD&Qo}࿫_~~f0K9 ^gЮX_ m.կFN)=TI q&+P|eaGl?x3)M]8/;]Y G] ػHn$W,vM/f1˰1;kR=%KbYR[`bR_wKUԖwڴsQT歁 $=z gSLγZjF)u[)f.Ȓ?_ԗRSPiJyS":NV3b[ ՙ4? 7Y-iٮF}LOHdbUig^B7ss!t}!; C7:?!Oz_ŝ׎i[ؠ|BBsG1=x+FAOt킘DctMGK%\_xM'ݠb 3%!L [lZzDx%GM\}mNrY \)c u#pk*Ώ0 G֯>">,GVR[4Ҫ!_^R30T:p J LTQWHp$9R+dw)R !RQ}?FoL(>bRƣ˭ Z-Ɲwx:[n2GndT'[e$iqh5;_*$[1i}R*5OzO=]B+ud=~b"DQ Dch'U~9/"n@w\|w/^vsˮYv_1HWVB)*?diU3@J*4V@]{>4( lw_qR6|Q0tWyDB,PD`!UU> ;cfmz.59 ^M)-u_CxlQXsR_nJ)%u[)f^H,<4dJ9Y)ڗ> +}NM)aWn:Ju)JVTX9!PR 3 YqjR\J8Ӿ2aMjyKs1$EZL wɬWl+4UQi b QKu&2H[i&T@|@j^?_@E]Ľt~]WֿyZU1r.wz)+xrI<oD:_DOP_Y9Jтy:I_ťo}';?gǛD5Բ彝>?A_s_?LNw"p8C qUt/]p4ۏEM?q3oB)Ǹi71dZh>Nzmr{LL;=En@9WS' {߄RRo`q񖥠=xҔk\vgmHo#PZic86pyiN!R [6lTQ<8'%Q> y!pn:zJ*iIϱScPVBPy# #KϨ'Z*eJmC ghl8 z{Q֡jz(WYkBawP@ Z-[U/g?0th O]F""W_u rE5?2ń*)PvIс8k VXgHM#3pŇh4R6DTfow *Wx"̄K@(ysUsfE“N>ѓTD p 9'xD% L􌁫0pU /׳B ZzԲsB E ~o ,[LJ1B z3 %yE EfL%X"GR) C.RjAyGDM*w*7!#:N:)P0}TzhLg 51hAwZ1Jd豄F)FC%YrV FķǟM\Tjפ 鹴l6n˨5wyY~^Y?ܿ?/AjeUA,;E;ơ:Oi}oRZY P0<Kc k),'BNNhq}}$Ϣ*`)ZH.8CAc=xifL:s]z (|%yي{F=Y7) *\FtM7ZpUtZ QWr|qD)GuX ~3rp:8fiUU ,S*.TZ}cpqJF/xvHVoW6hOuSuqYUcU^zE^~ݭ#;*n/޽}0Cg#@apDǐ&KhrsGL(;Jh_^DcG >YW:Ēi,oZ2e=iϱ=%vvbDk6_\wxzt-cv#k.7] 5#YXI 1eoz0|P@^ўrԐ`0*q;'+kϿ[y™+(2V!zLEk,O2. d=7w1nKi/#ןF8l\k,vj 뷅aBi|&]jE7)pꋥjiVZ`N{`qR+ڰǨchoT@,W:g\%ܒnvgRgy;c?ۇkiŨ/?_l&GZ}DD~^n} ϏƧ&6R 날j}D2]+r}L_~5X\ꅳɒ\d d!߸)h48uzۻŝ1xNon^I|<ֿH,ӻ a!߸)G oy7ozNonmzN炇n2hwJ(jV3yϵyV3qP9x,# 2Ih>ZIyϻC.I~ f,^f;A.YY1TJFs ~=diy3eu-\ D(;;Dޣ58"x9J`Sg~u놡Zp:ZýLSC <%aI ,/ňQ) umj9a/7Eը 7=W=s<-/ g;tMNZ8CqMU>Z`H ݇l\L*>JbR 3ԻJ7H{`EOJFdatL/ ӆ1_Ac$c AǠ7E4xHH텮q`RQE\{ %{P|㍨K2rԝTNԢ2բJ"c9Mim[҇#h΅p6Iм08 7inʥ:Mz$B7 y%!BrhL_ rVw}FQ__EW?aUc?(+LC24?4Ms;-(VA ` 6`e8閗&qk+HeRm}V{FS eJS)~ln7!2-oo]߰iͺ %AJsUk/-$U^h-q\i8[1DrF(bSrVZ'zGv/}F^~nH0LPFJmW #Lʸ TP:QL]Y¼iKMJMrtjE-cﮗFk@*zzCBErX' و ɲǐaaBfoCFޟAavZ|QÇ8qwT SEcc'$Ũ".jgeY49T%PGk–kpmFl~< wKї)38\lI'h~gSޡsMx'e 9YrH$/5tkJ} %u@[cвGL8Bhjd3܈Өv?uvSEyNgٿ_})qq\,Y5?wQ%JjnCfK4s{w*?3ӂUϺ6ȺS`NȞSӄB #N/t(r wDC uAֽ\IhP _~B>( =GHF9.@*yRN>ء&F-}[+qJ 1`0K3<0EMcRB-\Džuc"{ߥq=jEn'oo5)ؔUnS  \2TTw/EaL81 nl '(yayotAL`;zgC~֜WقR P»JyThNCIGL=騛LwM'ݠTW2j|e(Gx= 2&s5jDH99jSG4O 7#%IR2)^'m!Ki\PBܤ퉧dPY/ΡB*=pT3~P\dwєВhPrn8ٻHn%@,>B` ܗ;,ܕWZK;SH=/yH+g[⯊zv+n%̸.bL(CV2@Ka7ާf5uJ8J-3F m@Э^Ln{fDTf]CL1czz0ړYo FP<jyڬD:0Y,T3Ƚ\kj,䅜W_I aʫʥ3aks=#rsAx^s-pw 2&Q2/tcmG/ĚncՍYb"Sͤx˷X\r6}`C0u"-N#D0*8JbzV)be).zٗWOjija0A=7jҶ3hĒVs!>Ls{8֯2zX䱳ȟk/6E " /=aHd8=yRgN2"1HK)o'{rmϞٮn\(;e16\ h}HG JRBP|9C7xLԺ48ZAPOeWUH>eA!_ NeG[&.dY$ΠG#Rݣ҂9QTdvJ a B!]^dWӆh!dMPE^5hh~o2g;wr }dӦ=`05y҄كx@3V1jYr*lÞ3[P+́bZ7FS~*CӉ~49~+Şj7U _m[C30^4+SG({ n=; J68\x2ʛc;'ckTv|㩮K4~|tBChnf(ie|v6 z3;_l4C ĐgFGY .}%{ Vծp7HG 3&̒= aSnK35eD;"&}΀Jp&2%dXȎn%KYdV|Ff BıtZu)8quf ^ȓR\4LSJFQps0#R@HQ3Ar(8ߠ1P;ekK/O]HMFUFY(flOZ^5e>[#J oQ?At^8sղi]͹"Rq4|?V^2 dE#Բ2-6Ǻpb{q8Mbdw}L>UNY1d6YMSM9fi1o>gei Bͩm j!;hF,N d/;sZ0dAD.3Idu`6 @rodeZC)Me#Zh!J`ӑ,P0ua`CrL{C$I^ާ=kްpYTe̎lJ]Q}3h81QC Gc]oʞA 4~1Ӛpn];7w3vWKtd'Stn͎7d{n{\$y&\DGT(TM=bP#:cѢݎZ[Sݦ\DGT+ɩZMSnc1hn"޷vͼRM!!_)MT==P^`LNx?kJ˔PH+حyÎ]kTBCը,QxE&?XhUr ~+|iw}=횟ݰ -X^I Б]2QI,]&*ȯDXaƂ QTabdO'6=В t6dAj0+LBCD*(D'90fv66ceJ08UXw?*yT!9pƃMuI QcQDTt *Df'A j|;g iaPXDY$>2W1 # @;L.3TC%d^{ϥۇl7]qn $|$/HxEGK}ê?ueʓ˭~d-3rwe[2x~>.̌oJIFpg˔^F )A]_߾Y)֩}/rn1<$z # 6N1N0&er:m}9}fP 1U dWTETqvfŊ9St¬kS:.߆_~tܻH#.%f%W?q +BgeKLJͰ(ΌgO?w>΃6GZKJף$Z#B?c0e"6ҷ7'`ˠri"{Edu;[&T``Q2C`'1,^=GP_K` ed,\1U,v]6{t0 JT+&"Do}A-?@?_q[<lJvКmd>+Q6_wHƿ}W7D/e"5+_j u|݋.Ys$ 0S&+H߸.^*!jB^ўWi\_]ݎG٢Qj8t'# P_ȿ$%?.*Z*&Ĩ Iu6fXU>fsC2%ngǭlSk+'[gz[_2;2/U /xf sÞ`؉}s&Y_ղ[mb/< #Ů*Dz^I𷋿=n>uMOAumk~LcT/6-x^|V{_|%oq%Yʿz>ą=<-)w껽Ԍ1Gw%Sg)S'0s .[N#8Ojg<7 f=)&jK#jNA;c4/tOq ͙4ڡN@0Bbb,.rD!) TnƐM o02 4 o1\v>{@l9+ СT%KO0!nja|I.1\j{5'P%T WA}l 5o.0< x>x85' w8.vgm܆^A]XҍEx2PK"WŸX.iZ^[1* +wHe(]^8EskbVlvS>ݚ] 7\:,dh`#":^ do]hĈ<:: F'_}AaWqaijoW&noOS}| .uXTϜ~(q<m|=3󛨫÷E4yg'T/{ȯ6;Ĩ7N)??^5 \Amt-m7:ϭ~ν=D$Ӊt޸thVVMT`MӜyCN(%HC]l_D \=0{|A|Je*xөzEOaD+Ű@S=rzO"ΗoV ;-w7XWdNL`17FK~tVhIRW*yPphb2 #-@:NšY M+2$WЌ:Yv]􈝥LdENXho6o~yUgoA'd܅%;S АL!(<Ha#@r ag l=GoQ}դZYsxtO<iwLJߢ՚mi@5Q kWJߢIЌ􄭔'>%6D#RؑZQu+aO++11*Ia乡a74+%\ᩕ@iؠ4ic-uȹEi* λӞ[Y)cӰKloQ}ՠZ0VzVRrf%Ռ/=m+f #8)S(dT3λӶҰLRT劯QX[T_5V$l'lA_"Agl 1#QQ+y G7մv:?P0Vp]5Vƪ"Ɗ%4VҨH%X1diXDΥk ѧؙ.̉Z-鬠r7*c -Zxf֛mDc[K7ٌR3<ό) [oRFzhbJI,o?\QmT+jsQi:,(갠.m\Y)J(-TP}i[)$~N~[~R3O%fWg*hx؃&u,Q]`㨨TVJ#c$́)v"O_.1g<>5$Y;K/B l@!lP>%k%$#<7>affʷf!zL,}DexDeT`&~xk6<)|O;]݃ >t9@a~k~fee(Lj.-O{cyo"gξMзEp؛{~1A+x= Z-qR~( Io+_Vev:Sƭ,|Xc8Vк֎s_-1A0| 5񕶉޷Ugg])xgA.]`r=y \}?V/~KIKFxLZeoWs_yt%G-鰭:*F#9ȹ)dm^)Qx(k_Qh5Ё;jLKZ6'f7ĺԚƿԝrנ֡t Eփ6+! )0Jismbp"ݵ(pnknsՅ6K{d}FO֓ AόRt\:F |Bk GssZp$L[@LKh+!SޅؕN)v<bޠHJ 0vS8>La9[Z,!7U}=OH7yx^+k=(%`v~9ղ&x8_a<Wz8,ѠE/m2MCV ݢK"Jn:Pt<*mwE[2#32ߤ\9 xT?}zf[XAz$ڳ"%Txv=ݘ|Xޕ~qKDx ~{yiqϔ,FNNr0){9v֋!R'K&OG,N>"r } b)12 A-pDH€jJpkF=gW) 빸$U  kmlc]ʡ59"&"U%)#V:9>_33˜g̫}{*{Z.=93jO&v6_T O~|T;ד~직p13+7ƃiT(j<9eȁpE䡒O>"c ahpx*9b,ͷV8R>&^=C~}q/&w[xY!1bY!̛f1-fiIanYgl/j2TޟBIzØt|jOמSKF X쯓҅bK|3*TfyگddFso0Ul6jaH66t74vQՙ#Il5qD-i'j:\m\;& SSMR IG%9/}BG)гzShryЗ!c=$~XgL×/9) «—i*WbU^35ݭi/bnze(Q\GE5>d9z'@ lfOx/wōԸu_k(3v4 : ɞxU8Q8` 4"+ /\{)Brfz@ya؛j(D^yK)7N ?ι0ԹRu *.RyƼ҂ʊ`]-RxDz0^ɖ%ʶJLו6W p=蔻ZYۃ캝wv}[nj- @`!"Q9%E QTȆ;ZW1k(Gi#G3lo.` V 0WTðD# ~>0q:jem;dS㨶ԙ$ǃ 'C/]2FQG]sQQ8pۇ& !B|n9VD7Gxo7s$[BjOWCt8nfW%L"ؖ׺?&֙Px qo_}uR,҈w]24,RV?>|*Ϊ&b2Nj /?˟>A/ 28_@BشnL4 "UWq\l|=waxS;:Iy*d] Yg&aR=K fȭ0&l%7ٖwjHWhIA=9:OǧEIUHԄsEk](y켒4"8=wgE;hE+<4>IE?s{1:k| obv뺢5Y.-$`c U A{-G_`zoMu-\뵫$fl8aRSm %ވ oMYbيA(,NY='O-A8j.~S]Xa8lp$\Y AJ'zeׅ޹9[ Q! D Hnu"GoGCqy@F.Q& `.Q.B\9"du &2!Gf@L]ROY*fwhY#x q4.ha<=JSfCs8` 3!0k=-0 #2Dщ)j7] ddO?e7مnnZ*HM_`fp)w{%dHeaZd`\Xt!l;1eWSFOTG1E> -}-3өki+PL9mdmK-PV,H@j'p)PB4MtXr-*(EsBuNaU!,W_b`^P;u|p^|{>,LLn*9Sգ33y5:T x4~Cs[|S<\[VKwRb_}O9Im@Ւ[ՑJ83+:x/AU~%X+4B[%uEj~e#j/f .*JŁw|=N,u{cdc %8eYFdR0rH,~a?åm6"`DEr83M3F'cNUPIMN$w#KCqH@@od]zgIJ(%TQeQM3' ,9úQ5P408hv DRUFR-hЀx!?R ֕ubcwh- Ws-U]5e%*rG |d 9rv&epXM&:ZxBZK|o'5u+07G说5FP\a%Jkb;^r\1>-w *ђaoPfɂ*>j-nYew*>^2i6AbLTIG@64ӻ!rԑZJKOZN[ uf29(o39raLxtLZ^=ZPQB-j9I!l,fG:%`89z)͐퓘h GE͞]tS?},3p36Z*<</G&~6CW ]3tU e@Ds G>W fX 9,PT^TH F?a/~-l]`hn uբ4.lJoinseCK+ӏ.,@eBA$_?غWZp6m%~?~մ kwޓvk lSϳX@Ҟ6~X`8ϗ#TKoBc'6Y q0Y,g!C- ݵku]Çjڷ>UR8KRăZ;]d]$zzr'c"jAW5*V}LƤՆd&Nu7C!a:8TS&O{dYT; T%x3o rXRҖP!;P%  OUj6/$'l֩ 9FoQ)R5WBIrak$ mƌ+_@dd<u;K~Eïb~,ቐ='_[`F^8k3RB'wHp(',)`#-,**QH[h%*[XPliJgTuW?ȚRUyWŷ*9G:9d b w$ - U:Cߕγq5J,k2Xib! zqh~Z#3Ӥ;WlaENǹ*nNs5O0!YAѣ4=hzxkxXzk:}zzr'ck?ɉ{6͔!hIhaR o}/y j Z {[2%D'o}{m)\n߽*sީbfџ~f:9itQZpB r\Զ.B1`. vJJ0['5Xy3G=m2uW>vQEҐ\EstX{uGary:Hn3XHɬ[~H6|*S#4; 4uA#źuRg2>Һ !_)Mx0D)#Ȧ =9T-b$OK(q7qIOՅƌp.4iRJxqa&3¡G˜*Trj ڈs+Sl%LVp) C) Xa >|qw m7v"\ R|3!&®HT$%↥CΔǔFfQz[,ytzL>OaHF$_JYBzN 3 f8nyr`(sJ9X@3O?|iiUB>[x9$`b4PTļh,@,jat2+nO]Avurb'5)һ$J8 dO\Dh>ͿwC||2JPEB{k$pȳUKgGqL=۫D٘\T,eg] }Jb9.%Ġy7Ir)1 v5N`$>F.$Kᦩ %*%ʲEeQMYT9?/VkX a",s#IpRUBRI~l j`pbMm@,;lH6(!zld)*+]uH E]HgJ),;=G d4Ӗ"[^ v.=F-/ F6dF!X%ե,Kq*UjŲit MB26ljU9lvBϺ%M= Ue a~_W|ޗZ_ .% K篨|ۺCO C]?F~-Z }v`L/|͟^ֆb?zb8$$,%K xuq<| ?^%,{Nm'RgW&A?>G$TG.՗;TND.wzr5D噇Jr*ZXiuH3WU{ٔ\"TH)k`HnMrNKևois P]Y''*2Yṩ#V?L 3Q:'߷NeP|adt6ZRL:5f>n)BO_=!Zf ӷ,?Պƣ- J.bsߑ30򿔨&ߣ#qА\Et'ٵnsѺ Furߑbf<XAY۽H6|*Sc=qX7FT.bT')mF/6c[H6|*_ JϐS/^rK?B6[A;>}+FS9z4wKYG=3!I똰s<&4Z'=ra;Q)^Ka6cЫx_E&*iM 4f~R9y^*+ :_uX Z$#x#%%*8)h+io+diZAUUJqE"J/wmmy9.0HBV,Kv.%Y9n[e0yKUU_*wضf@v:E!x /7Sw-X2alm1m"O$ͦ! Z救,;M;G%#W/J"_f!jԉ`^珫vNb:Yt3F[@+ʒҞ0SJ:7Ow? @d}s1&'nuϙJ0-gra^M}\Y/}:e^V;2M4rv`yjkYNb:Xsg6L˫2 91b&A^9㾒,' KǁkG ZF &rA#n#.Kz$I]w^hobU.R_@9[ZMe Fk zp0+j۱N @Khj#+#w\ [,s' P Hd3$qLD+44W>X-JX%A_>2Njg3 vNyhJ| A-E>ތ^<],G+5OkWF8R,yv:߬o F=(}h5JXa= ULK Af2iueBK#bJ?*fH@M/A1XES9Q|>"Q4V 3z<,5>E1-2$qhȂB4/+JRU'!.? cU搁en[$.L'*Zͱ}6F LluWft\s6$` ϴTn$97d{=Hm /~3om}tV=D6bkVLg-hYg6A$I7.uw(k!q*m<-@ ʜBJ CI6EJPsB@Fx^͗OEj8QSLFaWY2=Oi7oCN͝!!ЏT&tb^8򲌹3Y(_4&BQ`TJȡrPGJJ)cx1 #5I0zGVE.u)p` 'S4 y`{z@w"e]eB$GXy'QECa~Ӿ$iJ@D]F 0c۶mБ>%xtqnreeϘ M}C#YgAGn'N/\0(LAj ѶL,dK+'A΂'=bLf g{h }%ڪ]MQB,.Ozx@j rr}9&pQ6«q's4ΠG'@O9LQ/ ii92pr5)3!1t)16Zs}b-Mي4Rp`0fD057Wlw /@U#?5@IV nݿ hN?m>}t#jw*\;DY:ʼn0Ӳ AgƋ͘H4#CSu992'xІ RX8s4$1sqO5,Q_\6ĔiiuO٬m/eTH JB>pblrw8gOw~bӻb/R(&3Poeu2F4#ؓg~P<[ufīPG@tR V'mJ}Ia>4kQz{HRfѳ% 5EnMWlrxl]߫ҍ I ONNb,Gs1]('Oaf R[!-yzD{@͟Wsw07[G\ROlIڐ&J4&(-91Tt <)wvu=q|6ؗq3V2'1vr\l긯d4,t>Tzs~˼ͤm\X%g]:܈P9羲J 94mu]e cȀfb@sQ,#GlJk.S }b׳\Ԣzy){V Ra4`<普xHH7jY뀷(Bޱ,,ArȬl6ē;`*ݥp{?$G0f>SDVQ@xЊw~:I1Շ㔀F3j"$>xNlU ύ OԲ2sij'm,j^r^EpWhbO⶞Z]Z??ŲQ%vc!G_w|ogI^jr9YgXFToR }4Jy jo+Rʞ=2 #C8B/?kE>ARߔvGre.4])W4W۝^gS_l CM>f/z,Q\܋[KZόWyX.^_C N!eb/ea%$}%F_ XYxprR>=Ԗ8r+MK$s/\3˚Y8/4buuY?I h#ult")gf]dO'do-t\ܵ"fT$Ǝ#q] 1x?tS(_ao#ew(@t\w<"=1&SWw l6o7b2oV1^N >3C$+ob##X3\u@KƔxTL`8Lr$~To^Vnz.eL Qx=S>1G58xFuD5W| =yJMv+*E.Ω\␊ iDj͠鮽 C @ըC(`kYoSSa1tQjTfjŰ_b\{\k_b4:]a{$P<8l] N%'&ٜbB6.Øy d`-yX,w^֔9}.R?zhA@ؕN4wDщ؞PB$7&4.Qٻ޶zW캠 9 見EwpkؖIPcّ#H$33 9hڍ&';+ {N̾Fb&)d(tzUt*v(ƀ:؊:v+ي~[K5jB`o!*&jŽhM6!(gmw\Z)][ەVMǘ /P= u@:nE^N`:)S֋jfcnmJ TMQ6'Y lzK~uŸx;K̫> , J<G oI.o[U^9 .WYRht]/ 2b}`(Ѣ_Q͕擹vl*Af ]29yy5~ _U\o=p.NJNBk7YkWAS$`G6wչ~vr|T;xLc97)JR@)Lrz(V);ަM bΊW?!v%ad, j~KqYZugs2zq~ظa%jh=vt[U Cf8ݳGť˝لX󖚧kOD6Ii<%6v|U=پ cϾ\tZ(?"P:Kn7؄'cs$H]a7GdϡO,7Km_gD+qŮX)c} yl"*JT ሀ/ q8q"lD]F{2@ iD=⌛HmXo+my-Q^L懝θI{ɮP)ld E$ɏb= wqui21b"AηAnE^>7s49e;dͬkQ,COAf'(d*EXJ &~判Y{}K l:VbEAe)H)8xc\fxn9&oj|V{nbk`xejr*$m;od`+_9U.4GeWA&S%X (Z~VfbSWW~7 "T80umeuR8Y %{͝IxZuꩦwW qlk:Yl_$9 ócaM/D( YgM9SR, ?j5Cibȹjizz]eIaҒE;ldWߦ?;#JW+3D(89><־ =.05W4vVe I ^ra0{e5` ؍(xĪQgӮP[FٺCB`XAJ(r /}x /^g3.[[k{~xblg)TkDqR`Լ7ٕ̾z#eY_!#,Lw]jD02*yjeue#`:Nƚ-ͰW ZI8.l/7.-oދɚc=[jSC*Փ7?.ʴ矖Y1QurpH0ఄ??ɧE:t)_/xR5q!Y=-y^õ6޵;)}!oa 6ܦS `>qԐߪmy5.'̉g>?k{pd%N< :b|QN݂vKᱻrdz@Iݱi}@y|9L>V쀱uZ1Yg3qLsqzr8.kJeMl ;`9 Iǁ.t.ڧWl0;=bWXJ6;`ioIx]R%hz=y;2xeI;Wߟ-(?hMH]gi3]q2/1kpV%{ yz-~;OKV6m,m\)ބ4劅ӧпSU FRT x;+v|C͵h} ŎWA ~b7w|_Tlζ<7µxXQO i\k*6G%Jq^ɹûCa#nV}dyAՒ~??"$jQQ WuU,؃cmzBWIF1nnMUs9en :uuR9SaKuݐsJֿF@<꾸VjT:d[hN\A ^ +0k8(ig+HED ]v5 *: М&xv n>@v ˅jT #%BԶQE*bMTw7tIZz"f 7D cZ5Sm;tSJ俐|ΨKNҚ6"e=cd~tDmʗk,rKȒ4}El|~hCƁZ+ZPfZl"7g' +r+K;z"&G٦yXG٦ГЊ5mO^>:;`2w[ n'q ׯ5C[';a_^I|>:/k{tԧܺ<\?O5zoi Gg7*γ.FIY9PiŲ$õs1W7mW̓ZԮvJىvq^bDI=?#qNz0`"̰dji1 &yruy 1\N|4;`MV I) V *o.t1՚*~qL5W~ǔ$fu | EًE|ǛgUՔ^?kmda.Bv""K];  ˸LHX޴L=.drCT4eTh%89h{OYvKZ0TT "qY-!N/]~omSfJxu: $ۢ1}-IJٳi ie!MHek9q;iQ߬}&4jW߲hYK9lh[3˝V;e_5u7 ޣN Tn4w(4A'̚{GY0+_V[PJ*q+GOjg'hby/~ue`{1fAŜ `]QT=XcĚ@zW>-q (2If_mv ~l&W'9} AZ) =#R'Jy/];BkDӄejb_ZX#-t69sAdɺAl+5Gj&WA X0qsp 60mtc11hG݋JJtv$~gW0 5c(a3DyIp>XŌ]dv,R2}:_Qer _@8)'&Y*8VL<RR$_%DD5K gi bLƋ40EHf4h*`d("d`66՗),gPJJRs>'. e-]˰*,M  Tw2cAco +FSLH̝rV16P&Ɵމ߂erZ 拋'jU%t$ɳ[pW:c':_օ&)11w` |gYhK }P-+,q)V?iQ6)Twҡj˖tҡ.;Y|C(GE7}hR 3*LC=c4@(Mc C2hO bh-]%Q]na ˳  FQl3e`|n@@y) +w  j\H4Iۺ>a!ZL$'LNsխ9X19mKVyWeOA),XүXK2{2;m>?"yf\$OG>pd*im[;nzmmxhۻW?4ٖp 3p+Ճsq7 (.(}Q=Z v vΛ" Pcx WII:onug8 ˏu6w˓g6Ng~9˫,9𻽼pu靛bz+T׫.FqԾnދpwˏ{,>Ol}S~j_] %#7q}g\R._%\7;g.U2E?Tv~ڭ9S6Ky`ڭ~@vBBVT2IP-ڍ tڭ9S6<h8V?*Su!!\D+dJG'(PWƾ%3#o؞[xkU}{IP-FG m.P1>K31E1RaˍSŔ)]f/_SPvp)KGu |3p>@i@6<|al?g:3W&:qmOO>Y֝UЄ'o5x 3uq ^F{j/U=TSYba?<%tomLg1? `kNV.`I~6_.C") 5K0]p={IfX떙aC #lj'VzQ; ?AзW?:#0dL$ 1ʷ8lJ)+xl"G&JՂ @ \PK4qLk&g^`%cb饘 Duw_?jcIWj@ZfG0Fi'|wnc1)ڛ6Pyo59ŝѷخŮ9̎9<%Ҳ @7дtܕtC3vN;^DdAqYf[U`o0h^T,ټg=,RmV)+룆T~"@~̉"?cmĥ#m2:bVݧeJBzS'H HLb9ni) IR&KS )Vp"QEůR!ڀS=2^LՊ6#k{hFKc3^צG b*;ksQ;QB>9o@YIH ynQj٭ /(z/1.ji_@/L__\ܻ)~#gQ>Ϣ|E'A"nψOK.@/\ש7/z1Y&0Ķ.0a} 빫~y{fcE;w7Q>=j$U4VsO`I9ssqU烹zVZI(V~J3TkNZfofKnNVu(ʡf3.v i$^:#<wQ!*4K V|_o! LEiƷHQBUUB>x!wrF6?|yrqˍ$\Zaՠ$bj7ʩ[ ࣥϻ +uù\k# w6FL0 y 46xA{G'2ZK8;[p׸LڲVS yїpe LѾEK?s M^HfG/syW +ū앷>,> lg=YR^}now櫹Y?c1p 7rX얲/w r}G·o،۹p,ZE$]έaHgQѶ`NERҲ -4D3IAV~l4qaMx6P&lʯ·tjJ?*碰i[; %|$#\q[_s&B'_mt[1SjMݬMݜŹ2V`t)\xA7KIIqvNV2=$IbtXޗU@o@ڬfloĿ= Jn[y&|AfЀ3ZN0jj*Ô5 `du}JLgh%FElxQm]Ţ=Vؕ]Gg1jǨ"{4)T=Wqry8]ZApy'Tks.PLCt p#-vZFĤI [dT>*JHP_*BRJC,23@j옟Ȭ;T{;7~bIwOxҝœ,=6Nb~T`)Nƹ(4F殑{64q6EuU'/ߒ)#$*džp8 ylHxGCX"z4`Dp TQ ]RlM vI4\?M/ ܋ddu5bQ̀鉚iマ ovr 2W3v9|LH1&`/p;#8GF(JX`bAeXcW =KgVb4Tb"Ոe$ ]Mw.F$dF%U=⌀F`~#׍_ &:fc1gVi J܂H1cy5dZ͑KYl~=XJ7;&Yꉻ^EOͣ^>,'F=]vJN_,;Q FtU \ ?).iFg&? ֌ys,hIGHwߞ||g&q jɅY\ U< I8GgZ*嬗k'xVDl"CR8{~蚼ʐBnu~e8F4{d.>7k: ߕkz" 7k`ihӘuA} 5A=}3xu$s>cR;Rt˰o'c6xa)wls e:W~5غڐ?r&b8bzV^D\iFͻ- ! 7urdII%HZާno݊uWI.Z՝*b+ 8 :r{\`%QI]A22"GӞkP0{%'`oX+,ƺU;ݪ@S-8KL7p?V BD|4:ݣePⶍj/C-bjˋ^L,"N\k5$I9xzj89kQ2}< ڭ y"%S $.2Uǥ@yJrffS2c llӇ>݆j٦O:۔e}&fL!e jRg$\F~ BJyfjХ(OXJ uuqh{{L)ǔRA% !iF93e<4[RzU1C} 5XyRJpRTŌ)A} 5K)yRJc>p r'-LI)3IWpz]}'V޳6r#W 9LdqHwal6=r$y6Cl^)#dG-vX,s5#uP?lZx`4G:klp0{NwnCz圎Q’JwCdرK ;KdxvKN OAWB[9>cnp4yO|y2~.OL͢_aa!{ ˫ʧ˸gMf0dc2Ϻ\j߼]"GwǎD&$\v,[i Z"*#)?^;k 4s2Cxfr-\+^:^v㩶w|Bޔ7%SpIUǏeLIe!@4M̋V^;鬼8LɎg\wE-%FAuE?sFPR Z>}k5#0.;,'@Z")]E+3ρy*᷼❣.(ePK( ZYȷ@DgU\paAK*Kl+ ޓ]@&yтHXI;d5?l@m9p~!FobAy]pq i"'`+U_KSXs)jc!bXØ^;`u[k mo|]y :ug?~gks=Ii`DfV0R#0 X?3yAm ߁2#-Q - ?d-^j:7LW<\8gXpŸ#iAsY.9ϴ6 [,<|q2+QbW.ɯ:Z_f$~x11#ƌͣMlZy/ϒ ;%~B3y}\\I}l6 SZ13\|0I*WF~riۜiUYZ2&'-J\d܉IEF~|ycFpӫ.ʫ>=@rK$KyЊ9#GLU*KutZn4juC¶2؆ ɈaBDs4bOt#v ^HnKHZKkkj&5ihSύ8r]y e;엇ܬjBN?haWbsS>t= 9)>}m_4tOWsT+AWf! s8G_jxY Dfs2\f>H0e*x%l5" z+[0j ЙGx^/8o)HㄊIMXuw[Lx- A&S܅ݩuJ\S)d-!,aU)%zH4+c$bPu?CI+O9iJ7ۅyV}M B`JdH@A)t6R.7Jm^"1ҌiBԡFHo6DaX qzevG0rę@& 4҂Ѡ멑VL yҳxDn[S171& +16㞒Z'۶nǸ$NSN\D"QTЃjfvDTԈ@29ؤҮmoyIεrɩ͉<9urfj\Lz8urN&(pA7K#Qh‰%BߋXuK_/&trËU` kN8vq} 3JS)+lqJCSkQ^:+3ϾS-*b_,*R暝*v^@JurdNx.NPa UQN[ RL֫qj$&i{( ];-z_HˏN߆UuՓKvz .i.Q_Ňsw1KGURaSu-Uea\<=FGIa9jFLعrP=QSӉF;d5B ; 5 T2r((ol%_\|xnnK,7:h^ZkX+Jf}ԇ 3dFT2c׎[PK;]*]ka ~Ns`YrtMW%fV;O Ix0]2zn0Ȉc혔ď]-3NpCC59o QG$_TC4y4s-Y2N׆Ӛa%f&ʝž5 yΙ *K"eϵ+ fsRK H 2*e;]a([Q$rmG\6q͋@%9YȕTJ.(| +ulNʗbUG%i3L}Ƈ ס\[uHBPՑ$ 4|R h(v.#I(P=q6Xl#8Q*9:% 4DӃ`t(P\Wt(0MA*4^Et@?*y%:Au%o=,H]퐨(d-1t5md/'g,,yAO1' gE.9*Q]5ď]~S fu;o0# HAP}+ DZeHCŁw6¯VG4בבבU&W`{猇6/ AAƕPļ&k:&kX廋zxuLxg(sRFĖ5 oI&~\붱F*fC$v]@]!!pܥyF_KYaG] IMҤou[ee0%P>!xIйv6{R֔$!AR,ı֔99aGjj8i;qG&L2÷ޑӂ1&͑d=5kMY ,Nc$`lDd1MJ4K;#IhV=YOTI4:SgYOͷwrew;]e=si*iP5ޛa9m${j jȓTBS3oJA(XLeu 8BuϞSdRE)%BJ_:B'-KF\+R,4.@}CgB o pLAvG(  ܭC Pٛ~µ ={_M|ѮS6E|9҉+Ivs]ӷ'ѷ&Z6,Wι>(*WXSNhH!;N^W%= 9vva5 +Ǯ 3e2g Hc0(˳ʺO^]Z)ۿ^e(0E+l3u7WSx"%cAo5 GPhcgD_g}2ZW=ΈYSFq yHeE3G4j/䘈_e"rQʌvB q[(22X ^dr(*kwK%Q]6Kz(ז fR" 7S{!/@X؈B. 8sJ" =0 [͹|:z57AE'R) C ++Qi ^?0ޠ2E /%|HEqu8FknVFF\L" ͈å|{jC9kRw`QTB#sH<]GRydЊE%J(|n=>\ps_g>-YܟS?LpZ>}Q<.Lt _oJca( 𧧏/UBfG 1Ȫ9ˠP&ْ?<~x~f3`GopCpyqg?/+H@bJzG$O\^aضQt>y'n[TI&Hhfs&yPǑG::֠&&yi.ݰ÷H+d2A5M_vqѭ&/ 䏟?z{7u`:JDNHt6p`[J=\Q UH#Mdt^>+kl=?g嗛5{C3"GT2 N_Z-e8J:lލ|K4T˲ ,McXx+:hx О-}FvlTP- 4/Nr#k~yY~g5#hųԣ+$T/hK)CQj`M "c z^ (|It.cPl D u*Oð$-DeC4*NJtD {=)ú.n0Y&TQ22k 7]>e2T-hk:/]!-Y"EB.nwtg>+Ux,PjkSk#^Z~EUc.A.2RTjN_1%P8QXѾ,eK$S`Uu1N`7yl iU$8-EBֶksJZuLhqirтfkKjO FOhU)09l!$2D)!m%z(+!|mYh?SpVzZVreDdJbNTv*'@Bvlnn'ѥ52X2wѓlO*!V_gdT(ٱ- jxlJ+cmqB16G! WoKh[ fDR%4Ϟ.'+ :ϫŗr#$J1Y:ͩ)W40d\|ZmŁqH et˜6>BH uxY{Fn6:T17*BC ݰ响"nC}S"iWeP64?rdt(D >K||]Kjx}G<hlXrSشSjkqON@U76?~sF™,`?vN:iTHkhPBqZs8)SX].R1qÚ\ u_ũdyؗbT`At7.b E)Lq&?0 9S|B'h5pQt12u~yWѭs}#ﮜSJJZ4okҿ̗O+~,Nb)yDsiT%SO+)"UׅZ&_(*W=0p?d%o>|?CԌd vQb1 jRFh zlߔ*U+m&Zԭ-z7RS˻i2&K D3&Dj)Z% e`~=-"TP }u?)eֱ5nNf\=;egY+U:Iyռ׶ Xǻ)4lƎQP\#19b gMPI]mGg=ݢ *xW_j]4!k4ZC4\L/>T>y|\=fv-[3o轍JB3DJ Q Lk!ª-=?A  ܥgpz#HAi"ܗ}6jcoK'Q!*k%v=*-{uoy]b^}wn?,lt,R4d0 O"DE:]Vhu7EGސ9q/k2UYG(:%hͲCvZX6 D: p.&&< _erѲM~8,z(^,#GS-}z JT){}c~o )͵-x;M18/ߡ\a E?h%.CPڠuRDF |s4tƻ;c€L;ATg ?}s4^_hak{T0[K-\wk do+CPD9wjp2 ]Auz T,+ /x^FSq3>m=J%S c!?{78DpcjG?iIP4sحf/>䨽kWX5g{$7}b8Iaɰl]Qj\һO6DZkË-Tyϣeo_nxo਒K~`ڹ<\ݔUGjnn 84/bbC߾gԯ CLgd#XKߥO}ӆ8U74bm vzEMǀ_Z9C X7I+U&7BPq<.ˤZx%-uŨMd[=_k&ǀ e-t>[oW,r_:!'uŅ\UD2VKJp&-9uQgR:IZפ"–Z:0OO&t̜;)j.u܈Q;%K2MTtM_|>N:WMny4ԒTI}S1^@#cn =p:m؋'F*V(@A5Ψ{H)'m'i6M-9׌YQQ@1&Yt W*9>RR.9l4¿'f/XZ[i#Lln7C­64dx~$͸dfk4;o0;4Rɀ%[`a='2ȨpRf.EЪ%Eշ\FJZbV>$S!e;dlQʠg~J5Jm PjjHC3A>uubոԌ0-a,rH++PBVUl=]Q~E[>2hB9[H]@tk!fF%jh r9dR|CnˉG59P8qyJ<ƙ{9얪)n4' /lbs?]nI/AJyY vg1yA!-IJ="%hJ,$YvnQd1ODdĉDB Z m6ϯ#9dV Iu(~et>H60oc\[0H~g ]2u\.\ę@oz^14K?Z\-X{^fjJ=P}@{:o8%3e?G lPz=`PI5%=Iu+Vu+8 ˖RB:'9({o%ּ0gXYoj"{FAEI/QF 'g+}b.-󆾷X;EVo~W?t';IhKPXv[fw?[fhy66׽|ߓkqs0#1Mj8}ۇFQF矛V \2a;_{rҴߕ=e4'T8FwΓd/ *E7bCՑTCϻ@{PW þ;)CDZ:è5Q>Rw7F`CJPVGT@:'ʨuAqBfQ^Ъz_`ݥdY|L%bh{j 1-*24R ]eo7TMR;n˛iZ/a^,OVJ.RGՍ.󴼘 "bu0A{IOҲVdOh'Go|[9[W1߱w{Fv:moڻu?һa!?vq~ p2h;~^ܭ}ܭ M6K>NxQ|5|L+-5V\C5<x(ϵj8>nPܪP5g gN'ePV$k#Ҭ8a`^&AI- O" 2gg,mzh0a`]av1,L?f5U % Hߵ=K%aatC(mpSЎˎ>_HG5gfGOw/|0 N&Ӄ0XQBaYD,H9qJE}.Ch+`Sd&a{O<\V>(BIiELHȤ4T_`z4O[>٨ D}\ʓRrH$!(bKI{3}O`ا3H%dORrUZ*t e}`YjP->CͧQxXc,mE9iNG  %%Q ,jrdXqQfԶxt+U\TݑѺVNcUv7O@[ Oz[5El9,}Y,YoÜ5jRS!єPS';Q3XM+`CD&*r52b60͔%!q)@4Y$Ys7wp22mZ& > |k"%LB;4 aF2?<2LȤ&y9J17*))LDd.wL>T$] Ɏ"J_cDA(3, A{/etW7`;T1!`ť g88BVޕ 1 &>a V:H y܊޴DorU厼Sm3HҘ*s<HN,Ϣ4g ǕݑG^N4N ƒg(E$9AcNg 棍]K*=Z[_| u=&\8-/kQbg̝'e=UcD_L*n )Z1փ&H(x47zpl/> !j,kL@He`w\=d2;X_qq/,J3*o2=@sI@ýؾb'ݸzқ|5<ȅ8C AԞϴyL`@-N^/@@.&'\%U7dc() Fٞ| g 8&zGx7t~VԂC-!oz|.湘2R+3<,׊F)Rq<D@?+z^XĒK+u+"')0g4X­f^B4=6>%xtâ,'(ۜ=94Wy \!Jc{F 4{IL&Ĭx+ֵ^4!IxHW enM22wvz}oوX@ W$G2*6V+>,kg _A[aeATBe%.ga4WS,1m~!e ƞ,,jtU52³dpunIJAeKĸAIvLsUO\}4],$,昿/)|VK\1ߌIZ7YRqc;{.sEZRդ?O競0n0b? u> Lב9`̝!9:"3#qZ|B :HZSDOL&~ tڃ${\$Rga[ONN^fa\Jd)K1M4v5e'8ZCZ<9=ˬZX.v얰[(ywi`n]Wo>w#0e5\L",#Zxukti-uלʦ"ܢv&-5zs;m< ~:mv}tN4xw~M*)O㠯f5INXA3([&|D 6F=q]̹6_UCgmC,9e9g)k! Ѩ-#^Kyf|ˮkB[mvu&af9uW׻οQzjhCmg_ӭ?VG\فt3ko*1|WBUmCZ>޲zS Y=?`b;'I7 >TI%ic ~ƲF'׳~ B ]3c {1i+B^.u&F3\L2e.YV"X}χڶ4wbtu7'IFB.mfy6;Ռb/UJ;eR@(R'ETe nŗ:>Md7)DPo=z3n,K{;ڣ^Zڢ4}} uE -I]ii_muhhRT6 PkA{A6l *b: 02LpAưl`)tleFbJ3 W%є,.KaSS敌\2w'ۖ]9YSn#Ǯe؋YUfWG$cFdyfZjMRG۟:7!_`DiFI9*Q_ օ, L,hҴr[J@B2Y3SqF}@ PHN\$dbeGj]RhYrgEgR8o,Ynr$e\;fZz+|mKQ^! Iy.k+BQԬ+/#IB/YLQwc;0vnG_ZSطlc7U?ó+;U`!(ط̺pA6oz2#*JT2*iktTWs& Ѧ58s V*[->|e-v~Dƺ|Z'ҝ'L4VORe ?#h-&`:㧕7^Zk\=Ʒv lJ$l>.}(pQ6|߀ Jumn" ]LB&DV/3q9`gCdRﯣwj9^S+eTN<>h*&UDM^v珞ǖ|e5T̫>AaVnP*6Ã#F;7/r8({h;šyhXCâKC]Gvt4 16[Gg6s{Nkk0':T$Ǿ;yz5Ê׀,E՝3#`BuxJœ2fgQdHeEA(NwݵtZe &c(D: Ӑs"yNBKW3|D&0SeSG[i .1B`Wy}MwhF$Lj*^F_[ n}oui4z!y@g6"sBGJ-rdEbVd2VJ!ʅϗي'r+ڨ;Y\bހ5хT3ހK!y[}E(٧+b4BvhZvb=y bx|5*? I$W'+C R!U|ez~ w!;eٚ[%H ^tbثVpJerono5~uO/9IIbTWE=KheR8RL_sïLq}vz {݉ma}/@HNV-͘n-:r>AVn~ OK$hkߔq(aD$KAar?V8RHA-|ݝr/ G1\مէ/DJxE+8eZ'mDmP jwlbT.By`}Qũj¬1Df\¬aU˴T̜*EY5c{gi%ˠ@rʡLeq2T2$廗Dϊf5+_\[:䪵! ^\ K} e k9S"iZ0Y*}&f3%x<_шZN})J=Q/kXPbHOfk8?7,)yʔm1,\ l=׵q&&̚YkcxJϖd|-k2/ߖuC}FXT;tƽCA; dN* 6᭍0S !IE`T׍ `6L%S g1|tߔ(a` Jq7+F~r?dEtW#E_PPk'# +u7r T{=hEర~1r(ܹ WŞ<ޏ/ڇ (n|\辿zr @\kT[KϽd}6w[?_^/ẻ)'gˁykݜ" #>2HoG ׊BY 'P_ad%eE iPR Oy6x!<x}(hL(#^8/1T2pfTҵ9ͨAq(Cj F.b$ H48.HTB󺠼b^ٺuA eEQVF㲢m +Pe'RJӍs3ֵiD>EkSjΫ>(hvFMOe EGYHϺGLYNV}31nѠ;?oϸq,An !7l Ӻ7`p^ Z@.'r)BMm3r8y'=dW8 mj>ȤMLG /i$-A)ephv7JW\dM3-5t?sq4 1 `Wp KK|/pxr+Z)jjShtSh_zBՋO415roC5R_I_iNڝj J6!VM`J+#>L|]2$'aB%P= ?N_Uhxӛ?AzxGѝ쥠tp!!=a6q>2wٵJ`EgP1μT*}d!8r#=I+Ev}4R1^+mEJz [7WL QƑ A4aҔJ#"Mhĩ auY$ M;ʕ` :=u7sPF*6gfR߶T݆*h2 b!EV:ʃ$7`EueaO B'6oI4II#qPPv~#23Ɯőoncn܂9Cu=U6,x$wrxY0%b j ŘfdQ\lrŬ)f+PQƏ G}}> $IDfZ N8+8 q'ٹڼlJn"0Ёa,G Sa`93'Q9HƸ;Hf!&OS:K6$Efrz?-Ϩ>L)E|Rs\Ϩ\y ޳ tX rrY َ ,KX[3!C,9{rJi*f,\H() n Ψ ,JY`sMőoCe6sc8Tg bD]IB Ꙉ>k @%}ی j@xb`Kz ǝ8`ǠX >)7[c+&6bԧ2pfTMzVDmª0 q\BonxT]!{hD߃&'u,am*{1U^zWot{0V#X;l*,GKMXaZ|ī,a1v!&z{{N$DHO7v([Koe$L2\NDTV{w+-xY ^՞#K`.[X A[Z!HopS>Ƙl(A~t?Pq&uxXL-RKu_reIs)T+Z@j%5Fb+%_u@>Ԉ**2qRks%ҵ*ʹڴ܂ZqHH Zh0#u+nԤhFI'VNۆղQְ(0]s++mjJF*vΣ 4g>WyZfl<@M:@[Er V`PU<KA@JZGߟZPis8ngI8~?BDC$$nz:ѿKedB.iSQe#nW5"}S4~cп/Mb]':dHsP(QR÷o=uE_#v7ݓCb-p|C8I>f3#6A >kY]^j)b QΛ9C5SiL G<ݍZvEwc{bI1ӬJq5cё%Hzv+1C4Z _Gmnv+ 0\3eՍQ"_4kd@Iٱk^iޚ45]K!5(mO#f[RF*@~53*2"Q X\3-_$@pg<@7$Dpx0w!ۉGK &9+Qi~vЭ{g΃|_ٝ!_i}~{϶<#nC~rz\WC[b\')m iE|]dtC~r򩀒\3&?-SQ c7rwZ[m]tOe`fh;z~)AP2 {wnGe7wvIOWioW|wjí.J%/ZJ>_7=qqs] .B3寛O+FgCO=fsS^lꇋ-:s/7$ڵ}o*G&8ExVonx_.w 8TkyvM'mT6L֮-p筰to7}#\aH[@g phOS e|8|Tvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004770614215156615367017727 0ustar rootrootMar 18 20:06:25 crc systemd[1]: Starting Kubernetes Kubelet... Mar 18 20:06:25 crc restorecon[4647]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 20:06:25 crc restorecon[4647]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 20:06:26 crc restorecon[4647]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 18 20:06:26 crc restorecon[4647]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Mar 18 20:06:27 crc kubenswrapper[4950]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 18 20:06:27 crc kubenswrapper[4950]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 18 20:06:27 crc kubenswrapper[4950]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 18 20:06:27 crc kubenswrapper[4950]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 18 20:06:27 crc kubenswrapper[4950]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 18 20:06:27 crc kubenswrapper[4950]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.172899 4950 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178154 4950 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178189 4950 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178200 4950 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178209 4950 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178220 4950 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178228 4950 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178239 4950 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178249 4950 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178257 4950 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178266 4950 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178273 4950 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178281 4950 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178290 4950 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178298 4950 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178306 4950 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178313 4950 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178321 4950 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178329 4950 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178337 4950 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178348 4950 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178358 4950 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178366 4950 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178375 4950 feature_gate.go:330] unrecognized feature gate: Example Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178384 4950 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178391 4950 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178399 4950 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178407 4950 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178445 4950 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178453 4950 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178461 4950 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178468 4950 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178476 4950 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178483 4950 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178491 4950 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178500 4950 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178508 4950 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178516 4950 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178524 4950 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178531 4950 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178542 4950 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178550 4950 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178557 4950 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178566 4950 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178573 4950 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178581 4950 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178588 4950 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178596 4950 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178604 4950 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178612 4950 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178620 4950 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178628 4950 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178636 4950 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178658 4950 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178669 4950 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178680 4950 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178688 4950 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178697 4950 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178708 4950 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178718 4950 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178727 4950 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178735 4950 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178746 4950 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178755 4950 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178764 4950 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178772 4950 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178780 4950 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178787 4950 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178808 4950 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178816 4950 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178823 4950 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.178831 4950 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.179761 4950 flags.go:64] FLAG: --address="0.0.0.0" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.179782 4950 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.179807 4950 flags.go:64] FLAG: --anonymous-auth="true" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.179819 4950 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.179833 4950 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.179844 4950 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.179857 4950 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.179868 4950 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.179877 4950 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.179887 4950 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.179896 4950 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.179905 4950 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.179914 4950 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.179923 4950 flags.go:64] FLAG: --cgroup-root="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.179932 4950 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.179941 4950 flags.go:64] FLAG: --client-ca-file="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.179950 4950 flags.go:64] FLAG: --cloud-config="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.179958 4950 flags.go:64] FLAG: --cloud-provider="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.179967 4950 flags.go:64] FLAG: --cluster-dns="[]" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.179982 4950 flags.go:64] FLAG: --cluster-domain="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.179991 4950 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180000 4950 flags.go:64] FLAG: --config-dir="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180009 4950 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180018 4950 flags.go:64] FLAG: --container-log-max-files="5" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180029 4950 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180038 4950 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180047 4950 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180056 4950 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180065 4950 flags.go:64] FLAG: --contention-profiling="false" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180074 4950 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180084 4950 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180093 4950 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180113 4950 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180124 4950 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180133 4950 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180143 4950 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180213 4950 flags.go:64] FLAG: --enable-load-reader="false" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180223 4950 flags.go:64] FLAG: --enable-server="true" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180232 4950 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180244 4950 flags.go:64] FLAG: --event-burst="100" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180253 4950 flags.go:64] FLAG: --event-qps="50" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180262 4950 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180271 4950 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180280 4950 flags.go:64] FLAG: --eviction-hard="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180290 4950 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180299 4950 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180309 4950 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180318 4950 flags.go:64] FLAG: --eviction-soft="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180327 4950 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180336 4950 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180345 4950 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180354 4950 flags.go:64] FLAG: --experimental-mounter-path="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180362 4950 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180371 4950 flags.go:64] FLAG: --fail-swap-on="true" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180380 4950 flags.go:64] FLAG: --feature-gates="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180390 4950 flags.go:64] FLAG: --file-check-frequency="20s" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180399 4950 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180408 4950 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180448 4950 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180458 4950 flags.go:64] FLAG: --healthz-port="10248" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180466 4950 flags.go:64] FLAG: --help="false" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180476 4950 flags.go:64] FLAG: --hostname-override="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180486 4950 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180496 4950 flags.go:64] FLAG: --http-check-frequency="20s" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180504 4950 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180515 4950 flags.go:64] FLAG: --image-credential-provider-config="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180524 4950 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180533 4950 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180543 4950 flags.go:64] FLAG: --image-service-endpoint="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180551 4950 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180560 4950 flags.go:64] FLAG: --kube-api-burst="100" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180569 4950 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180578 4950 flags.go:64] FLAG: --kube-api-qps="50" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180587 4950 flags.go:64] FLAG: --kube-reserved="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180596 4950 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180604 4950 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180613 4950 flags.go:64] FLAG: --kubelet-cgroups="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180622 4950 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180631 4950 flags.go:64] FLAG: --lock-file="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180640 4950 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180649 4950 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180658 4950 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180671 4950 flags.go:64] FLAG: --log-json-split-stream="false" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180680 4950 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180688 4950 flags.go:64] FLAG: --log-text-split-stream="false" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180697 4950 flags.go:64] FLAG: --logging-format="text" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180705 4950 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180715 4950 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180724 4950 flags.go:64] FLAG: --manifest-url="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180732 4950 flags.go:64] FLAG: --manifest-url-header="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180744 4950 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180753 4950 flags.go:64] FLAG: --max-open-files="1000000" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180763 4950 flags.go:64] FLAG: --max-pods="110" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180773 4950 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180782 4950 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180790 4950 flags.go:64] FLAG: --memory-manager-policy="None" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180799 4950 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180808 4950 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180817 4950 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180826 4950 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180876 4950 flags.go:64] FLAG: --node-status-max-images="50" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180885 4950 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180894 4950 flags.go:64] FLAG: --oom-score-adj="-999" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180903 4950 flags.go:64] FLAG: --pod-cidr="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180912 4950 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180927 4950 flags.go:64] FLAG: --pod-manifest-path="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180936 4950 flags.go:64] FLAG: --pod-max-pids="-1" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180945 4950 flags.go:64] FLAG: --pods-per-core="0" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180954 4950 flags.go:64] FLAG: --port="10250" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180963 4950 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180972 4950 flags.go:64] FLAG: --provider-id="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180982 4950 flags.go:64] FLAG: --qos-reserved="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.180992 4950 flags.go:64] FLAG: --read-only-port="10255" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181001 4950 flags.go:64] FLAG: --register-node="true" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181011 4950 flags.go:64] FLAG: --register-schedulable="true" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181020 4950 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181034 4950 flags.go:64] FLAG: --registry-burst="10" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181043 4950 flags.go:64] FLAG: --registry-qps="5" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181052 4950 flags.go:64] FLAG: --reserved-cpus="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181061 4950 flags.go:64] FLAG: --reserved-memory="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181072 4950 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181081 4950 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181090 4950 flags.go:64] FLAG: --rotate-certificates="false" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181099 4950 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181107 4950 flags.go:64] FLAG: --runonce="false" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181117 4950 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181127 4950 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181136 4950 flags.go:64] FLAG: --seccomp-default="false" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181145 4950 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181155 4950 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181164 4950 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181173 4950 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181182 4950 flags.go:64] FLAG: --storage-driver-password="root" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181191 4950 flags.go:64] FLAG: --storage-driver-secure="false" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181201 4950 flags.go:64] FLAG: --storage-driver-table="stats" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181210 4950 flags.go:64] FLAG: --storage-driver-user="root" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181219 4950 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181228 4950 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181237 4950 flags.go:64] FLAG: --system-cgroups="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181246 4950 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181261 4950 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181270 4950 flags.go:64] FLAG: --tls-cert-file="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181278 4950 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181289 4950 flags.go:64] FLAG: --tls-min-version="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181298 4950 flags.go:64] FLAG: --tls-private-key-file="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181307 4950 flags.go:64] FLAG: --topology-manager-policy="none" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181315 4950 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181324 4950 flags.go:64] FLAG: --topology-manager-scope="container" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181333 4950 flags.go:64] FLAG: --v="2" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181344 4950 flags.go:64] FLAG: --version="false" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181355 4950 flags.go:64] FLAG: --vmodule="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181365 4950 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.181374 4950 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181646 4950 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181660 4950 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181672 4950 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181683 4950 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181692 4950 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181703 4950 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181713 4950 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181722 4950 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181731 4950 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181740 4950 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181748 4950 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181756 4950 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181764 4950 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181773 4950 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181781 4950 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181788 4950 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181796 4950 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181805 4950 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181813 4950 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181820 4950 feature_gate.go:330] unrecognized feature gate: Example Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181828 4950 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181835 4950 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181843 4950 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181852 4950 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181860 4950 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181868 4950 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181875 4950 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181883 4950 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181892 4950 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181900 4950 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181908 4950 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181916 4950 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181924 4950 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181935 4950 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181945 4950 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181953 4950 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181962 4950 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181971 4950 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181980 4950 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181989 4950 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.181997 4950 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182006 4950 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182015 4950 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182023 4950 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182032 4950 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182040 4950 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182048 4950 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182056 4950 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182064 4950 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182078 4950 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182086 4950 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182094 4950 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182102 4950 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182110 4950 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182118 4950 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182125 4950 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182133 4950 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182141 4950 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182149 4950 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182161 4950 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182171 4950 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182179 4950 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182187 4950 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182195 4950 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182205 4950 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182212 4950 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182220 4950 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182228 4950 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182236 4950 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182244 4950 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.182252 4950 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.182265 4950 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.196698 4950 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.196749 4950 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.196935 4950 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.196951 4950 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.196961 4950 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.196973 4950 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.196982 4950 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.196991 4950 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197002 4950 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197012 4950 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197021 4950 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197031 4950 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197040 4950 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197048 4950 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197059 4950 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197070 4950 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197081 4950 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197091 4950 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197100 4950 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197109 4950 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197119 4950 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197129 4950 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197139 4950 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197149 4950 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197158 4950 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197168 4950 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197176 4950 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197185 4950 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197193 4950 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197201 4950 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197209 4950 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197217 4950 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197226 4950 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197234 4950 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197242 4950 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197250 4950 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197259 4950 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197267 4950 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197275 4950 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197283 4950 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197291 4950 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197299 4950 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197306 4950 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197315 4950 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197323 4950 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197331 4950 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197339 4950 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197347 4950 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197355 4950 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197363 4950 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197372 4950 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197379 4950 feature_gate.go:330] unrecognized feature gate: Example Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197387 4950 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197395 4950 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197403 4950 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197437 4950 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197447 4950 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197454 4950 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197462 4950 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197472 4950 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197481 4950 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197491 4950 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197499 4950 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197506 4950 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197514 4950 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197522 4950 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197530 4950 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197538 4950 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197545 4950 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197553 4950 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197561 4950 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197568 4950 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197577 4950 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.197590 4950 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197835 4950 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197848 4950 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197856 4950 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197864 4950 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197872 4950 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197882 4950 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197893 4950 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197904 4950 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197914 4950 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197924 4950 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197934 4950 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197943 4950 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197952 4950 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197959 4950 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197967 4950 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197978 4950 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197987 4950 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.197996 4950 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198005 4950 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198013 4950 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198022 4950 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198030 4950 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198037 4950 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198045 4950 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198053 4950 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198061 4950 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198069 4950 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198076 4950 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198084 4950 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198092 4950 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198100 4950 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198108 4950 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198116 4950 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198123 4950 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198131 4950 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198141 4950 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198149 4950 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198157 4950 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198166 4950 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198174 4950 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198182 4950 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198191 4950 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198199 4950 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198206 4950 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198214 4950 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198222 4950 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198230 4950 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198238 4950 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198246 4950 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198253 4950 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198261 4950 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198268 4950 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198276 4950 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198284 4950 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198292 4950 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198300 4950 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198307 4950 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198316 4950 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198323 4950 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198331 4950 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198338 4950 feature_gate.go:330] unrecognized feature gate: Example Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198346 4950 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198354 4950 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198364 4950 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198373 4950 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198382 4950 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198390 4950 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198398 4950 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198407 4950 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198440 4950 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.198460 4950 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.198473 4950 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.200478 4950 server.go:940] "Client rotation is on, will bootstrap in background" Mar 18 20:06:27 crc kubenswrapper[4950]: E0318 20:06:27.204903 4950 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.209499 4950 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.209647 4950 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.211316 4950 server.go:997] "Starting client certificate rotation" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.211370 4950 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.211699 4950 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.235221 4950 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.241576 4950 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 18 20:06:27 crc kubenswrapper[4950]: E0318 20:06:27.241628 4950 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.163:6443: connect: connection refused" logger="UnhandledError" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.265952 4950 log.go:25] "Validated CRI v1 runtime API" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.311272 4950 log.go:25] "Validated CRI v1 image API" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.314524 4950 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.319574 4950 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-03-18-20-00-36-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.319660 4950 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.340847 4950 manager.go:217] Machine: {Timestamp:2026-03-18 20:06:27.339067848 +0000 UTC m=+0.579909796 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2800000 MemoryCapacity:25199476736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:96c0961c-2d17-44f9-aece-16ddcd69eb85 BootID:28057439-b38c-49e6-908d-4b8eb91daf63 Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599738368 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:00:15:6e Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:00:15:6e Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:0b:30:1b Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:88:74:37 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:f3:e6:bb Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:49:25:dc Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:41:87:4a Speed:-1 Mtu:1496} {Name:eth10 MacAddress:d6:ce:cc:53:d9:61 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:76:3e:f6:cb:c1:ca Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199476736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.341160 4950 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.341337 4950 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.343454 4950 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.343769 4950 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.343812 4950 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.344122 4950 topology_manager.go:138] "Creating topology manager with none policy" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.344141 4950 container_manager_linux.go:303] "Creating device plugin manager" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.344810 4950 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.344860 4950 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.345215 4950 state_mem.go:36] "Initialized new in-memory state store" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.345344 4950 server.go:1245] "Using root directory" path="/var/lib/kubelet" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.348793 4950 kubelet.go:418] "Attempting to sync node with API server" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.348829 4950 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.348867 4950 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.348887 4950 kubelet.go:324] "Adding apiserver pod source" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.348906 4950 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.353223 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.163:6443: connect: connection refused Mar 18 20:06:27 crc kubenswrapper[4950]: E0318 20:06:27.353335 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.163:6443: connect: connection refused" logger="UnhandledError" Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.353285 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.163:6443: connect: connection refused Mar 18 20:06:27 crc kubenswrapper[4950]: E0318 20:06:27.354247 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.163:6443: connect: connection refused" logger="UnhandledError" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.354384 4950 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.358756 4950 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.362506 4950 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.364348 4950 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.364391 4950 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.364406 4950 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.364449 4950 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.364474 4950 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.364488 4950 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.364502 4950 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.364526 4950 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.364541 4950 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.364556 4950 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.364598 4950 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.364611 4950 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.366190 4950 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.366846 4950 server.go:1280] "Started kubelet" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.367164 4950 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.367934 4950 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.368401 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.163:6443: connect: connection refused Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.369351 4950 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 18 20:06:27 crc systemd[1]: Started Kubernetes Kubelet. Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.371813 4950 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.371922 4950 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.372257 4950 server.go:460] "Adding debug handlers to kubelet server" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.372852 4950 volume_manager.go:287] "The desired_state_of_world populator starts" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.372889 4950 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.373066 4950 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 18 20:06:27 crc kubenswrapper[4950]: E0318 20:06:27.373343 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:06:27 crc kubenswrapper[4950]: E0318 20:06:27.374169 4950 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.163:6443: connect: connection refused" interval="200ms" Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.375663 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.163:6443: connect: connection refused Mar 18 20:06:27 crc kubenswrapper[4950]: E0318 20:06:27.375831 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.163:6443: connect: connection refused" logger="UnhandledError" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.376318 4950 factory.go:55] Registering systemd factory Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.376370 4950 factory.go:221] Registration of the systemd container factory successfully Mar 18 20:06:27 crc kubenswrapper[4950]: E0318 20:06:27.374240 4950 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.163:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189e084310b6ed5c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.366808924 +0000 UTC m=+0.607650823,LastTimestamp:2026-03-18 20:06:27.366808924 +0000 UTC m=+0.607650823,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.377311 4950 factory.go:153] Registering CRI-O factory Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.377344 4950 factory.go:221] Registration of the crio container factory successfully Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.377516 4950 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.377558 4950 factory.go:103] Registering Raw factory Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.377657 4950 manager.go:1196] Started watching for new ooms in manager Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.378982 4950 manager.go:319] Starting recovery of all containers Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.405798 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.405924 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.405958 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.405986 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406014 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406041 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406069 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406097 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406127 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406151 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406177 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406200 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406226 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406256 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406287 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406328 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406355 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406380 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406407 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406472 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406500 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406528 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406555 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406588 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406616 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406642 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406676 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406705 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406740 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406766 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406792 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406860 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406887 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406914 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406942 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.406994 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.407018 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.407044 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.407096 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.407122 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.407178 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.407205 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.407233 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.407258 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.407287 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.407315 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.407381 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.407452 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.407485 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.407516 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.407545 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.407574 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.414625 4950 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.414716 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.414754 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.414785 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.414816 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.414849 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.414883 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.414912 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.414965 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.414994 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415024 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415051 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415078 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415105 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415133 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415159 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415192 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415224 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415254 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415283 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415313 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415343 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415374 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415445 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415483 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415516 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415544 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415571 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415599 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415625 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415653 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415681 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415710 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415740 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415768 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415797 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415815 4950 manager.go:324] Recovery completed Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415828 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415859 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415889 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415917 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415951 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.415981 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.416011 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.416044 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.416074 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.416105 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.416133 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.416162 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.421632 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.421701 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.421718 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.421735 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.421750 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.421803 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.421818 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.421833 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.421849 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.421863 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.421878 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.421891 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.421907 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.421921 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.421933 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.421962 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.421975 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.421989 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422003 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422016 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422030 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422043 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422056 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422069 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422082 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422096 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422107 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422119 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422133 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422144 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422156 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422169 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422182 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422195 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422207 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422219 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422230 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422245 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422257 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422268 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422281 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422295 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422306 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422319 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422332 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422346 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422358 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422370 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422383 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422395 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422428 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422442 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422456 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422471 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422484 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422497 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422513 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422527 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422540 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422555 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422570 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422583 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422596 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422609 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422622 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422636 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422651 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422664 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422677 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422690 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422702 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422715 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422728 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422742 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422755 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422767 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422779 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422793 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422811 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422824 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422854 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422865 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422879 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422893 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422905 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422918 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422933 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422947 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422965 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422977 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.422990 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.423003 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.423017 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.423031 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.423045 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.423060 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.423074 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.423088 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.423101 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.423147 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.423161 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.423175 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.423191 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.423204 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.423217 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.423230 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.423244 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.423257 4950 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.423268 4950 reconstruct.go:97] "Volume reconstruction finished" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.423277 4950 reconciler.go:26] "Reconciler: start to sync state" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.432452 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.435248 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.435332 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.435355 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.436673 4950 cpu_manager.go:225] "Starting CPU manager" policy="none" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.436707 4950 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.436737 4950 state_mem.go:36] "Initialized new in-memory state store" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.455921 4950 policy_none.go:49] "None policy: Start" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.457088 4950 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.457137 4950 state_mem.go:35] "Initializing new in-memory state store" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.470398 4950 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 18 20:06:27 crc kubenswrapper[4950]: E0318 20:06:27.473908 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.478341 4950 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.478388 4950 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.478439 4950 kubelet.go:2335] "Starting kubelet main sync loop" Mar 18 20:06:27 crc kubenswrapper[4950]: E0318 20:06:27.478578 4950 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.480400 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.163:6443: connect: connection refused Mar 18 20:06:27 crc kubenswrapper[4950]: E0318 20:06:27.480589 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.163:6443: connect: connection refused" logger="UnhandledError" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.506538 4950 manager.go:334] "Starting Device Plugin manager" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.506602 4950 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.506619 4950 server.go:79] "Starting device plugin registration server" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.507220 4950 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.507254 4950 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.509698 4950 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.509836 4950 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.509850 4950 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 18 20:06:27 crc kubenswrapper[4950]: E0318 20:06:27.517755 4950 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 20:06:27 crc kubenswrapper[4950]: E0318 20:06:27.575886 4950 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.163:6443: connect: connection refused" interval="400ms" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.578706 4950 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.578823 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.580201 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.580247 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.580265 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.580448 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.581227 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.581328 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.581366 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.581368 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.581378 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.581542 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.582051 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.582145 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.582883 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.582919 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.582936 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.583125 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.583290 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.583323 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.584584 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.584616 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.584644 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.585062 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.585098 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.585111 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.586318 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.586372 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.586438 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.586734 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.586759 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.586778 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.587089 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.587763 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.587816 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.589319 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.589370 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.589389 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.589867 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.589877 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.589921 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.589942 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.589945 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.591944 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.591993 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.592005 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.607944 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.609345 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.609391 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.609439 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.609480 4950 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 20:06:27 crc kubenswrapper[4950]: E0318 20:06:27.610136 4950 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.163:6443: connect: connection refused" node="crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.726307 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.726399 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.726467 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.726602 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.726654 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.726729 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.726884 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.726984 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.727008 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.727076 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.727146 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.727167 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.727227 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.727246 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.727335 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.811319 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.813359 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.813458 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.813475 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.813541 4950 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 20:06:27 crc kubenswrapper[4950]: E0318 20:06:27.814594 4950 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.163:6443: connect: connection refused" node="crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.828856 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.828945 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.829048 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.829112 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.829143 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.829202 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.829230 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.829287 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.829316 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.829371 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.829401 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.829473 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.829510 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.829570 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.829602 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.830267 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.830386 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.830595 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.830664 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.830745 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.830746 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.830774 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.830808 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.830812 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.830825 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.830840 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.830805 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.830883 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.831025 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.831150 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.923466 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.944912 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.965018 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.974705 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: E0318 20:06:27.977540 4950 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.163:6443: connect: connection refused" interval="800ms" Mar 18 20:06:27 crc kubenswrapper[4950]: I0318 20:06:27.982010 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.988913 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-840ebfa1d554c2c8cd67c866c0829e322718b84340d38e0a06832f0091235805 WatchSource:0}: Error finding container 840ebfa1d554c2c8cd67c866c0829e322718b84340d38e0a06832f0091235805: Status 404 returned error can't find the container with id 840ebfa1d554c2c8cd67c866c0829e322718b84340d38e0a06832f0091235805 Mar 18 20:06:27 crc kubenswrapper[4950]: W0318 20:06:27.991228 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-f993d634f4d46c21866bc9af6e8819aa5ffc707ffbaab3a1c00cdc5795a1ce50 WatchSource:0}: Error finding container f993d634f4d46c21866bc9af6e8819aa5ffc707ffbaab3a1c00cdc5795a1ce50: Status 404 returned error can't find the container with id f993d634f4d46c21866bc9af6e8819aa5ffc707ffbaab3a1c00cdc5795a1ce50 Mar 18 20:06:28 crc kubenswrapper[4950]: W0318 20:06:28.010132 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-48ca542574c58c3bde5cee87c249b02ee36c0b057aaf5f40580d9dc596b07387 WatchSource:0}: Error finding container 48ca542574c58c3bde5cee87c249b02ee36c0b057aaf5f40580d9dc596b07387: Status 404 returned error can't find the container with id 48ca542574c58c3bde5cee87c249b02ee36c0b057aaf5f40580d9dc596b07387 Mar 18 20:06:28 crc kubenswrapper[4950]: W0318 20:06:28.014386 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-a033049b29e9c5d15a7d2d2ec8386ea3d2d14ec9a2ad8d91e2fbfbcd6a5860f1 WatchSource:0}: Error finding container a033049b29e9c5d15a7d2d2ec8386ea3d2d14ec9a2ad8d91e2fbfbcd6a5860f1: Status 404 returned error can't find the container with id a033049b29e9c5d15a7d2d2ec8386ea3d2d14ec9a2ad8d91e2fbfbcd6a5860f1 Mar 18 20:06:28 crc kubenswrapper[4950]: W0318 20:06:28.016570 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-919ea8912065f591e15cb6704801dfe61100bb53f3bb7b25edbb3c9857fd26cd WatchSource:0}: Error finding container 919ea8912065f591e15cb6704801dfe61100bb53f3bb7b25edbb3c9857fd26cd: Status 404 returned error can't find the container with id 919ea8912065f591e15cb6704801dfe61100bb53f3bb7b25edbb3c9857fd26cd Mar 18 20:06:28 crc kubenswrapper[4950]: I0318 20:06:28.215327 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:28 crc kubenswrapper[4950]: I0318 20:06:28.217711 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:28 crc kubenswrapper[4950]: I0318 20:06:28.217764 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:28 crc kubenswrapper[4950]: I0318 20:06:28.217782 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:28 crc kubenswrapper[4950]: I0318 20:06:28.217815 4950 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 20:06:28 crc kubenswrapper[4950]: E0318 20:06:28.218464 4950 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.163:6443: connect: connection refused" node="crc" Mar 18 20:06:28 crc kubenswrapper[4950]: I0318 20:06:28.370182 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.163:6443: connect: connection refused Mar 18 20:06:28 crc kubenswrapper[4950]: W0318 20:06:28.483541 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.163:6443: connect: connection refused Mar 18 20:06:28 crc kubenswrapper[4950]: E0318 20:06:28.483635 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.163:6443: connect: connection refused" logger="UnhandledError" Mar 18 20:06:28 crc kubenswrapper[4950]: I0318 20:06:28.483663 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"48ca542574c58c3bde5cee87c249b02ee36c0b057aaf5f40580d9dc596b07387"} Mar 18 20:06:28 crc kubenswrapper[4950]: I0318 20:06:28.487316 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a033049b29e9c5d15a7d2d2ec8386ea3d2d14ec9a2ad8d91e2fbfbcd6a5860f1"} Mar 18 20:06:28 crc kubenswrapper[4950]: I0318 20:06:28.489337 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"840ebfa1d554c2c8cd67c866c0829e322718b84340d38e0a06832f0091235805"} Mar 18 20:06:28 crc kubenswrapper[4950]: I0318 20:06:28.491531 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f993d634f4d46c21866bc9af6e8819aa5ffc707ffbaab3a1c00cdc5795a1ce50"} Mar 18 20:06:28 crc kubenswrapper[4950]: I0318 20:06:28.493159 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"919ea8912065f591e15cb6704801dfe61100bb53f3bb7b25edbb3c9857fd26cd"} Mar 18 20:06:28 crc kubenswrapper[4950]: W0318 20:06:28.666102 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.163:6443: connect: connection refused Mar 18 20:06:28 crc kubenswrapper[4950]: E0318 20:06:28.666258 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.163:6443: connect: connection refused" logger="UnhandledError" Mar 18 20:06:28 crc kubenswrapper[4950]: E0318 20:06:28.778320 4950 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.163:6443: connect: connection refused" interval="1.6s" Mar 18 20:06:28 crc kubenswrapper[4950]: W0318 20:06:28.874052 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.163:6443: connect: connection refused Mar 18 20:06:28 crc kubenswrapper[4950]: E0318 20:06:28.874122 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.163:6443: connect: connection refused" logger="UnhandledError" Mar 18 20:06:28 crc kubenswrapper[4950]: W0318 20:06:28.937927 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.163:6443: connect: connection refused Mar 18 20:06:28 crc kubenswrapper[4950]: E0318 20:06:28.938024 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.163:6443: connect: connection refused" logger="UnhandledError" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.019250 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.022226 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.022262 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.022274 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.022299 4950 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 20:06:29 crc kubenswrapper[4950]: E0318 20:06:29.022814 4950 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.163:6443: connect: connection refused" node="crc" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.370306 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.163:6443: connect: connection refused Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.406597 4950 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 18 20:06:29 crc kubenswrapper[4950]: E0318 20:06:29.408401 4950 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.163:6443: connect: connection refused" logger="UnhandledError" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.498010 4950 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d72fd3bb32eb2e59901013085fcebe6149876bd4c48316cf631d769deaba01d0" exitCode=0 Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.498098 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d72fd3bb32eb2e59901013085fcebe6149876bd4c48316cf631d769deaba01d0"} Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.498221 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.499562 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.499612 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.499632 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.501288 4950 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="9ea96d85a7524dbdb86733ec82ee52e98ad1de83fc26e1651d039637e489523f" exitCode=0 Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.501396 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"9ea96d85a7524dbdb86733ec82ee52e98ad1de83fc26e1651d039637e489523f"} Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.501588 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.503394 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.503492 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.503512 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.504755 4950 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb" exitCode=0 Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.504803 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb"} Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.505269 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.506604 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.506642 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.506657 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.508742 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.510096 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8e89899cf4b5cbc8566cda6ebfc9a858c7bb0f743bc4ed17b63de28c47fcfa30"} Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.510133 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3e5979160660396987502257fbb0f04771de4c0de21fe3d1be16deb1da3aacf2"} Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.510175 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1c09d46a4d315127ff42d2eb02cf657f6fbe2bdc23eb05addee4f8da9d736498"} Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.510192 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5e662b624a83fcb0b0eab1048193fd00ce4a744f1a4653d776e0584aec856c49"} Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.510308 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.510326 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.510349 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.510364 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.511731 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.511771 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.511793 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.513350 4950 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="6e66d6bfce75c340eebe5f45e62b451508a52e5b83caf85af2c318bb87e47d2d" exitCode=0 Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.513429 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"6e66d6bfce75c340eebe5f45e62b451508a52e5b83caf85af2c318bb87e47d2d"} Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.513492 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.514329 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.514363 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:29 crc kubenswrapper[4950]: I0318 20:06:29.514381 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.369998 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.163:6443: connect: connection refused Mar 18 20:06:30 crc kubenswrapper[4950]: E0318 20:06:30.379140 4950 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.163:6443: connect: connection refused" interval="3.2s" Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.520645 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"88a63d65bec16e87319bce4a3265538d83f89e35ce181fe8013917a17c4c88ba"} Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.520787 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.522809 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.522850 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.522900 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.524044 4950 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="0e62355cd468bbed7dd16a39a4a92131ff015ae83f22c529f92269ccc314427d" exitCode=0 Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.524282 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.524925 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"0e62355cd468bbed7dd16a39a4a92131ff015ae83f22c529f92269ccc314427d"} Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.525481 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.525501 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.525514 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.528621 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"4c4987ceacf9ceda23f6c9f9b3b79cc2984e0e088b89d99ba7c2b0560baf5f2b"} Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.528649 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"973ac70ac0e653b6a420f5bfb070f6304dc2d828022a97a5804ac3c050187888"} Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.528660 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c8c38d8f163e6271ccc589aac12d588aabc2ed0963ec5d13976b3a56e94c64a8"} Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.528757 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.530216 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.530239 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.530248 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.537216 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.537177 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3"} Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.537348 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db"} Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.537369 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4"} Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.537380 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc"} Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.549693 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.549745 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.549759 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.623854 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.625260 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.625305 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.625316 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:30 crc kubenswrapper[4950]: I0318 20:06:30.625351 4950 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 20:06:30 crc kubenswrapper[4950]: E0318 20:06:30.626219 4950 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.163:6443: connect: connection refused" node="crc" Mar 18 20:06:30 crc kubenswrapper[4950]: W0318 20:06:30.693623 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.163:6443: connect: connection refused Mar 18 20:06:30 crc kubenswrapper[4950]: E0318 20:06:30.693710 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.163:6443: connect: connection refused" logger="UnhandledError" Mar 18 20:06:31 crc kubenswrapper[4950]: I0318 20:06:31.543275 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a409769d46f79e92f9f3a68f0e552693d4a01073ec09d3cbcdcd9ba61a73a3f5"} Mar 18 20:06:31 crc kubenswrapper[4950]: I0318 20:06:31.543455 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:31 crc kubenswrapper[4950]: I0318 20:06:31.544366 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:31 crc kubenswrapper[4950]: I0318 20:06:31.544391 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:31 crc kubenswrapper[4950]: I0318 20:06:31.544401 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:31 crc kubenswrapper[4950]: I0318 20:06:31.548083 4950 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="90d8c839ec81d38c9ec6781beb967c1e03b5ce16681d2eb6598b41e0e442ff13" exitCode=0 Mar 18 20:06:31 crc kubenswrapper[4950]: I0318 20:06:31.548167 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:31 crc kubenswrapper[4950]: I0318 20:06:31.548146 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"90d8c839ec81d38c9ec6781beb967c1e03b5ce16681d2eb6598b41e0e442ff13"} Mar 18 20:06:31 crc kubenswrapper[4950]: I0318 20:06:31.548225 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:31 crc kubenswrapper[4950]: I0318 20:06:31.548211 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:31 crc kubenswrapper[4950]: I0318 20:06:31.548773 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:31 crc kubenswrapper[4950]: I0318 20:06:31.548806 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:31 crc kubenswrapper[4950]: I0318 20:06:31.548816 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:31 crc kubenswrapper[4950]: I0318 20:06:31.549034 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 20:06:31 crc kubenswrapper[4950]: I0318 20:06:31.549329 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:31 crc kubenswrapper[4950]: I0318 20:06:31.549343 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:31 crc kubenswrapper[4950]: I0318 20:06:31.549351 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:31 crc kubenswrapper[4950]: I0318 20:06:31.549804 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:31 crc kubenswrapper[4950]: I0318 20:06:31.549846 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:31 crc kubenswrapper[4950]: I0318 20:06:31.549868 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:32 crc kubenswrapper[4950]: I0318 20:06:32.555189 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a01c6a910b4dd0ba361a582c3e72676c18d0bb195f5d0dd147ab902eb854e128"} Mar 18 20:06:32 crc kubenswrapper[4950]: I0318 20:06:32.555254 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:32 crc kubenswrapper[4950]: I0318 20:06:32.555256 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c9f1796cf60d36ed85a1907de5325cc8482727130ca1a9b2a5a7c97fc53382d8"} Mar 18 20:06:32 crc kubenswrapper[4950]: I0318 20:06:32.555432 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c94c6c814522ce9983e141882d2ca02eff923305a529c151db9263608ab74231"} Mar 18 20:06:32 crc kubenswrapper[4950]: I0318 20:06:32.555488 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"72a41fef981b904b4171cf45a17bf870ba45b9807e4fee2e32bacacd6ef2ef9c"} Mar 18 20:06:32 crc kubenswrapper[4950]: I0318 20:06:32.555508 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:32 crc kubenswrapper[4950]: I0318 20:06:32.555518 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:06:32 crc kubenswrapper[4950]: I0318 20:06:32.555978 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:32 crc kubenswrapper[4950]: I0318 20:06:32.555998 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:32 crc kubenswrapper[4950]: I0318 20:06:32.556007 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:32 crc kubenswrapper[4950]: I0318 20:06:32.562156 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:32 crc kubenswrapper[4950]: I0318 20:06:32.562212 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:32 crc kubenswrapper[4950]: I0318 20:06:32.562223 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:33 crc kubenswrapper[4950]: I0318 20:06:33.418033 4950 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 18 20:06:33 crc kubenswrapper[4950]: I0318 20:06:33.564265 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"646923e44cb45eb1e0358fbcd8d6b29c5c5e11c80733588730575018dc0dea7e"} Mar 18 20:06:33 crc kubenswrapper[4950]: I0318 20:06:33.564343 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:33 crc kubenswrapper[4950]: I0318 20:06:33.564351 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:33 crc kubenswrapper[4950]: I0318 20:06:33.565774 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:33 crc kubenswrapper[4950]: I0318 20:06:33.565945 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:33 crc kubenswrapper[4950]: I0318 20:06:33.566079 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:33 crc kubenswrapper[4950]: I0318 20:06:33.565778 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:33 crc kubenswrapper[4950]: I0318 20:06:33.566306 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:33 crc kubenswrapper[4950]: I0318 20:06:33.566316 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:33 crc kubenswrapper[4950]: I0318 20:06:33.702122 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 18 20:06:33 crc kubenswrapper[4950]: I0318 20:06:33.754779 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:06:33 crc kubenswrapper[4950]: I0318 20:06:33.770128 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:06:33 crc kubenswrapper[4950]: I0318 20:06:33.826687 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:33 crc kubenswrapper[4950]: I0318 20:06:33.828111 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:33 crc kubenswrapper[4950]: I0318 20:06:33.828285 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:33 crc kubenswrapper[4950]: I0318 20:06:33.828494 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:33 crc kubenswrapper[4950]: I0318 20:06:33.828641 4950 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 20:06:34 crc kubenswrapper[4950]: I0318 20:06:34.567044 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:34 crc kubenswrapper[4950]: I0318 20:06:34.567044 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:34 crc kubenswrapper[4950]: I0318 20:06:34.569138 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:34 crc kubenswrapper[4950]: I0318 20:06:34.569382 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:34 crc kubenswrapper[4950]: I0318 20:06:34.569617 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:34 crc kubenswrapper[4950]: I0318 20:06:34.569763 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:34 crc kubenswrapper[4950]: I0318 20:06:34.569933 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:34 crc kubenswrapper[4950]: I0318 20:06:34.569968 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:34 crc kubenswrapper[4950]: I0318 20:06:34.958668 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:06:34 crc kubenswrapper[4950]: I0318 20:06:34.958945 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:34 crc kubenswrapper[4950]: I0318 20:06:34.960563 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:34 crc kubenswrapper[4950]: I0318 20:06:34.960680 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:34 crc kubenswrapper[4950]: I0318 20:06:34.960796 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:35 crc kubenswrapper[4950]: I0318 20:06:35.456673 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:06:35 crc kubenswrapper[4950]: I0318 20:06:35.471071 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Mar 18 20:06:35 crc kubenswrapper[4950]: I0318 20:06:35.569237 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:35 crc kubenswrapper[4950]: I0318 20:06:35.569362 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:35 crc kubenswrapper[4950]: I0318 20:06:35.569362 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:35 crc kubenswrapper[4950]: I0318 20:06:35.570843 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:35 crc kubenswrapper[4950]: I0318 20:06:35.570893 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:35 crc kubenswrapper[4950]: I0318 20:06:35.570909 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:35 crc kubenswrapper[4950]: I0318 20:06:35.570933 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:35 crc kubenswrapper[4950]: I0318 20:06:35.570969 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:35 crc kubenswrapper[4950]: I0318 20:06:35.570990 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:35 crc kubenswrapper[4950]: I0318 20:06:35.571004 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:35 crc kubenswrapper[4950]: I0318 20:06:35.571057 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:35 crc kubenswrapper[4950]: I0318 20:06:35.571081 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:36 crc kubenswrapper[4950]: I0318 20:06:36.571380 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:36 crc kubenswrapper[4950]: I0318 20:06:36.572560 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:36 crc kubenswrapper[4950]: I0318 20:06:36.572597 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:36 crc kubenswrapper[4950]: I0318 20:06:36.572611 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:37 crc kubenswrapper[4950]: I0318 20:06:37.471373 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:06:37 crc kubenswrapper[4950]: I0318 20:06:37.471626 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:37 crc kubenswrapper[4950]: I0318 20:06:37.473086 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:37 crc kubenswrapper[4950]: I0318 20:06:37.473162 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:37 crc kubenswrapper[4950]: I0318 20:06:37.473191 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:37 crc kubenswrapper[4950]: E0318 20:06:37.517891 4950 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 20:06:37 crc kubenswrapper[4950]: I0318 20:06:37.959717 4950 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 20:06:37 crc kubenswrapper[4950]: I0318 20:06:37.959829 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 18 20:06:38 crc kubenswrapper[4950]: I0318 20:06:38.299629 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:06:38 crc kubenswrapper[4950]: I0318 20:06:38.299887 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:38 crc kubenswrapper[4950]: I0318 20:06:38.302530 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:38 crc kubenswrapper[4950]: I0318 20:06:38.302599 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:38 crc kubenswrapper[4950]: I0318 20:06:38.302618 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:38 crc kubenswrapper[4950]: I0318 20:06:38.308153 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:06:38 crc kubenswrapper[4950]: I0318 20:06:38.576289 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:38 crc kubenswrapper[4950]: I0318 20:06:38.577521 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:38 crc kubenswrapper[4950]: I0318 20:06:38.577587 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:38 crc kubenswrapper[4950]: I0318 20:06:38.577609 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:38 crc kubenswrapper[4950]: I0318 20:06:38.583310 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:06:39 crc kubenswrapper[4950]: I0318 20:06:39.580160 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:39 crc kubenswrapper[4950]: I0318 20:06:39.581586 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:39 crc kubenswrapper[4950]: I0318 20:06:39.581619 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:39 crc kubenswrapper[4950]: I0318 20:06:39.581636 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:41 crc kubenswrapper[4950]: W0318 20:06:41.358170 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 18 20:06:41 crc kubenswrapper[4950]: I0318 20:06:41.358274 4950 trace.go:236] Trace[333622104]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (18-Mar-2026 20:06:31.356) (total time: 10001ms): Mar 18 20:06:41 crc kubenswrapper[4950]: Trace[333622104]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (20:06:41.358) Mar 18 20:06:41 crc kubenswrapper[4950]: Trace[333622104]: [10.00169233s] [10.00169233s] END Mar 18 20:06:41 crc kubenswrapper[4950]: E0318 20:06:41.358301 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 18 20:06:41 crc kubenswrapper[4950]: I0318 20:06:41.369844 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Mar 18 20:06:41 crc kubenswrapper[4950]: I0318 20:06:41.585949 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 18 20:06:41 crc kubenswrapper[4950]: I0318 20:06:41.587972 4950 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a409769d46f79e92f9f3a68f0e552693d4a01073ec09d3cbcdcd9ba61a73a3f5" exitCode=255 Mar 18 20:06:41 crc kubenswrapper[4950]: I0318 20:06:41.588026 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a409769d46f79e92f9f3a68f0e552693d4a01073ec09d3cbcdcd9ba61a73a3f5"} Mar 18 20:06:41 crc kubenswrapper[4950]: I0318 20:06:41.588192 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:41 crc kubenswrapper[4950]: I0318 20:06:41.589117 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:41 crc kubenswrapper[4950]: I0318 20:06:41.589157 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:41 crc kubenswrapper[4950]: I0318 20:06:41.589180 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:41 crc kubenswrapper[4950]: I0318 20:06:41.589690 4950 scope.go:117] "RemoveContainer" containerID="a409769d46f79e92f9f3a68f0e552693d4a01073ec09d3cbcdcd9ba61a73a3f5" Mar 18 20:06:41 crc kubenswrapper[4950]: W0318 20:06:41.998589 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 18 20:06:41 crc kubenswrapper[4950]: I0318 20:06:41.998683 4950 trace.go:236] Trace[1717230918]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (18-Mar-2026 20:06:31.996) (total time: 10001ms): Mar 18 20:06:41 crc kubenswrapper[4950]: Trace[1717230918]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (20:06:41.998) Mar 18 20:06:41 crc kubenswrapper[4950]: Trace[1717230918]: [10.00171215s] [10.00171215s] END Mar 18 20:06:41 crc kubenswrapper[4950]: E0318 20:06:41.998706 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 18 20:06:42 crc kubenswrapper[4950]: W0318 20:06:42.004106 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 18 20:06:42 crc kubenswrapper[4950]: I0318 20:06:42.004163 4950 trace.go:236] Trace[471399066]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (18-Mar-2026 20:06:32.003) (total time: 10001ms): Mar 18 20:06:42 crc kubenswrapper[4950]: Trace[471399066]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (20:06:42.004) Mar 18 20:06:42 crc kubenswrapper[4950]: Trace[471399066]: [10.0011367s] [10.0011367s] END Mar 18 20:06:42 crc kubenswrapper[4950]: E0318 20:06:42.004178 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 18 20:06:42 crc kubenswrapper[4950]: E0318 20:06:42.318483 4950 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:42Z is after 2026-02-23T05:33:13Z" interval="6.4s" Mar 18 20:06:42 crc kubenswrapper[4950]: E0318 20:06:42.318686 4950 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:42Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189e084310b6ed5c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.366808924 +0000 UTC m=+0.607650823,LastTimestamp:2026-03-18 20:06:27.366808924 +0000 UTC m=+0.607650823,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:06:42 crc kubenswrapper[4950]: W0318 20:06:42.319385 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:42Z is after 2026-02-23T05:33:13Z Mar 18 20:06:42 crc kubenswrapper[4950]: E0318 20:06:42.319533 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:42Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 20:06:42 crc kubenswrapper[4950]: E0318 20:06:42.322721 4950 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:42Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 20:06:42 crc kubenswrapper[4950]: E0318 20:06:42.323733 4950 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:42Z is after 2026-02-23T05:33:13Z" node="crc" Mar 18 20:06:42 crc kubenswrapper[4950]: I0318 20:06:42.325227 4950 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 18 20:06:42 crc kubenswrapper[4950]: I0318 20:06:42.325293 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 18 20:06:42 crc kubenswrapper[4950]: I0318 20:06:42.330488 4950 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 18 20:06:42 crc kubenswrapper[4950]: I0318 20:06:42.330553 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 18 20:06:42 crc kubenswrapper[4950]: I0318 20:06:42.372499 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:42Z is after 2026-02-23T05:33:13Z Mar 18 20:06:42 crc kubenswrapper[4950]: I0318 20:06:42.525356 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:06:42 crc kubenswrapper[4950]: I0318 20:06:42.592692 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 18 20:06:42 crc kubenswrapper[4950]: I0318 20:06:42.594695 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"327de9b5499b181c822c66c141e9902576cc76d3165bc79651ff27525b6cc391"} Mar 18 20:06:42 crc kubenswrapper[4950]: I0318 20:06:42.594837 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:42 crc kubenswrapper[4950]: I0318 20:06:42.595827 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:42 crc kubenswrapper[4950]: I0318 20:06:42.595858 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:42 crc kubenswrapper[4950]: I0318 20:06:42.595871 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:43 crc kubenswrapper[4950]: I0318 20:06:43.384959 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:43Z is after 2026-02-23T05:33:13Z Mar 18 20:06:43 crc kubenswrapper[4950]: I0318 20:06:43.599369 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 18 20:06:43 crc kubenswrapper[4950]: I0318 20:06:43.600246 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 18 20:06:43 crc kubenswrapper[4950]: I0318 20:06:43.602915 4950 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="327de9b5499b181c822c66c141e9902576cc76d3165bc79651ff27525b6cc391" exitCode=255 Mar 18 20:06:43 crc kubenswrapper[4950]: I0318 20:06:43.603017 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:43 crc kubenswrapper[4950]: I0318 20:06:43.602997 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"327de9b5499b181c822c66c141e9902576cc76d3165bc79651ff27525b6cc391"} Mar 18 20:06:43 crc kubenswrapper[4950]: I0318 20:06:43.603260 4950 scope.go:117] "RemoveContainer" containerID="a409769d46f79e92f9f3a68f0e552693d4a01073ec09d3cbcdcd9ba61a73a3f5" Mar 18 20:06:43 crc kubenswrapper[4950]: I0318 20:06:43.604033 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:43 crc kubenswrapper[4950]: I0318 20:06:43.604220 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:43 crc kubenswrapper[4950]: I0318 20:06:43.604582 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:43 crc kubenswrapper[4950]: I0318 20:06:43.605278 4950 scope.go:117] "RemoveContainer" containerID="327de9b5499b181c822c66c141e9902576cc76d3165bc79651ff27525b6cc391" Mar 18 20:06:43 crc kubenswrapper[4950]: E0318 20:06:43.605810 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 20:06:43 crc kubenswrapper[4950]: I0318 20:06:43.745386 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 18 20:06:43 crc kubenswrapper[4950]: I0318 20:06:43.745659 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:43 crc kubenswrapper[4950]: I0318 20:06:43.747202 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:43 crc kubenswrapper[4950]: I0318 20:06:43.747409 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:43 crc kubenswrapper[4950]: I0318 20:06:43.747636 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:43 crc kubenswrapper[4950]: I0318 20:06:43.765670 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 18 20:06:43 crc kubenswrapper[4950]: I0318 20:06:43.777155 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:06:44 crc kubenswrapper[4950]: I0318 20:06:44.375002 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:44Z is after 2026-02-23T05:33:13Z Mar 18 20:06:44 crc kubenswrapper[4950]: I0318 20:06:44.610350 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 18 20:06:44 crc kubenswrapper[4950]: I0318 20:06:44.613504 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:44 crc kubenswrapper[4950]: I0318 20:06:44.614088 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:44 crc kubenswrapper[4950]: I0318 20:06:44.615060 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:44 crc kubenswrapper[4950]: I0318 20:06:44.615278 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:44 crc kubenswrapper[4950]: I0318 20:06:44.615507 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:44 crc kubenswrapper[4950]: I0318 20:06:44.615508 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:44 crc kubenswrapper[4950]: I0318 20:06:44.615828 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:44 crc kubenswrapper[4950]: I0318 20:06:44.615851 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:44 crc kubenswrapper[4950]: I0318 20:06:44.617458 4950 scope.go:117] "RemoveContainer" containerID="327de9b5499b181c822c66c141e9902576cc76d3165bc79651ff27525b6cc391" Mar 18 20:06:44 crc kubenswrapper[4950]: E0318 20:06:44.617775 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 20:06:44 crc kubenswrapper[4950]: I0318 20:06:44.623308 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:06:45 crc kubenswrapper[4950]: I0318 20:06:45.375137 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:45Z is after 2026-02-23T05:33:13Z Mar 18 20:06:45 crc kubenswrapper[4950]: I0318 20:06:45.616400 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:45 crc kubenswrapper[4950]: I0318 20:06:45.617898 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:45 crc kubenswrapper[4950]: I0318 20:06:45.617961 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:45 crc kubenswrapper[4950]: I0318 20:06:45.617985 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:45 crc kubenswrapper[4950]: I0318 20:06:45.619142 4950 scope.go:117] "RemoveContainer" containerID="327de9b5499b181c822c66c141e9902576cc76d3165bc79651ff27525b6cc391" Mar 18 20:06:45 crc kubenswrapper[4950]: E0318 20:06:45.619496 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 20:06:46 crc kubenswrapper[4950]: I0318 20:06:46.373910 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:46Z is after 2026-02-23T05:33:13Z Mar 18 20:06:46 crc kubenswrapper[4950]: W0318 20:06:46.563053 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:46Z is after 2026-02-23T05:33:13Z Mar 18 20:06:46 crc kubenswrapper[4950]: E0318 20:06:46.563180 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:46Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 20:06:46 crc kubenswrapper[4950]: I0318 20:06:46.620310 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:46 crc kubenswrapper[4950]: I0318 20:06:46.621924 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:46 crc kubenswrapper[4950]: I0318 20:06:46.622088 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:46 crc kubenswrapper[4950]: I0318 20:06:46.622109 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:46 crc kubenswrapper[4950]: I0318 20:06:46.623062 4950 scope.go:117] "RemoveContainer" containerID="327de9b5499b181c822c66c141e9902576cc76d3165bc79651ff27525b6cc391" Mar 18 20:06:46 crc kubenswrapper[4950]: E0318 20:06:46.623378 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 20:06:47 crc kubenswrapper[4950]: I0318 20:06:47.374902 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:47Z is after 2026-02-23T05:33:13Z Mar 18 20:06:47 crc kubenswrapper[4950]: E0318 20:06:47.517987 4950 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 20:06:47 crc kubenswrapper[4950]: W0318 20:06:47.521512 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:47Z is after 2026-02-23T05:33:13Z Mar 18 20:06:47 crc kubenswrapper[4950]: E0318 20:06:47.521598 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:47Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 20:06:47 crc kubenswrapper[4950]: W0318 20:06:47.727139 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:47Z is after 2026-02-23T05:33:13Z Mar 18 20:06:47 crc kubenswrapper[4950]: E0318 20:06:47.727260 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:47Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 20:06:47 crc kubenswrapper[4950]: I0318 20:06:47.960543 4950 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 20:06:47 crc kubenswrapper[4950]: I0318 20:06:47.960651 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 18 20:06:48 crc kubenswrapper[4950]: I0318 20:06:48.375895 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:48Z is after 2026-02-23T05:33:13Z Mar 18 20:06:48 crc kubenswrapper[4950]: I0318 20:06:48.691114 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:06:48 crc kubenswrapper[4950]: I0318 20:06:48.691390 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:48 crc kubenswrapper[4950]: I0318 20:06:48.693296 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:48 crc kubenswrapper[4950]: I0318 20:06:48.693359 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:48 crc kubenswrapper[4950]: I0318 20:06:48.693374 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:48 crc kubenswrapper[4950]: I0318 20:06:48.694010 4950 scope.go:117] "RemoveContainer" containerID="327de9b5499b181c822c66c141e9902576cc76d3165bc79651ff27525b6cc391" Mar 18 20:06:48 crc kubenswrapper[4950]: E0318 20:06:48.694270 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 20:06:48 crc kubenswrapper[4950]: I0318 20:06:48.724406 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:48 crc kubenswrapper[4950]: I0318 20:06:48.727187 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:48 crc kubenswrapper[4950]: I0318 20:06:48.727774 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:48 crc kubenswrapper[4950]: I0318 20:06:48.727797 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:48 crc kubenswrapper[4950]: I0318 20:06:48.727834 4950 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 20:06:48 crc kubenswrapper[4950]: E0318 20:06:48.727938 4950 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:48Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 18 20:06:48 crc kubenswrapper[4950]: E0318 20:06:48.734270 4950 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:48Z is after 2026-02-23T05:33:13Z" node="crc" Mar 18 20:06:49 crc kubenswrapper[4950]: I0318 20:06:49.374400 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:49Z is after 2026-02-23T05:33:13Z Mar 18 20:06:50 crc kubenswrapper[4950]: I0318 20:06:50.374483 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:50Z is after 2026-02-23T05:33:13Z Mar 18 20:06:51 crc kubenswrapper[4950]: I0318 20:06:51.069660 4950 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 18 20:06:51 crc kubenswrapper[4950]: E0318 20:06:51.076371 4950 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:51Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 20:06:51 crc kubenswrapper[4950]: I0318 20:06:51.374911 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:51Z is after 2026-02-23T05:33:13Z Mar 18 20:06:52 crc kubenswrapper[4950]: E0318 20:06:52.324350 4950 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:52Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189e084310b6ed5c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.366808924 +0000 UTC m=+0.607650823,LastTimestamp:2026-03-18 20:06:27.366808924 +0000 UTC m=+0.607650823,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:06:52 crc kubenswrapper[4950]: I0318 20:06:52.374457 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:52Z is after 2026-02-23T05:33:13Z Mar 18 20:06:52 crc kubenswrapper[4950]: I0318 20:06:52.524475 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:06:52 crc kubenswrapper[4950]: I0318 20:06:52.524696 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:52 crc kubenswrapper[4950]: I0318 20:06:52.526269 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:52 crc kubenswrapper[4950]: I0318 20:06:52.526367 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:52 crc kubenswrapper[4950]: I0318 20:06:52.526398 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:52 crc kubenswrapper[4950]: I0318 20:06:52.527944 4950 scope.go:117] "RemoveContainer" containerID="327de9b5499b181c822c66c141e9902576cc76d3165bc79651ff27525b6cc391" Mar 18 20:06:52 crc kubenswrapper[4950]: E0318 20:06:52.528515 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 20:06:53 crc kubenswrapper[4950]: I0318 20:06:53.373549 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:53Z is after 2026-02-23T05:33:13Z Mar 18 20:06:53 crc kubenswrapper[4950]: W0318 20:06:53.899121 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:53Z is after 2026-02-23T05:33:13Z Mar 18 20:06:53 crc kubenswrapper[4950]: E0318 20:06:53.899231 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:53Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 20:06:54 crc kubenswrapper[4950]: W0318 20:06:54.178205 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:54Z is after 2026-02-23T05:33:13Z Mar 18 20:06:54 crc kubenswrapper[4950]: E0318 20:06:54.178312 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:54Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 20:06:54 crc kubenswrapper[4950]: I0318 20:06:54.374160 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:54Z is after 2026-02-23T05:33:13Z Mar 18 20:06:55 crc kubenswrapper[4950]: I0318 20:06:55.374099 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:55Z is after 2026-02-23T05:33:13Z Mar 18 20:06:55 crc kubenswrapper[4950]: E0318 20:06:55.732927 4950 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:55Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 18 20:06:55 crc kubenswrapper[4950]: I0318 20:06:55.735161 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:55 crc kubenswrapper[4950]: I0318 20:06:55.736811 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:55 crc kubenswrapper[4950]: I0318 20:06:55.736870 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:55 crc kubenswrapper[4950]: I0318 20:06:55.736895 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:55 crc kubenswrapper[4950]: I0318 20:06:55.736941 4950 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 20:06:55 crc kubenswrapper[4950]: E0318 20:06:55.740592 4950 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:55Z is after 2026-02-23T05:33:13Z" node="crc" Mar 18 20:06:56 crc kubenswrapper[4950]: I0318 20:06:56.374816 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:56Z is after 2026-02-23T05:33:13Z Mar 18 20:06:57 crc kubenswrapper[4950]: I0318 20:06:57.374129 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:57Z is after 2026-02-23T05:33:13Z Mar 18 20:06:57 crc kubenswrapper[4950]: E0318 20:06:57.518146 4950 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 20:06:57 crc kubenswrapper[4950]: I0318 20:06:57.960168 4950 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 20:06:57 crc kubenswrapper[4950]: I0318 20:06:57.960267 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 18 20:06:57 crc kubenswrapper[4950]: I0318 20:06:57.960345 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:06:57 crc kubenswrapper[4950]: I0318 20:06:57.960569 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:57 crc kubenswrapper[4950]: I0318 20:06:57.962248 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:57 crc kubenswrapper[4950]: I0318 20:06:57.962280 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:57 crc kubenswrapper[4950]: I0318 20:06:57.962291 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:57 crc kubenswrapper[4950]: I0318 20:06:57.962714 4950 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"1c09d46a4d315127ff42d2eb02cf657f6fbe2bdc23eb05addee4f8da9d736498"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 18 20:06:57 crc kubenswrapper[4950]: I0318 20:06:57.962861 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://1c09d46a4d315127ff42d2eb02cf657f6fbe2bdc23eb05addee4f8da9d736498" gracePeriod=30 Mar 18 20:06:58 crc kubenswrapper[4950]: I0318 20:06:58.374807 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:58Z is after 2026-02-23T05:33:13Z Mar 18 20:06:58 crc kubenswrapper[4950]: I0318 20:06:58.661875 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 18 20:06:58 crc kubenswrapper[4950]: I0318 20:06:58.662496 4950 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="1c09d46a4d315127ff42d2eb02cf657f6fbe2bdc23eb05addee4f8da9d736498" exitCode=255 Mar 18 20:06:58 crc kubenswrapper[4950]: I0318 20:06:58.662560 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"1c09d46a4d315127ff42d2eb02cf657f6fbe2bdc23eb05addee4f8da9d736498"} Mar 18 20:06:58 crc kubenswrapper[4950]: I0318 20:06:58.662662 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ced76909ac5d1fd3962332517244a448367b89662e932e93daa0de34f70b4a57"} Mar 18 20:06:58 crc kubenswrapper[4950]: I0318 20:06:58.662838 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:06:58 crc kubenswrapper[4950]: I0318 20:06:58.664683 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:06:58 crc kubenswrapper[4950]: I0318 20:06:58.664759 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:06:58 crc kubenswrapper[4950]: I0318 20:06:58.664790 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:06:59 crc kubenswrapper[4950]: W0318 20:06:59.011701 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:59Z is after 2026-02-23T05:33:13Z Mar 18 20:06:59 crc kubenswrapper[4950]: E0318 20:06:59.011852 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:59Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 20:06:59 crc kubenswrapper[4950]: I0318 20:06:59.374655 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:59Z is after 2026-02-23T05:33:13Z Mar 18 20:06:59 crc kubenswrapper[4950]: W0318 20:06:59.972713 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:59Z is after 2026-02-23T05:33:13Z Mar 18 20:06:59 crc kubenswrapper[4950]: E0318 20:06:59.972832 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:59Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 18 20:07:00 crc kubenswrapper[4950]: I0318 20:07:00.373078 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:07:00Z is after 2026-02-23T05:33:13Z Mar 18 20:07:01 crc kubenswrapper[4950]: I0318 20:07:01.374095 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:07:01Z is after 2026-02-23T05:33:13Z Mar 18 20:07:02 crc kubenswrapper[4950]: E0318 20:07:02.329403 4950 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:07:02Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189e084310b6ed5c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.366808924 +0000 UTC m=+0.607650823,LastTimestamp:2026-03-18 20:06:27.366808924 +0000 UTC m=+0.607650823,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:02 crc kubenswrapper[4950]: I0318 20:07:02.374575 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:07:02Z is after 2026-02-23T05:33:13Z Mar 18 20:07:02 crc kubenswrapper[4950]: E0318 20:07:02.739074 4950 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:07:02Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 18 20:07:02 crc kubenswrapper[4950]: I0318 20:07:02.741150 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:02 crc kubenswrapper[4950]: I0318 20:07:02.742823 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:02 crc kubenswrapper[4950]: I0318 20:07:02.742875 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:02 crc kubenswrapper[4950]: I0318 20:07:02.742892 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:02 crc kubenswrapper[4950]: I0318 20:07:02.742932 4950 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 20:07:02 crc kubenswrapper[4950]: E0318 20:07:02.748524 4950 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:07:02Z is after 2026-02-23T05:33:13Z" node="crc" Mar 18 20:07:03 crc kubenswrapper[4950]: I0318 20:07:03.375867 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:07:03Z is after 2026-02-23T05:33:13Z Mar 18 20:07:04 crc kubenswrapper[4950]: I0318 20:07:04.375290 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:07:04Z is after 2026-02-23T05:33:13Z Mar 18 20:07:04 crc kubenswrapper[4950]: I0318 20:07:04.480303 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:04 crc kubenswrapper[4950]: I0318 20:07:04.481691 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:04 crc kubenswrapper[4950]: I0318 20:07:04.481725 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:04 crc kubenswrapper[4950]: I0318 20:07:04.481768 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:04 crc kubenswrapper[4950]: I0318 20:07:04.482351 4950 scope.go:117] "RemoveContainer" containerID="327de9b5499b181c822c66c141e9902576cc76d3165bc79651ff27525b6cc391" Mar 18 20:07:04 crc kubenswrapper[4950]: I0318 20:07:04.959019 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:07:04 crc kubenswrapper[4950]: I0318 20:07:04.959268 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:04 crc kubenswrapper[4950]: I0318 20:07:04.961132 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:04 crc kubenswrapper[4950]: I0318 20:07:04.961226 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:04 crc kubenswrapper[4950]: I0318 20:07:04.961251 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:05 crc kubenswrapper[4950]: I0318 20:07:05.381351 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:05 crc kubenswrapper[4950]: I0318 20:07:05.688880 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 18 20:07:05 crc kubenswrapper[4950]: I0318 20:07:05.689584 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 18 20:07:05 crc kubenswrapper[4950]: I0318 20:07:05.692028 4950 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ee36133d3e999100537e52bd5423ae1b162565e92a954db3af452449fe0e7a21" exitCode=255 Mar 18 20:07:05 crc kubenswrapper[4950]: I0318 20:07:05.692084 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"ee36133d3e999100537e52bd5423ae1b162565e92a954db3af452449fe0e7a21"} Mar 18 20:07:05 crc kubenswrapper[4950]: I0318 20:07:05.692150 4950 scope.go:117] "RemoveContainer" containerID="327de9b5499b181c822c66c141e9902576cc76d3165bc79651ff27525b6cc391" Mar 18 20:07:05 crc kubenswrapper[4950]: I0318 20:07:05.692315 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:05 crc kubenswrapper[4950]: I0318 20:07:05.693570 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:05 crc kubenswrapper[4950]: I0318 20:07:05.693634 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:05 crc kubenswrapper[4950]: I0318 20:07:05.693652 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:05 crc kubenswrapper[4950]: I0318 20:07:05.694390 4950 scope.go:117] "RemoveContainer" containerID="ee36133d3e999100537e52bd5423ae1b162565e92a954db3af452449fe0e7a21" Mar 18 20:07:05 crc kubenswrapper[4950]: E0318 20:07:05.694725 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 20:07:06 crc kubenswrapper[4950]: I0318 20:07:06.377141 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:06 crc kubenswrapper[4950]: I0318 20:07:06.697027 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 18 20:07:07 crc kubenswrapper[4950]: I0318 20:07:07.375120 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:07 crc kubenswrapper[4950]: I0318 20:07:07.471975 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:07:07 crc kubenswrapper[4950]: I0318 20:07:07.472236 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:07 crc kubenswrapper[4950]: I0318 20:07:07.473955 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:07 crc kubenswrapper[4950]: I0318 20:07:07.474068 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:07 crc kubenswrapper[4950]: I0318 20:07:07.474098 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:07 crc kubenswrapper[4950]: E0318 20:07:07.518503 4950 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 20:07:07 crc kubenswrapper[4950]: I0318 20:07:07.959774 4950 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 20:07:07 crc kubenswrapper[4950]: I0318 20:07:07.960898 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 18 20:07:08 crc kubenswrapper[4950]: I0318 20:07:08.379784 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:08 crc kubenswrapper[4950]: I0318 20:07:08.402303 4950 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 18 20:07:08 crc kubenswrapper[4950]: I0318 20:07:08.430402 4950 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 18 20:07:08 crc kubenswrapper[4950]: I0318 20:07:08.690822 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:07:08 crc kubenswrapper[4950]: I0318 20:07:08.691111 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:08 crc kubenswrapper[4950]: I0318 20:07:08.693056 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:08 crc kubenswrapper[4950]: I0318 20:07:08.693127 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:08 crc kubenswrapper[4950]: I0318 20:07:08.693147 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:08 crc kubenswrapper[4950]: I0318 20:07:08.694139 4950 scope.go:117] "RemoveContainer" containerID="ee36133d3e999100537e52bd5423ae1b162565e92a954db3af452449fe0e7a21" Mar 18 20:07:08 crc kubenswrapper[4950]: E0318 20:07:08.694524 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 20:07:09 crc kubenswrapper[4950]: I0318 20:07:09.376665 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:09 crc kubenswrapper[4950]: I0318 20:07:09.928164 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:09 crc kubenswrapper[4950]: I0318 20:07:09.930011 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:09 crc kubenswrapper[4950]: I0318 20:07:09.930088 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:09 crc kubenswrapper[4950]: I0318 20:07:09.930109 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:09 crc kubenswrapper[4950]: I0318 20:07:09.930273 4950 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 20:07:09 crc kubenswrapper[4950]: E0318 20:07:09.935614 4950 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 18 20:07:09 crc kubenswrapper[4950]: E0318 20:07:09.936722 4950 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 18 20:07:10 crc kubenswrapper[4950]: I0318 20:07:10.378655 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:11 crc kubenswrapper[4950]: I0318 20:07:11.377259 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.339459 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084310b6ed5c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.366808924 +0000 UTC m=+0.607650823,LastTimestamp:2026-03-18 20:06:27.366808924 +0000 UTC m=+0.607650823,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.349652 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084314cc0b90 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.435301776 +0000 UTC m=+0.676143684,LastTimestamp:2026-03-18 20:06:27.435301776 +0000 UTC m=+0.676143684,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.356254 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084314ccbbdb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.435346907 +0000 UTC m=+0.676188815,LastTimestamp:2026-03-18 20:06:27.435346907 +0000 UTC m=+0.676188815,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.358246 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084314cd0be5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.435367397 +0000 UTC m=+0.676209305,LastTimestamp:2026-03-18 20:06:27.435367397 +0000 UTC m=+0.676209305,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.362280 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084319526a16 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.511216662 +0000 UTC m=+0.752058540,LastTimestamp:2026-03-18 20:06:27.511216662 +0000 UTC m=+0.752058540,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.366508 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e084314cc0b90\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084314cc0b90 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.435301776 +0000 UTC m=+0.676143684,LastTimestamp:2026-03-18 20:06:27.580228985 +0000 UTC m=+0.821070863,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.371997 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e084314ccbbdb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084314ccbbdb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.435346907 +0000 UTC m=+0.676188815,LastTimestamp:2026-03-18 20:06:27.580259685 +0000 UTC m=+0.821101563,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: I0318 20:07:12.372250 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.374024 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e084314cd0be5\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084314cd0be5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.435367397 +0000 UTC m=+0.676209305,LastTimestamp:2026-03-18 20:06:27.580274926 +0000 UTC m=+0.821116814,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.379310 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e084314cc0b90\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084314cc0b90 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.435301776 +0000 UTC m=+0.676143684,LastTimestamp:2026-03-18 20:06:27.58135148 +0000 UTC m=+0.822193358,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.383901 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e084314ccbbdb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084314ccbbdb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.435346907 +0000 UTC m=+0.676188815,LastTimestamp:2026-03-18 20:06:27.58137448 +0000 UTC m=+0.822216358,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.391455 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e084314cd0be5\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084314cd0be5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.435367397 +0000 UTC m=+0.676209305,LastTimestamp:2026-03-18 20:06:27.581449062 +0000 UTC m=+0.822290940,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.396445 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e084314cc0b90\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084314cc0b90 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.435301776 +0000 UTC m=+0.676143684,LastTimestamp:2026-03-18 20:06:27.582911824 +0000 UTC m=+0.823753702,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.403303 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e084314ccbbdb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084314ccbbdb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.435346907 +0000 UTC m=+0.676188815,LastTimestamp:2026-03-18 20:06:27.582925695 +0000 UTC m=+0.823767573,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.409021 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e084314cd0be5\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084314cd0be5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.435367397 +0000 UTC m=+0.676209305,LastTimestamp:2026-03-18 20:06:27.582945725 +0000 UTC m=+0.823787603,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.417261 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e084314cc0b90\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084314cc0b90 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.435301776 +0000 UTC m=+0.676143684,LastTimestamp:2026-03-18 20:06:27.584606542 +0000 UTC m=+0.825448410,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.424090 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e084314ccbbdb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084314ccbbdb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.435346907 +0000 UTC m=+0.676188815,LastTimestamp:2026-03-18 20:06:27.584623062 +0000 UTC m=+0.825464920,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.430205 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e084314cd0be5\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084314cd0be5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.435367397 +0000 UTC m=+0.676209305,LastTimestamp:2026-03-18 20:06:27.584650063 +0000 UTC m=+0.825491931,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.436364 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e084314cc0b90\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084314cc0b90 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.435301776 +0000 UTC m=+0.676143684,LastTimestamp:2026-03-18 20:06:27.585089373 +0000 UTC m=+0.825931251,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.441942 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e084314ccbbdb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084314ccbbdb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.435346907 +0000 UTC m=+0.676188815,LastTimestamp:2026-03-18 20:06:27.585105543 +0000 UTC m=+0.825947421,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.447397 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e084314cd0be5\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084314cd0be5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.435367397 +0000 UTC m=+0.676209305,LastTimestamp:2026-03-18 20:06:27.585117653 +0000 UTC m=+0.825959531,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.452555 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e084314cc0b90\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084314cc0b90 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.435301776 +0000 UTC m=+0.676143684,LastTimestamp:2026-03-18 20:06:27.58633797 +0000 UTC m=+0.827179848,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.457110 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e084314ccbbdb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084314ccbbdb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.435346907 +0000 UTC m=+0.676188815,LastTimestamp:2026-03-18 20:06:27.586430512 +0000 UTC m=+0.827272390,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.461572 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e084314cd0be5\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084314cd0be5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.435367397 +0000 UTC m=+0.676209305,LastTimestamp:2026-03-18 20:06:27.586446253 +0000 UTC m=+0.827288131,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.466007 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e084314cc0b90\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084314cc0b90 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.435301776 +0000 UTC m=+0.676143684,LastTimestamp:2026-03-18 20:06:27.58675345 +0000 UTC m=+0.827595328,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.471611 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e084314ccbbdb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e084314ccbbdb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:27.435346907 +0000 UTC m=+0.676188815,LastTimestamp:2026-03-18 20:06:27.58676731 +0000 UTC m=+0.827609188,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.481651 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e08433678af94 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:28.000264084 +0000 UTC m=+1.241105972,LastTimestamp:2026-03-18 20:06:28.000264084 +0000 UTC m=+1.241105972,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.488619 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e08433698d16d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:28.002369901 +0000 UTC m=+1.243211779,LastTimestamp:2026-03-18 20:06:28.002369901 +0000 UTC m=+1.243211779,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.493763 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e084337403d16 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:28.013341974 +0000 UTC m=+1.254183852,LastTimestamp:2026-03-18 20:06:28.013341974 +0000 UTC m=+1.254183852,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.499115 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e084337b3a736 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:28.020905782 +0000 UTC m=+1.261747690,LastTimestamp:2026-03-18 20:06:28.020905782 +0000 UTC m=+1.261747690,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.503835 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e084337eb7225 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:28.024562213 +0000 UTC m=+1.265404091,LastTimestamp:2026-03-18 20:06:28.024562213 +0000 UTC m=+1.265404091,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.509021 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e0843597ac18e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:28.587602318 +0000 UTC m=+1.828444196,LastTimestamp:2026-03-18 20:06:28.587602318 +0000 UTC m=+1.828444196,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.514920 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e084359849ba4 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:28.588247972 +0000 UTC m=+1.829089850,LastTimestamp:2026-03-18 20:06:28.588247972 +0000 UTC m=+1.829089850,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.520455 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e08435986cb9d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:28.588391325 +0000 UTC m=+1.829233203,LastTimestamp:2026-03-18 20:06:28.588391325 +0000 UTC m=+1.829233203,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: I0318 20:07:12.525214 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:07:12 crc kubenswrapper[4950]: I0318 20:07:12.525542 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:12 crc kubenswrapper[4950]: I0318 20:07:12.526952 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:12 crc kubenswrapper[4950]: I0318 20:07:12.527070 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.526921 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e08435993b4c4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:28.589237444 +0000 UTC m=+1.830079322,LastTimestamp:2026-03-18 20:06:28.589237444 +0000 UTC m=+1.830079322,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: I0318 20:07:12.527156 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:12 crc kubenswrapper[4950]: I0318 20:07:12.528162 4950 scope.go:117] "RemoveContainer" containerID="ee36133d3e999100537e52bd5423ae1b162565e92a954db3af452449fe0e7a21" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.528396 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.534956 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e08435a008372 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:28.596368242 +0000 UTC m=+1.837210150,LastTimestamp:2026-03-18 20:06:28.596368242 +0000 UTC m=+1.837210150,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.540739 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e08435a46f5a3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:28.600984995 +0000 UTC m=+1.841826863,LastTimestamp:2026-03-18 20:06:28.600984995 +0000 UTC m=+1.841826863,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.545527 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e08435a8bd2a9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:28.605498025 +0000 UTC m=+1.846339893,LastTimestamp:2026-03-18 20:06:28.605498025 +0000 UTC m=+1.846339893,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.550568 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e08435aa773f1 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:28.607308785 +0000 UTC m=+1.848150673,LastTimestamp:2026-03-18 20:06:28.607308785 +0000 UTC m=+1.848150673,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.557213 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e08435aacfa95 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:28.607670933 +0000 UTC m=+1.848512811,LastTimestamp:2026-03-18 20:06:28.607670933 +0000 UTC m=+1.848512811,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.563283 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e08435ab392b5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:28.608103093 +0000 UTC m=+1.848944981,LastTimestamp:2026-03-18 20:06:28.608103093 +0000 UTC m=+1.848944981,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.568216 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e08435ac9ebc6 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:28.609567686 +0000 UTC m=+1.850409564,LastTimestamp:2026-03-18 20:06:28.609567686 +0000 UTC m=+1.850409564,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.574044 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e08436be4d60d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:28.896544269 +0000 UTC m=+2.137386137,LastTimestamp:2026-03-18 20:06:28.896544269 +0000 UTC m=+2.137386137,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.579582 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e08436c9a4650 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:28.908435024 +0000 UTC m=+2.149276902,LastTimestamp:2026-03-18 20:06:28.908435024 +0000 UTC m=+2.149276902,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.584829 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e08436cae2cd7 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:28.909739223 +0000 UTC m=+2.150581081,LastTimestamp:2026-03-18 20:06:28.909739223 +0000 UTC m=+2.150581081,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.589537 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e08437998cffe openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:29.126443006 +0000 UTC m=+2.367284884,LastTimestamp:2026-03-18 20:06:29.126443006 +0000 UTC m=+2.367284884,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.593957 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e08437a8ddbf8 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:29.142502392 +0000 UTC m=+2.383344270,LastTimestamp:2026-03-18 20:06:29.142502392 +0000 UTC m=+2.383344270,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.598336 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e08437aa2ba29 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:29.143869993 +0000 UTC m=+2.384711871,LastTimestamp:2026-03-18 20:06:29.143869993 +0000 UTC m=+2.384711871,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.603584 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e084386c961da openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:29.347729882 +0000 UTC m=+2.588571780,LastTimestamp:2026-03-18 20:06:29.347729882 +0000 UTC m=+2.588571780,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.609468 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e084387d660d6 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:29.365358806 +0000 UTC m=+2.606200714,LastTimestamp:2026-03-18 20:06:29.365358806 +0000 UTC m=+2.606200714,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.613928 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e08438ffeb0dd openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:29.502218461 +0000 UTC m=+2.743060339,LastTimestamp:2026-03-18 20:06:29.502218461 +0000 UTC m=+2.743060339,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.621316 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e084390462ca6 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:29.506903206 +0000 UTC m=+2.747745084,LastTimestamp:2026-03-18 20:06:29.506903206 +0000 UTC m=+2.747745084,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.627606 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e0843905da7fc openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:29.508442108 +0000 UTC m=+2.749284016,LastTimestamp:2026-03-18 20:06:29.508442108 +0000 UTC m=+2.749284016,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.634454 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e084391cfcb20 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:29.532699424 +0000 UTC m=+2.773541292,LastTimestamp:2026-03-18 20:06:29.532699424 +0000 UTC m=+2.773541292,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.640806 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e0843a1bdda9b openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:29.799959195 +0000 UTC m=+3.040801063,LastTimestamp:2026-03-18 20:06:29.799959195 +0000 UTC m=+3.040801063,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.645247 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e0843a1bfca97 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:29.800086167 +0000 UTC m=+3.040928035,LastTimestamp:2026-03-18 20:06:29.800086167 +0000 UTC m=+3.040928035,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.651355 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e0843a1c71f2d openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:29.800566573 +0000 UTC m=+3.041408441,LastTimestamp:2026-03-18 20:06:29.800566573 +0000 UTC m=+3.041408441,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.655262 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e0843a1ce84a4 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:29.8010513 +0000 UTC m=+3.041893168,LastTimestamp:2026-03-18 20:06:29.8010513 +0000 UTC m=+3.041893168,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.666987 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e0843a27d2480 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:29.812495488 +0000 UTC m=+3.053337356,LastTimestamp:2026-03-18 20:06:29.812495488 +0000 UTC m=+3.053337356,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.670479 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e0843a28e9fae openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:29.813641134 +0000 UTC m=+3.054483002,LastTimestamp:2026-03-18 20:06:29.813641134 +0000 UTC m=+3.054483002,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.673282 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e0843a29ebecf openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:29.814697679 +0000 UTC m=+3.055539547,LastTimestamp:2026-03-18 20:06:29.814697679 +0000 UTC m=+3.055539547,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.675487 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e0843a2b383a4 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:29.816058788 +0000 UTC m=+3.056900656,LastTimestamp:2026-03-18 20:06:29.816058788 +0000 UTC m=+3.056900656,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.682170 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e0843a31357b3 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:29.822338995 +0000 UTC m=+3.063180863,LastTimestamp:2026-03-18 20:06:29.822338995 +0000 UTC m=+3.063180863,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.687060 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e0843a382e12e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:29.829648686 +0000 UTC m=+3.070490554,LastTimestamp:2026-03-18 20:06:29.829648686 +0000 UTC m=+3.070490554,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.693997 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e0843af0d0ff7 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:30.023254007 +0000 UTC m=+3.264095875,LastTimestamp:2026-03-18 20:06:30.023254007 +0000 UTC m=+3.264095875,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.700574 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e0843af0eb449 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:30.023361609 +0000 UTC m=+3.264203487,LastTimestamp:2026-03-18 20:06:30.023361609 +0000 UTC m=+3.264203487,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.706556 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e0843afdd0dbe openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:30.036884926 +0000 UTC m=+3.277726794,LastTimestamp:2026-03-18 20:06:30.036884926 +0000 UTC m=+3.277726794,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.711011 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e0843aff00461 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:30.038127713 +0000 UTC m=+3.278969581,LastTimestamp:2026-03-18 20:06:30.038127713 +0000 UTC m=+3.278969581,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.715202 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e0843affaebc9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:30.038842313 +0000 UTC m=+3.279684181,LastTimestamp:2026-03-18 20:06:30.038842313 +0000 UTC m=+3.279684181,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.722434 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e0843b01237ee openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:30.040369134 +0000 UTC m=+3.281211002,LastTimestamp:2026-03-18 20:06:30.040369134 +0000 UTC m=+3.281211002,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.728875 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e0843bae5f526 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:30.22201783 +0000 UTC m=+3.462859698,LastTimestamp:2026-03-18 20:06:30.22201783 +0000 UTC m=+3.462859698,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.732600 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e0843bb0cb53f openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:30.224557375 +0000 UTC m=+3.465399243,LastTimestamp:2026-03-18 20:06:30.224557375 +0000 UTC m=+3.465399243,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.738153 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e0843bbc70dc6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:30.236769734 +0000 UTC m=+3.477611602,LastTimestamp:2026-03-18 20:06:30.236769734 +0000 UTC m=+3.477611602,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.743906 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e0843bbd6918c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:30.237786508 +0000 UTC m=+3.478628376,LastTimestamp:2026-03-18 20:06:30.237786508 +0000 UTC m=+3.478628376,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.750466 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e0843bc1f5922 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:30.242556194 +0000 UTC m=+3.483398062,LastTimestamp:2026-03-18 20:06:30.242556194 +0000 UTC m=+3.483398062,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.757114 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e0843c814e093 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:30.443196563 +0000 UTC m=+3.684038441,LastTimestamp:2026-03-18 20:06:30.443196563 +0000 UTC m=+3.684038441,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.761758 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e0843c8e5e39b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:30.456894363 +0000 UTC m=+3.697736241,LastTimestamp:2026-03-18 20:06:30.456894363 +0000 UTC m=+3.697736241,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.765406 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e0843c8f6d886 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:30.458005638 +0000 UTC m=+3.698847506,LastTimestamp:2026-03-18 20:06:30.458005638 +0000 UTC m=+3.698847506,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.772505 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e0843cd0c0df5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:30.526504437 +0000 UTC m=+3.767346305,LastTimestamp:2026-03-18 20:06:30.526504437 +0000 UTC m=+3.767346305,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.776185 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e0843d8b0b426 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:30.721844262 +0000 UTC m=+3.962686130,LastTimestamp:2026-03-18 20:06:30.721844262 +0000 UTC m=+3.962686130,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.782369 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e0843dbcd2bfb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:30.774041595 +0000 UTC m=+4.014883463,LastTimestamp:2026-03-18 20:06:30.774041595 +0000 UTC m=+4.014883463,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.787692 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e0843dbe10a9d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:30.775343773 +0000 UTC m=+4.016185641,LastTimestamp:2026-03-18 20:06:30.775343773 +0000 UTC m=+4.016185641,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.792296 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e0843dc7f8541 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:30.785729857 +0000 UTC m=+4.026571725,LastTimestamp:2026-03-18 20:06:30.785729857 +0000 UTC m=+4.026571725,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.798151 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e08440a1f4ec3 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:31.551176387 +0000 UTC m=+4.792018255,LastTimestamp:2026-03-18 20:06:31.551176387 +0000 UTC m=+4.792018255,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.802941 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e0844154a0f18 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:31.738527512 +0000 UTC m=+4.979369390,LastTimestamp:2026-03-18 20:06:31.738527512 +0000 UTC m=+4.979369390,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.807473 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e084416461797 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:31.755044759 +0000 UTC m=+4.995886637,LastTimestamp:2026-03-18 20:06:31.755044759 +0000 UTC m=+4.995886637,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.812131 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e084416646eb1 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:31.757033137 +0000 UTC m=+4.997875025,LastTimestamp:2026-03-18 20:06:31.757033137 +0000 UTC m=+4.997875025,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.816854 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e08442498b7b7 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:31.995340727 +0000 UTC m=+5.236182605,LastTimestamp:2026-03-18 20:06:31.995340727 +0000 UTC m=+5.236182605,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.820704 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e08442572fb69 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:32.009644905 +0000 UTC m=+5.250486783,LastTimestamp:2026-03-18 20:06:32.009644905 +0000 UTC m=+5.250486783,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.825662 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e0844258578c2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:32.010856642 +0000 UTC m=+5.251698500,LastTimestamp:2026-03-18 20:06:32.010856642 +0000 UTC m=+5.251698500,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.830084 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e08443455b9a4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:32.259385764 +0000 UTC m=+5.500227642,LastTimestamp:2026-03-18 20:06:32.259385764 +0000 UTC m=+5.500227642,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.835647 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e084434fbf0e5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:32.270278885 +0000 UTC m=+5.511120763,LastTimestamp:2026-03-18 20:06:32.270278885 +0000 UTC m=+5.511120763,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.841703 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e08443512cfbd openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:32.271777725 +0000 UTC m=+5.512619633,LastTimestamp:2026-03-18 20:06:32.271777725 +0000 UTC m=+5.512619633,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.846175 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e084443abc35b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:32.516682587 +0000 UTC m=+5.757524495,LastTimestamp:2026-03-18 20:06:32.516682587 +0000 UTC m=+5.757524495,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.851322 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e084444866649 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:32.531011145 +0000 UTC m=+5.771853053,LastTimestamp:2026-03-18 20:06:32.531011145 +0000 UTC m=+5.771853053,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.856843 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e08444498c584 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:32.532215172 +0000 UTC m=+5.773057060,LastTimestamp:2026-03-18 20:06:32.532215172 +0000 UTC m=+5.773057060,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.862264 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e084451f08168 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:32.756068712 +0000 UTC m=+5.996910590,LastTimestamp:2026-03-18 20:06:32.756068712 +0000 UTC m=+5.996910590,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.867684 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e084452fc0f03 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:32.773603075 +0000 UTC m=+6.014444983,LastTimestamp:2026-03-18 20:06:32.773603075 +0000 UTC m=+6.014444983,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.874982 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 18 20:07:12 crc kubenswrapper[4950]: &Event{ObjectMeta:{kube-controller-manager-crc.189e0845881b0fb7 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Mar 18 20:07:12 crc kubenswrapper[4950]: body: Mar 18 20:07:12 crc kubenswrapper[4950]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:37.959794615 +0000 UTC m=+11.200636513,LastTimestamp:2026-03-18 20:06:37.959794615 +0000 UTC m=+11.200636513,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 18 20:07:12 crc kubenswrapper[4950]: > Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.881379 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e0845881c4cb8 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:37.959875768 +0000 UTC m=+11.200717666,LastTimestamp:2026-03-18 20:06:37.959875768 +0000 UTC m=+11.200717666,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.889130 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189e0843c8f6d886\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e0843c8f6d886 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:30.458005638 +0000 UTC m=+3.698847506,LastTimestamp:2026-03-18 20:06:41.590635485 +0000 UTC m=+14.831477353,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.895491 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189e0843d8b0b426\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e0843d8b0b426 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:30.721844262 +0000 UTC m=+3.962686130,LastTimestamp:2026-03-18 20:06:41.770019055 +0000 UTC m=+15.010860923,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.903995 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189e0843dbcd2bfb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e0843dbcd2bfb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:30.774041595 +0000 UTC m=+4.014883463,LastTimestamp:2026-03-18 20:06:41.780841559 +0000 UTC m=+15.021683427,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.909937 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 18 20:07:12 crc kubenswrapper[4950]: &Event{ObjectMeta:{kube-apiserver-crc.189e08468c4f0335 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 18 20:07:12 crc kubenswrapper[4950]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 18 20:07:12 crc kubenswrapper[4950]: Mar 18 20:07:12 crc kubenswrapper[4950]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:42.325275445 +0000 UTC m=+15.566117313,LastTimestamp:2026-03-18 20:06:42.325275445 +0000 UTC m=+15.566117313,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 18 20:07:12 crc kubenswrapper[4950]: > Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.916640 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e08468c4fac51 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:42.325318737 +0000 UTC m=+15.566160605,LastTimestamp:2026-03-18 20:06:42.325318737 +0000 UTC m=+15.566160605,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.924700 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189e08468c4f0335\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 18 20:07:12 crc kubenswrapper[4950]: &Event{ObjectMeta:{kube-apiserver-crc.189e08468c4f0335 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 18 20:07:12 crc kubenswrapper[4950]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 18 20:07:12 crc kubenswrapper[4950]: Mar 18 20:07:12 crc kubenswrapper[4950]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:42.325275445 +0000 UTC m=+15.566117313,LastTimestamp:2026-03-18 20:06:42.330534454 +0000 UTC m=+15.571376322,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 18 20:07:12 crc kubenswrapper[4950]: > Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.929155 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189e08468c4fac51\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e08468c4fac51 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:42.325318737 +0000 UTC m=+15.566160605,LastTimestamp:2026-03-18 20:06:42.330573155 +0000 UTC m=+15.571415023,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.937992 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 18 20:07:12 crc kubenswrapper[4950]: &Event{ObjectMeta:{kube-controller-manager-crc.189e0847dc336d85 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 18 20:07:12 crc kubenswrapper[4950]: body: Mar 18 20:07:12 crc kubenswrapper[4950]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:47.960612229 +0000 UTC m=+21.201454127,LastTimestamp:2026-03-18 20:06:47.960612229 +0000 UTC m=+21.201454127,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 18 20:07:12 crc kubenswrapper[4950]: > Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.944068 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e0847dc34b403 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:47.960695811 +0000 UTC m=+21.201537719,LastTimestamp:2026-03-18 20:06:47.960695811 +0000 UTC m=+21.201537719,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.952066 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e0847dc336d85\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 18 20:07:12 crc kubenswrapper[4950]: &Event{ObjectMeta:{kube-controller-manager-crc.189e0847dc336d85 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 18 20:07:12 crc kubenswrapper[4950]: body: Mar 18 20:07:12 crc kubenswrapper[4950]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:47.960612229 +0000 UTC m=+21.201454127,LastTimestamp:2026-03-18 20:06:57.960237574 +0000 UTC m=+31.201079472,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 18 20:07:12 crc kubenswrapper[4950]: > Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.959730 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e0847dc34b403\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e0847dc34b403 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:47.960695811 +0000 UTC m=+21.201537719,LastTimestamp:2026-03-18 20:06:57.960308666 +0000 UTC m=+31.201150564,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.967984 4950 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e084a30617557 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:57.962849623 +0000 UTC m=+31.203691501,LastTimestamp:2026-03-18 20:06:57.962849623 +0000 UTC m=+31.203691501,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.978785 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e08435aa773f1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e08435aa773f1 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:28.607308785 +0000 UTC m=+1.848150673,LastTimestamp:2026-03-18 20:06:58.087465035 +0000 UTC m=+31.328306913,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.986587 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e08436be4d60d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e08436be4d60d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:28.896544269 +0000 UTC m=+2.137386137,LastTimestamp:2026-03-18 20:06:58.324839608 +0000 UTC m=+31.565681516,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:12 crc kubenswrapper[4950]: E0318 20:07:12.992271 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e08436c9a4650\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e08436c9a4650 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:28.908435024 +0000 UTC m=+2.149276902,LastTimestamp:2026-03-18 20:06:58.337129621 +0000 UTC m=+31.577971509,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:13 crc kubenswrapper[4950]: E0318 20:07:13.002147 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e0847dc336d85\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 18 20:07:13 crc kubenswrapper[4950]: &Event{ObjectMeta:{kube-controller-manager-crc.189e0847dc336d85 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 18 20:07:13 crc kubenswrapper[4950]: body: Mar 18 20:07:13 crc kubenswrapper[4950]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:47.960612229 +0000 UTC m=+21.201454127,LastTimestamp:2026-03-18 20:07:07.960793823 +0000 UTC m=+41.201635781,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 18 20:07:13 crc kubenswrapper[4950]: > Mar 18 20:07:13 crc kubenswrapper[4950]: E0318 20:07:13.009507 4950 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e0847dc34b403\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e0847dc34b403 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:06:47.960695811 +0000 UTC m=+21.201537719,LastTimestamp:2026-03-18 20:07:07.961005998 +0000 UTC m=+41.201847896,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:07:13 crc kubenswrapper[4950]: I0318 20:07:13.377370 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:14 crc kubenswrapper[4950]: W0318 20:07:14.251251 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 18 20:07:14 crc kubenswrapper[4950]: E0318 20:07:14.251335 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 18 20:07:14 crc kubenswrapper[4950]: I0318 20:07:14.376520 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:14 crc kubenswrapper[4950]: I0318 20:07:14.964644 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:07:14 crc kubenswrapper[4950]: I0318 20:07:14.965672 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:14 crc kubenswrapper[4950]: I0318 20:07:14.967491 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:14 crc kubenswrapper[4950]: I0318 20:07:14.967561 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:14 crc kubenswrapper[4950]: I0318 20:07:14.967581 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:14 crc kubenswrapper[4950]: I0318 20:07:14.984661 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:07:15 crc kubenswrapper[4950]: W0318 20:07:15.142457 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Mar 18 20:07:15 crc kubenswrapper[4950]: E0318 20:07:15.142525 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 18 20:07:15 crc kubenswrapper[4950]: I0318 20:07:15.372725 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:15 crc kubenswrapper[4950]: I0318 20:07:15.951014 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:15 crc kubenswrapper[4950]: I0318 20:07:15.952025 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:15 crc kubenswrapper[4950]: I0318 20:07:15.952056 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:15 crc kubenswrapper[4950]: I0318 20:07:15.952066 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:16 crc kubenswrapper[4950]: I0318 20:07:16.373446 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:16 crc kubenswrapper[4950]: I0318 20:07:16.936253 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:16 crc kubenswrapper[4950]: I0318 20:07:16.938079 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:16 crc kubenswrapper[4950]: I0318 20:07:16.938144 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:16 crc kubenswrapper[4950]: I0318 20:07:16.938159 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:16 crc kubenswrapper[4950]: I0318 20:07:16.938199 4950 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 20:07:16 crc kubenswrapper[4950]: E0318 20:07:16.943737 4950 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 18 20:07:16 crc kubenswrapper[4950]: E0318 20:07:16.943771 4950 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 18 20:07:17 crc kubenswrapper[4950]: W0318 20:07:17.197205 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:17 crc kubenswrapper[4950]: E0318 20:07:17.197276 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 18 20:07:17 crc kubenswrapper[4950]: I0318 20:07:17.375126 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:17 crc kubenswrapper[4950]: E0318 20:07:17.518638 4950 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 20:07:17 crc kubenswrapper[4950]: W0318 20:07:17.584274 4950 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 18 20:07:17 crc kubenswrapper[4950]: E0318 20:07:17.584338 4950 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 18 20:07:18 crc kubenswrapper[4950]: I0318 20:07:18.373459 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:19 crc kubenswrapper[4950]: I0318 20:07:19.374266 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:20 crc kubenswrapper[4950]: I0318 20:07:20.230113 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 18 20:07:20 crc kubenswrapper[4950]: I0318 20:07:20.230329 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:20 crc kubenswrapper[4950]: I0318 20:07:20.231585 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:20 crc kubenswrapper[4950]: I0318 20:07:20.231626 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:20 crc kubenswrapper[4950]: I0318 20:07:20.231641 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:20 crc kubenswrapper[4950]: I0318 20:07:20.375114 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:21 crc kubenswrapper[4950]: I0318 20:07:21.376155 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:22 crc kubenswrapper[4950]: I0318 20:07:22.378790 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:23 crc kubenswrapper[4950]: I0318 20:07:23.377055 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:23 crc kubenswrapper[4950]: I0318 20:07:23.479371 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:23 crc kubenswrapper[4950]: I0318 20:07:23.481041 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:23 crc kubenswrapper[4950]: I0318 20:07:23.481265 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:23 crc kubenswrapper[4950]: I0318 20:07:23.481476 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:23 crc kubenswrapper[4950]: I0318 20:07:23.482475 4950 scope.go:117] "RemoveContainer" containerID="ee36133d3e999100537e52bd5423ae1b162565e92a954db3af452449fe0e7a21" Mar 18 20:07:23 crc kubenswrapper[4950]: E0318 20:07:23.482934 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 20:07:23 crc kubenswrapper[4950]: I0318 20:07:23.944607 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:23 crc kubenswrapper[4950]: I0318 20:07:23.946811 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:23 crc kubenswrapper[4950]: I0318 20:07:23.946891 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:23 crc kubenswrapper[4950]: I0318 20:07:23.946916 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:23 crc kubenswrapper[4950]: I0318 20:07:23.946967 4950 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 20:07:23 crc kubenswrapper[4950]: E0318 20:07:23.954952 4950 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 18 20:07:23 crc kubenswrapper[4950]: E0318 20:07:23.955715 4950 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 18 20:07:24 crc kubenswrapper[4950]: I0318 20:07:24.376050 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:25 crc kubenswrapper[4950]: I0318 20:07:25.376577 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:26 crc kubenswrapper[4950]: I0318 20:07:26.377159 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:27 crc kubenswrapper[4950]: I0318 20:07:27.377340 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:27 crc kubenswrapper[4950]: E0318 20:07:27.522340 4950 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 20:07:28 crc kubenswrapper[4950]: I0318 20:07:28.376913 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:29 crc kubenswrapper[4950]: I0318 20:07:29.378357 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:30 crc kubenswrapper[4950]: I0318 20:07:30.376507 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:30 crc kubenswrapper[4950]: I0318 20:07:30.955594 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:30 crc kubenswrapper[4950]: I0318 20:07:30.958435 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:30 crc kubenswrapper[4950]: I0318 20:07:30.958568 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:30 crc kubenswrapper[4950]: I0318 20:07:30.958590 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:30 crc kubenswrapper[4950]: I0318 20:07:30.958636 4950 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 20:07:30 crc kubenswrapper[4950]: E0318 20:07:30.964405 4950 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 18 20:07:30 crc kubenswrapper[4950]: E0318 20:07:30.965474 4950 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 18 20:07:31 crc kubenswrapper[4950]: I0318 20:07:31.374564 4950 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 18 20:07:31 crc kubenswrapper[4950]: I0318 20:07:31.941247 4950 csr.go:261] certificate signing request csr-g9fz5 is approved, waiting to be issued Mar 18 20:07:31 crc kubenswrapper[4950]: I0318 20:07:31.947871 4950 csr.go:257] certificate signing request csr-g9fz5 is issued Mar 18 20:07:32 crc kubenswrapper[4950]: I0318 20:07:32.029316 4950 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 18 20:07:32 crc kubenswrapper[4950]: I0318 20:07:32.210928 4950 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 18 20:07:32 crc kubenswrapper[4950]: I0318 20:07:32.949303 4950 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-22 02:22:06.958669656 +0000 UTC Mar 18 20:07:32 crc kubenswrapper[4950]: I0318 20:07:32.949346 4950 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 5958h14m34.009326485s for next certificate rotation Mar 18 20:07:34 crc kubenswrapper[4950]: I0318 20:07:34.478848 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:34 crc kubenswrapper[4950]: I0318 20:07:34.480092 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:34 crc kubenswrapper[4950]: I0318 20:07:34.480762 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:34 crc kubenswrapper[4950]: I0318 20:07:34.480834 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:34 crc kubenswrapper[4950]: I0318 20:07:34.482330 4950 scope.go:117] "RemoveContainer" containerID="ee36133d3e999100537e52bd5423ae1b162565e92a954db3af452449fe0e7a21" Mar 18 20:07:35 crc kubenswrapper[4950]: I0318 20:07:35.011478 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 18 20:07:35 crc kubenswrapper[4950]: I0318 20:07:35.013353 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb"} Mar 18 20:07:35 crc kubenswrapper[4950]: I0318 20:07:35.013515 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:35 crc kubenswrapper[4950]: I0318 20:07:35.014341 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:35 crc kubenswrapper[4950]: I0318 20:07:35.014387 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:35 crc kubenswrapper[4950]: I0318 20:07:35.014405 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:36 crc kubenswrapper[4950]: I0318 20:07:36.017515 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 18 20:07:36 crc kubenswrapper[4950]: I0318 20:07:36.018017 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 18 20:07:36 crc kubenswrapper[4950]: I0318 20:07:36.020435 4950 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb" exitCode=255 Mar 18 20:07:36 crc kubenswrapper[4950]: I0318 20:07:36.020476 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb"} Mar 18 20:07:36 crc kubenswrapper[4950]: I0318 20:07:36.020579 4950 scope.go:117] "RemoveContainer" containerID="ee36133d3e999100537e52bd5423ae1b162565e92a954db3af452449fe0e7a21" Mar 18 20:07:36 crc kubenswrapper[4950]: I0318 20:07:36.020691 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:36 crc kubenswrapper[4950]: I0318 20:07:36.022948 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:36 crc kubenswrapper[4950]: I0318 20:07:36.023018 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:36 crc kubenswrapper[4950]: I0318 20:07:36.023037 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:36 crc kubenswrapper[4950]: I0318 20:07:36.024101 4950 scope.go:117] "RemoveContainer" containerID="7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb" Mar 18 20:07:36 crc kubenswrapper[4950]: E0318 20:07:36.024483 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 20:07:37 crc kubenswrapper[4950]: I0318 20:07:37.030826 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 18 20:07:37 crc kubenswrapper[4950]: E0318 20:07:37.537698 4950 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 20:07:37 crc kubenswrapper[4950]: I0318 20:07:37.966163 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:37 crc kubenswrapper[4950]: I0318 20:07:37.969149 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:37 crc kubenswrapper[4950]: I0318 20:07:37.969308 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:37 crc kubenswrapper[4950]: I0318 20:07:37.969400 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:37 crc kubenswrapper[4950]: I0318 20:07:37.969621 4950 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 18 20:07:37 crc kubenswrapper[4950]: I0318 20:07:37.979515 4950 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 18 20:07:37 crc kubenswrapper[4950]: I0318 20:07:37.979833 4950 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 18 20:07:37 crc kubenswrapper[4950]: E0318 20:07:37.979857 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 18 20:07:37 crc kubenswrapper[4950]: I0318 20:07:37.982889 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:37 crc kubenswrapper[4950]: I0318 20:07:37.982912 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:37 crc kubenswrapper[4950]: I0318 20:07:37.982922 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:37 crc kubenswrapper[4950]: I0318 20:07:37.982939 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:07:37 crc kubenswrapper[4950]: I0318 20:07:37.982952 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:07:37Z","lastTransitionTime":"2026-03-18T20:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:07:37 crc kubenswrapper[4950]: E0318 20:07:37.994218 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:07:38 crc kubenswrapper[4950]: I0318 20:07:38.002116 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:38 crc kubenswrapper[4950]: I0318 20:07:38.002503 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:38 crc kubenswrapper[4950]: I0318 20:07:38.002515 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:38 crc kubenswrapper[4950]: I0318 20:07:38.002531 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:07:38 crc kubenswrapper[4950]: I0318 20:07:38.002541 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:07:38Z","lastTransitionTime":"2026-03-18T20:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:07:38 crc kubenswrapper[4950]: E0318 20:07:38.010996 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:07:38 crc kubenswrapper[4950]: I0318 20:07:38.017378 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:38 crc kubenswrapper[4950]: I0318 20:07:38.017400 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:38 crc kubenswrapper[4950]: I0318 20:07:38.017441 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:38 crc kubenswrapper[4950]: I0318 20:07:38.017458 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:07:38 crc kubenswrapper[4950]: I0318 20:07:38.017470 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:07:38Z","lastTransitionTime":"2026-03-18T20:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:07:38 crc kubenswrapper[4950]: E0318 20:07:38.025978 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:07:38 crc kubenswrapper[4950]: I0318 20:07:38.031905 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:38 crc kubenswrapper[4950]: I0318 20:07:38.031932 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:38 crc kubenswrapper[4950]: I0318 20:07:38.031941 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:38 crc kubenswrapper[4950]: I0318 20:07:38.031953 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:07:38 crc kubenswrapper[4950]: I0318 20:07:38.031962 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:07:38Z","lastTransitionTime":"2026-03-18T20:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:07:38 crc kubenswrapper[4950]: E0318 20:07:38.042118 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:07:38 crc kubenswrapper[4950]: E0318 20:07:38.042224 4950 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 20:07:38 crc kubenswrapper[4950]: E0318 20:07:38.042241 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:38 crc kubenswrapper[4950]: E0318 20:07:38.142709 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:38 crc kubenswrapper[4950]: E0318 20:07:38.242957 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:38 crc kubenswrapper[4950]: E0318 20:07:38.343870 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:38 crc kubenswrapper[4950]: E0318 20:07:38.444627 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:38 crc kubenswrapper[4950]: E0318 20:07:38.545469 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:38 crc kubenswrapper[4950]: E0318 20:07:38.646267 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:38 crc kubenswrapper[4950]: I0318 20:07:38.691112 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:07:38 crc kubenswrapper[4950]: I0318 20:07:38.691293 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:38 crc kubenswrapper[4950]: I0318 20:07:38.692232 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:38 crc kubenswrapper[4950]: I0318 20:07:38.692260 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:38 crc kubenswrapper[4950]: I0318 20:07:38.692268 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:38 crc kubenswrapper[4950]: I0318 20:07:38.692800 4950 scope.go:117] "RemoveContainer" containerID="7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb" Mar 18 20:07:38 crc kubenswrapper[4950]: E0318 20:07:38.692958 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 20:07:38 crc kubenswrapper[4950]: E0318 20:07:38.747126 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:38 crc kubenswrapper[4950]: E0318 20:07:38.847717 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:38 crc kubenswrapper[4950]: E0318 20:07:38.948860 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:39 crc kubenswrapper[4950]: E0318 20:07:39.049649 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:39 crc kubenswrapper[4950]: E0318 20:07:39.151561 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:39 crc kubenswrapper[4950]: E0318 20:07:39.251889 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:39 crc kubenswrapper[4950]: E0318 20:07:39.353009 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:39 crc kubenswrapper[4950]: E0318 20:07:39.453824 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:39 crc kubenswrapper[4950]: E0318 20:07:39.554878 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:39 crc kubenswrapper[4950]: E0318 20:07:39.655336 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:39 crc kubenswrapper[4950]: E0318 20:07:39.756400 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:39 crc kubenswrapper[4950]: E0318 20:07:39.856850 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:39 crc kubenswrapper[4950]: E0318 20:07:39.957459 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:40 crc kubenswrapper[4950]: E0318 20:07:40.070552 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:40 crc kubenswrapper[4950]: E0318 20:07:40.170746 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:40 crc kubenswrapper[4950]: E0318 20:07:40.271908 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:40 crc kubenswrapper[4950]: E0318 20:07:40.372588 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:40 crc kubenswrapper[4950]: E0318 20:07:40.473242 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:40 crc kubenswrapper[4950]: E0318 20:07:40.574010 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:40 crc kubenswrapper[4950]: E0318 20:07:40.674096 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:40 crc kubenswrapper[4950]: E0318 20:07:40.774224 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:40 crc kubenswrapper[4950]: E0318 20:07:40.874725 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:40 crc kubenswrapper[4950]: E0318 20:07:40.975728 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:41 crc kubenswrapper[4950]: E0318 20:07:41.076122 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:41 crc kubenswrapper[4950]: E0318 20:07:41.177156 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:41 crc kubenswrapper[4950]: E0318 20:07:41.277539 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:41 crc kubenswrapper[4950]: E0318 20:07:41.378136 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:41 crc kubenswrapper[4950]: E0318 20:07:41.478954 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:41 crc kubenswrapper[4950]: E0318 20:07:41.579503 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:41 crc kubenswrapper[4950]: E0318 20:07:41.679881 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:41 crc kubenswrapper[4950]: E0318 20:07:41.780928 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:41 crc kubenswrapper[4950]: E0318 20:07:41.881460 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:41 crc kubenswrapper[4950]: E0318 20:07:41.981915 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:42 crc kubenswrapper[4950]: E0318 20:07:42.082483 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:42 crc kubenswrapper[4950]: E0318 20:07:42.183474 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:42 crc kubenswrapper[4950]: I0318 20:07:42.215442 4950 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 18 20:07:42 crc kubenswrapper[4950]: E0318 20:07:42.284610 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:42 crc kubenswrapper[4950]: E0318 20:07:42.384780 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:42 crc kubenswrapper[4950]: E0318 20:07:42.485905 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:42 crc kubenswrapper[4950]: I0318 20:07:42.525167 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:07:42 crc kubenswrapper[4950]: I0318 20:07:42.525305 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:42 crc kubenswrapper[4950]: I0318 20:07:42.526742 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:42 crc kubenswrapper[4950]: I0318 20:07:42.526808 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:42 crc kubenswrapper[4950]: I0318 20:07:42.526828 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:42 crc kubenswrapper[4950]: I0318 20:07:42.527873 4950 scope.go:117] "RemoveContainer" containerID="7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb" Mar 18 20:07:42 crc kubenswrapper[4950]: E0318 20:07:42.528161 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 20:07:42 crc kubenswrapper[4950]: E0318 20:07:42.590204 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:42 crc kubenswrapper[4950]: E0318 20:07:42.691388 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:42 crc kubenswrapper[4950]: E0318 20:07:42.791554 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:42 crc kubenswrapper[4950]: E0318 20:07:42.892008 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:42 crc kubenswrapper[4950]: E0318 20:07:42.993107 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:43 crc kubenswrapper[4950]: E0318 20:07:43.094207 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:43 crc kubenswrapper[4950]: E0318 20:07:43.194539 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:43 crc kubenswrapper[4950]: E0318 20:07:43.295485 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:43 crc kubenswrapper[4950]: E0318 20:07:43.395668 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:43 crc kubenswrapper[4950]: E0318 20:07:43.496247 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:43 crc kubenswrapper[4950]: E0318 20:07:43.597375 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:43 crc kubenswrapper[4950]: E0318 20:07:43.698928 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:43 crc kubenswrapper[4950]: I0318 20:07:43.769262 4950 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 18 20:07:43 crc kubenswrapper[4950]: E0318 20:07:43.799529 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:43 crc kubenswrapper[4950]: E0318 20:07:43.900495 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:44 crc kubenswrapper[4950]: E0318 20:07:44.001731 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:44 crc kubenswrapper[4950]: E0318 20:07:44.102524 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:44 crc kubenswrapper[4950]: E0318 20:07:44.203199 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:44 crc kubenswrapper[4950]: E0318 20:07:44.303991 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:44 crc kubenswrapper[4950]: E0318 20:07:44.404692 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:44 crc kubenswrapper[4950]: E0318 20:07:44.505030 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:44 crc kubenswrapper[4950]: E0318 20:07:44.605860 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:44 crc kubenswrapper[4950]: E0318 20:07:44.706893 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:44 crc kubenswrapper[4950]: E0318 20:07:44.807319 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:44 crc kubenswrapper[4950]: E0318 20:07:44.907854 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:45 crc kubenswrapper[4950]: E0318 20:07:45.008261 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:45 crc kubenswrapper[4950]: E0318 20:07:45.109346 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:45 crc kubenswrapper[4950]: E0318 20:07:45.209505 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:45 crc kubenswrapper[4950]: E0318 20:07:45.309888 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:45 crc kubenswrapper[4950]: E0318 20:07:45.410565 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:45 crc kubenswrapper[4950]: E0318 20:07:45.511042 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:45 crc kubenswrapper[4950]: E0318 20:07:45.611514 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:45 crc kubenswrapper[4950]: E0318 20:07:45.711739 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:45 crc kubenswrapper[4950]: E0318 20:07:45.812352 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:45 crc kubenswrapper[4950]: E0318 20:07:45.913105 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:46 crc kubenswrapper[4950]: E0318 20:07:46.014225 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:46 crc kubenswrapper[4950]: E0318 20:07:46.114750 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:46 crc kubenswrapper[4950]: E0318 20:07:46.215478 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:46 crc kubenswrapper[4950]: E0318 20:07:46.316527 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:46 crc kubenswrapper[4950]: E0318 20:07:46.417103 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:46 crc kubenswrapper[4950]: E0318 20:07:46.517756 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:46 crc kubenswrapper[4950]: E0318 20:07:46.618483 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:46 crc kubenswrapper[4950]: E0318 20:07:46.719829 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:46 crc kubenswrapper[4950]: E0318 20:07:46.821178 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:46 crc kubenswrapper[4950]: E0318 20:07:46.921963 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:47 crc kubenswrapper[4950]: E0318 20:07:47.023114 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:47 crc kubenswrapper[4950]: E0318 20:07:47.123873 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:47 crc kubenswrapper[4950]: E0318 20:07:47.224437 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:47 crc kubenswrapper[4950]: E0318 20:07:47.324640 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:47 crc kubenswrapper[4950]: E0318 20:07:47.425823 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:47 crc kubenswrapper[4950]: I0318 20:07:47.478783 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:47 crc kubenswrapper[4950]: I0318 20:07:47.480587 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:47 crc kubenswrapper[4950]: I0318 20:07:47.480852 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:47 crc kubenswrapper[4950]: I0318 20:07:47.481005 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:47 crc kubenswrapper[4950]: E0318 20:07:47.526682 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:47 crc kubenswrapper[4950]: E0318 20:07:47.538957 4950 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 20:07:47 crc kubenswrapper[4950]: E0318 20:07:47.627705 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:47 crc kubenswrapper[4950]: E0318 20:07:47.728257 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:47 crc kubenswrapper[4950]: E0318 20:07:47.828929 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:47 crc kubenswrapper[4950]: E0318 20:07:47.930121 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:48 crc kubenswrapper[4950]: E0318 20:07:48.031265 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:48 crc kubenswrapper[4950]: E0318 20:07:48.131620 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:48 crc kubenswrapper[4950]: E0318 20:07:48.232197 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:48 crc kubenswrapper[4950]: E0318 20:07:48.256831 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 18 20:07:48 crc kubenswrapper[4950]: I0318 20:07:48.262953 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:48 crc kubenswrapper[4950]: I0318 20:07:48.263016 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:48 crc kubenswrapper[4950]: I0318 20:07:48.263041 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:48 crc kubenswrapper[4950]: I0318 20:07:48.263071 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:07:48 crc kubenswrapper[4950]: I0318 20:07:48.263091 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:07:48Z","lastTransitionTime":"2026-03-18T20:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:07:48 crc kubenswrapper[4950]: E0318 20:07:48.282922 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:07:48 crc kubenswrapper[4950]: I0318 20:07:48.289004 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:48 crc kubenswrapper[4950]: I0318 20:07:48.289351 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:48 crc kubenswrapper[4950]: I0318 20:07:48.289776 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:48 crc kubenswrapper[4950]: I0318 20:07:48.290181 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:07:48 crc kubenswrapper[4950]: I0318 20:07:48.290746 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:07:48Z","lastTransitionTime":"2026-03-18T20:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:07:48 crc kubenswrapper[4950]: E0318 20:07:48.308941 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:07:48 crc kubenswrapper[4950]: I0318 20:07:48.317044 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:48 crc kubenswrapper[4950]: I0318 20:07:48.317101 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:48 crc kubenswrapper[4950]: I0318 20:07:48.317117 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:48 crc kubenswrapper[4950]: I0318 20:07:48.317143 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:07:48 crc kubenswrapper[4950]: I0318 20:07:48.317163 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:07:48Z","lastTransitionTime":"2026-03-18T20:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:07:48 crc kubenswrapper[4950]: E0318 20:07:48.333081 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:07:48 crc kubenswrapper[4950]: I0318 20:07:48.338990 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:48 crc kubenswrapper[4950]: I0318 20:07:48.339051 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:48 crc kubenswrapper[4950]: I0318 20:07:48.339070 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:48 crc kubenswrapper[4950]: I0318 20:07:48.339095 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:07:48 crc kubenswrapper[4950]: I0318 20:07:48.339112 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:07:48Z","lastTransitionTime":"2026-03-18T20:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:07:48 crc kubenswrapper[4950]: E0318 20:07:48.357037 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:07:48 crc kubenswrapper[4950]: E0318 20:07:48.357365 4950 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 20:07:48 crc kubenswrapper[4950]: E0318 20:07:48.357402 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:48 crc kubenswrapper[4950]: E0318 20:07:48.458121 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:48 crc kubenswrapper[4950]: E0318 20:07:48.558319 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:48 crc kubenswrapper[4950]: E0318 20:07:48.658728 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:48 crc kubenswrapper[4950]: E0318 20:07:48.759528 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:48 crc kubenswrapper[4950]: E0318 20:07:48.860987 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:48 crc kubenswrapper[4950]: E0318 20:07:48.962189 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:49 crc kubenswrapper[4950]: E0318 20:07:49.064612 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:49 crc kubenswrapper[4950]: E0318 20:07:49.165322 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:49 crc kubenswrapper[4950]: E0318 20:07:49.265509 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:49 crc kubenswrapper[4950]: E0318 20:07:49.366682 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:49 crc kubenswrapper[4950]: E0318 20:07:49.467830 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:49 crc kubenswrapper[4950]: E0318 20:07:49.568689 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:49 crc kubenswrapper[4950]: E0318 20:07:49.669551 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:49 crc kubenswrapper[4950]: E0318 20:07:49.770475 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:49 crc kubenswrapper[4950]: E0318 20:07:49.871641 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:49 crc kubenswrapper[4950]: E0318 20:07:49.972599 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:50 crc kubenswrapper[4950]: E0318 20:07:50.073520 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:50 crc kubenswrapper[4950]: E0318 20:07:50.173643 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:50 crc kubenswrapper[4950]: E0318 20:07:50.274873 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:50 crc kubenswrapper[4950]: E0318 20:07:50.375587 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:50 crc kubenswrapper[4950]: E0318 20:07:50.476888 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:50 crc kubenswrapper[4950]: E0318 20:07:50.577755 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:50 crc kubenswrapper[4950]: E0318 20:07:50.678897 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:50 crc kubenswrapper[4950]: E0318 20:07:50.780199 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:50 crc kubenswrapper[4950]: E0318 20:07:50.880747 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:50 crc kubenswrapper[4950]: E0318 20:07:50.982317 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:51 crc kubenswrapper[4950]: E0318 20:07:51.082957 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:51 crc kubenswrapper[4950]: E0318 20:07:51.183086 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:51 crc kubenswrapper[4950]: E0318 20:07:51.283340 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:51 crc kubenswrapper[4950]: E0318 20:07:51.385657 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:51 crc kubenswrapper[4950]: E0318 20:07:51.486600 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:51 crc kubenswrapper[4950]: E0318 20:07:51.587853 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:51 crc kubenswrapper[4950]: E0318 20:07:51.688951 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:51 crc kubenswrapper[4950]: E0318 20:07:51.789801 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:51 crc kubenswrapper[4950]: E0318 20:07:51.889912 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:51 crc kubenswrapper[4950]: E0318 20:07:51.991020 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:52 crc kubenswrapper[4950]: E0318 20:07:52.092331 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:52 crc kubenswrapper[4950]: E0318 20:07:52.193823 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:52 crc kubenswrapper[4950]: E0318 20:07:52.294547 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:52 crc kubenswrapper[4950]: E0318 20:07:52.396185 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:52 crc kubenswrapper[4950]: E0318 20:07:52.496658 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:52 crc kubenswrapper[4950]: E0318 20:07:52.597270 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:52 crc kubenswrapper[4950]: E0318 20:07:52.698173 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:52 crc kubenswrapper[4950]: E0318 20:07:52.799161 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:52 crc kubenswrapper[4950]: E0318 20:07:52.899797 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:53 crc kubenswrapper[4950]: E0318 20:07:53.000471 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:53 crc kubenswrapper[4950]: E0318 20:07:53.101140 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:53 crc kubenswrapper[4950]: E0318 20:07:53.201249 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:53 crc kubenswrapper[4950]: E0318 20:07:53.301358 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:53 crc kubenswrapper[4950]: E0318 20:07:53.401823 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:53 crc kubenswrapper[4950]: E0318 20:07:53.502720 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:53 crc kubenswrapper[4950]: E0318 20:07:53.603832 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:53 crc kubenswrapper[4950]: E0318 20:07:53.704943 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:53 crc kubenswrapper[4950]: E0318 20:07:53.805813 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:53 crc kubenswrapper[4950]: E0318 20:07:53.906897 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:54 crc kubenswrapper[4950]: E0318 20:07:54.007504 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:54 crc kubenswrapper[4950]: E0318 20:07:54.108640 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:54 crc kubenswrapper[4950]: E0318 20:07:54.209263 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:54 crc kubenswrapper[4950]: E0318 20:07:54.310022 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:54 crc kubenswrapper[4950]: E0318 20:07:54.410717 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:54 crc kubenswrapper[4950]: E0318 20:07:54.511289 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:54 crc kubenswrapper[4950]: E0318 20:07:54.612690 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:54 crc kubenswrapper[4950]: E0318 20:07:54.713113 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:54 crc kubenswrapper[4950]: E0318 20:07:54.814032 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:54 crc kubenswrapper[4950]: E0318 20:07:54.915236 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:55 crc kubenswrapper[4950]: E0318 20:07:55.015385 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:55 crc kubenswrapper[4950]: E0318 20:07:55.115969 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:55 crc kubenswrapper[4950]: E0318 20:07:55.216897 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:55 crc kubenswrapper[4950]: E0318 20:07:55.317322 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:55 crc kubenswrapper[4950]: E0318 20:07:55.418450 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:55 crc kubenswrapper[4950]: E0318 20:07:55.518804 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:55 crc kubenswrapper[4950]: E0318 20:07:55.619512 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:55 crc kubenswrapper[4950]: E0318 20:07:55.720445 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:55 crc kubenswrapper[4950]: E0318 20:07:55.820895 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:55 crc kubenswrapper[4950]: E0318 20:07:55.922011 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:56 crc kubenswrapper[4950]: E0318 20:07:56.022966 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:56 crc kubenswrapper[4950]: E0318 20:07:56.123466 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:56 crc kubenswrapper[4950]: E0318 20:07:56.224337 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:56 crc kubenswrapper[4950]: E0318 20:07:56.325370 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:56 crc kubenswrapper[4950]: E0318 20:07:56.426312 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:56 crc kubenswrapper[4950]: E0318 20:07:56.527464 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:56 crc kubenswrapper[4950]: E0318 20:07:56.628205 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:56 crc kubenswrapper[4950]: E0318 20:07:56.729284 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:56 crc kubenswrapper[4950]: E0318 20:07:56.830232 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:56 crc kubenswrapper[4950]: E0318 20:07:56.931283 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:57 crc kubenswrapper[4950]: E0318 20:07:57.032184 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:57 crc kubenswrapper[4950]: E0318 20:07:57.133424 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:57 crc kubenswrapper[4950]: E0318 20:07:57.233751 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:57 crc kubenswrapper[4950]: E0318 20:07:57.334843 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:57 crc kubenswrapper[4950]: E0318 20:07:57.436046 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:57 crc kubenswrapper[4950]: I0318 20:07:57.479444 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:57 crc kubenswrapper[4950]: I0318 20:07:57.481119 4950 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 18 20:07:57 crc kubenswrapper[4950]: I0318 20:07:57.481919 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:57 crc kubenswrapper[4950]: I0318 20:07:57.481982 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:57 crc kubenswrapper[4950]: I0318 20:07:57.482001 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:57 crc kubenswrapper[4950]: I0318 20:07:57.482908 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:57 crc kubenswrapper[4950]: I0318 20:07:57.483134 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:57 crc kubenswrapper[4950]: I0318 20:07:57.483317 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:57 crc kubenswrapper[4950]: I0318 20:07:57.484984 4950 scope.go:117] "RemoveContainer" containerID="7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb" Mar 18 20:07:57 crc kubenswrapper[4950]: E0318 20:07:57.485471 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 20:07:57 crc kubenswrapper[4950]: E0318 20:07:57.537106 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:57 crc kubenswrapper[4950]: E0318 20:07:57.539538 4950 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 18 20:07:57 crc kubenswrapper[4950]: E0318 20:07:57.638273 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:57 crc kubenswrapper[4950]: E0318 20:07:57.739348 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:57 crc kubenswrapper[4950]: E0318 20:07:57.840217 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:57 crc kubenswrapper[4950]: E0318 20:07:57.941665 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:58 crc kubenswrapper[4950]: E0318 20:07:58.042645 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:58 crc kubenswrapper[4950]: E0318 20:07:58.143882 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:58 crc kubenswrapper[4950]: E0318 20:07:58.244587 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:58 crc kubenswrapper[4950]: E0318 20:07:58.344710 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:58 crc kubenswrapper[4950]: E0318 20:07:58.445506 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:58 crc kubenswrapper[4950]: E0318 20:07:58.518300 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 18 20:07:58 crc kubenswrapper[4950]: I0318 20:07:58.521240 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:58 crc kubenswrapper[4950]: I0318 20:07:58.521363 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:58 crc kubenswrapper[4950]: I0318 20:07:58.521534 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:58 crc kubenswrapper[4950]: I0318 20:07:58.521705 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:07:58 crc kubenswrapper[4950]: I0318 20:07:58.521865 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:07:58Z","lastTransitionTime":"2026-03-18T20:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:07:58 crc kubenswrapper[4950]: E0318 20:07:58.529955 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:07:58 crc kubenswrapper[4950]: I0318 20:07:58.532937 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:58 crc kubenswrapper[4950]: I0318 20:07:58.533052 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:58 crc kubenswrapper[4950]: I0318 20:07:58.533146 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:58 crc kubenswrapper[4950]: I0318 20:07:58.533235 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:07:58 crc kubenswrapper[4950]: I0318 20:07:58.533314 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:07:58Z","lastTransitionTime":"2026-03-18T20:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:07:58 crc kubenswrapper[4950]: E0318 20:07:58.542351 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:07:58 crc kubenswrapper[4950]: I0318 20:07:58.546231 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:58 crc kubenswrapper[4950]: I0318 20:07:58.546400 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:58 crc kubenswrapper[4950]: I0318 20:07:58.546549 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:58 crc kubenswrapper[4950]: I0318 20:07:58.546661 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:07:58 crc kubenswrapper[4950]: I0318 20:07:58.546770 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:07:58Z","lastTransitionTime":"2026-03-18T20:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:07:58 crc kubenswrapper[4950]: E0318 20:07:58.556683 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:07:58 crc kubenswrapper[4950]: I0318 20:07:58.560324 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:07:58 crc kubenswrapper[4950]: I0318 20:07:58.560356 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:07:58 crc kubenswrapper[4950]: I0318 20:07:58.560366 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:07:58 crc kubenswrapper[4950]: I0318 20:07:58.560384 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:07:58 crc kubenswrapper[4950]: I0318 20:07:58.560394 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:07:58Z","lastTransitionTime":"2026-03-18T20:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:07:58 crc kubenswrapper[4950]: E0318 20:07:58.571385 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:07:58 crc kubenswrapper[4950]: E0318 20:07:58.571587 4950 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 20:07:58 crc kubenswrapper[4950]: E0318 20:07:58.571613 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:58 crc kubenswrapper[4950]: E0318 20:07:58.671926 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:58 crc kubenswrapper[4950]: E0318 20:07:58.772512 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:58 crc kubenswrapper[4950]: E0318 20:07:58.872823 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:58 crc kubenswrapper[4950]: E0318 20:07:58.973776 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:59 crc kubenswrapper[4950]: E0318 20:07:59.074780 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:59 crc kubenswrapper[4950]: E0318 20:07:59.176143 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:59 crc kubenswrapper[4950]: E0318 20:07:59.277342 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:59 crc kubenswrapper[4950]: E0318 20:07:59.377892 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:59 crc kubenswrapper[4950]: E0318 20:07:59.479334 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:59 crc kubenswrapper[4950]: E0318 20:07:59.580118 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:59 crc kubenswrapper[4950]: E0318 20:07:59.680867 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:59 crc kubenswrapper[4950]: E0318 20:07:59.781978 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:59 crc kubenswrapper[4950]: E0318 20:07:59.882831 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:07:59 crc kubenswrapper[4950]: E0318 20:07:59.983654 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:00 crc kubenswrapper[4950]: E0318 20:08:00.084125 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:00 crc kubenswrapper[4950]: E0318 20:08:00.184236 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:00 crc kubenswrapper[4950]: E0318 20:08:00.285585 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:00 crc kubenswrapper[4950]: E0318 20:08:00.386756 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:00 crc kubenswrapper[4950]: E0318 20:08:00.487929 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:00 crc kubenswrapper[4950]: E0318 20:08:00.588378 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:00 crc kubenswrapper[4950]: E0318 20:08:00.688881 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:00 crc kubenswrapper[4950]: E0318 20:08:00.789864 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:00 crc kubenswrapper[4950]: E0318 20:08:00.890807 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:00 crc kubenswrapper[4950]: E0318 20:08:00.991745 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:01 crc kubenswrapper[4950]: E0318 20:08:01.091964 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:01 crc kubenswrapper[4950]: E0318 20:08:01.192517 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:01 crc kubenswrapper[4950]: E0318 20:08:01.293495 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:01 crc kubenswrapper[4950]: E0318 20:08:01.394655 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:01 crc kubenswrapper[4950]: E0318 20:08:01.496404 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:01 crc kubenswrapper[4950]: E0318 20:08:01.597618 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:01 crc kubenswrapper[4950]: E0318 20:08:01.698732 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:01 crc kubenswrapper[4950]: E0318 20:08:01.799085 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:01 crc kubenswrapper[4950]: E0318 20:08:01.899358 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:01 crc kubenswrapper[4950]: E0318 20:08:01.999707 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:02 crc kubenswrapper[4950]: E0318 20:08:02.100267 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:02 crc kubenswrapper[4950]: E0318 20:08:02.201244 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:02 crc kubenswrapper[4950]: E0318 20:08:02.301457 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:02 crc kubenswrapper[4950]: E0318 20:08:02.401786 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:02 crc kubenswrapper[4950]: E0318 20:08:02.502520 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:02 crc kubenswrapper[4950]: E0318 20:08:02.603453 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:02 crc kubenswrapper[4950]: E0318 20:08:02.704175 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:02 crc kubenswrapper[4950]: E0318 20:08:02.805028 4950 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 18 20:08:02 crc kubenswrapper[4950]: I0318 20:08:02.805754 4950 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 18 20:08:02 crc kubenswrapper[4950]: I0318 20:08:02.893801 4950 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 18 20:08:02 crc kubenswrapper[4950]: I0318 20:08:02.908311 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:02 crc kubenswrapper[4950]: I0318 20:08:02.908398 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:02 crc kubenswrapper[4950]: I0318 20:08:02.908466 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:02 crc kubenswrapper[4950]: I0318 20:08:02.908492 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:02 crc kubenswrapper[4950]: I0318 20:08:02.908512 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:02Z","lastTransitionTime":"2026-03-18T20:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.012023 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.012144 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.012162 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.012221 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.012241 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:03Z","lastTransitionTime":"2026-03-18T20:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.115786 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.115839 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.115860 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.115894 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.115911 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:03Z","lastTransitionTime":"2026-03-18T20:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.219248 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.219288 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.219304 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.219326 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.219343 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:03Z","lastTransitionTime":"2026-03-18T20:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.322187 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.322664 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.322850 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.323083 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.323269 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:03Z","lastTransitionTime":"2026-03-18T20:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.409986 4950 apiserver.go:52] "Watching apiserver" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.420329 4950 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.421103 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-image-registry/node-ca-7bhrr","openshift-machine-config-operator/machine-config-daemon-k2n6r","openshift-multus/multus-additional-cni-plugins-pwmzl","openshift-multus/multus-lqkxm","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-operator/iptables-alerter-4ln5h","openshift-dns/node-resolver-vl9h2","openshift-ovn-kubernetes/ovnkube-node-p2zdp","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.421718 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.422077 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:03 crc kubenswrapper[4950]: E0318 20:08:03.422189 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.423044 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.424007 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.424195 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7bhrr" Mar 18 20:08:03 crc kubenswrapper[4950]: E0318 20:08:03.424362 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.424507 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.425294 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:03 crc kubenswrapper[4950]: E0318 20:08:03.425516 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.425444 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.426956 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.427131 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.427264 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.427398 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.427578 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:03Z","lastTransitionTime":"2026-03-18T20:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.432746 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.432747 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.433097 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.432835 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.432920 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.432953 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.432970 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.433029 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.433462 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.433592 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.433761 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.433815 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.433908 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.434031 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.434040 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.434231 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.434298 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.434576 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.435574 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.435683 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.436009 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-vl9h2" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.440251 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.440299 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.440592 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.445942 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.446013 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.446402 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.446607 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.446880 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.446918 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.447043 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.447519 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.447521 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.447676 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.447920 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.448135 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.448576 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.448657 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.448850 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.475404 4950 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.479675 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.492372 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.510993 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.530596 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.531577 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.531625 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.531644 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.531668 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.531688 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:03Z","lastTransitionTime":"2026-03-18T20:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.551379 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.557717 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.557767 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.557790 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.557808 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.557829 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.557845 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.557861 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.557876 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.557904 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.557919 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.557932 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.557947 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.557965 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.557979 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.557993 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558012 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558030 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558051 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558065 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558080 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558123 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558139 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558159 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558176 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558193 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558209 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558225 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558247 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558263 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558278 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558295 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558310 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558325 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558362 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558379 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558395 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558446 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558470 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558490 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558518 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558541 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558563 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558585 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558604 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558621 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558638 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558656 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558674 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558706 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558731 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558753 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558774 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558795 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558811 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558830 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558845 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558861 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558878 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558895 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558909 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558926 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558939 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558954 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558971 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.558989 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.559004 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.559019 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.559035 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.559055 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.559069 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.559086 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.559102 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.559133 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.559159 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.559181 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.559200 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.559228 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.559255 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.559277 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.559254 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.559298 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.559575 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.559890 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.560134 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.560452 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.560666 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.560834 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.560837 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.560933 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.560996 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.561022 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.561068 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.561094 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.561117 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.561122 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.561163 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.561190 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.561283 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.561329 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.561354 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.561617 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.561650 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.561736 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.561749 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.561798 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.561827 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.561873 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.561903 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.561931 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.561956 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.561985 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562009 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562036 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562062 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562089 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562117 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562142 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562171 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562198 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562254 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562293 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562319 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562347 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562373 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562446 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562476 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562500 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562526 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562552 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562578 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562607 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562636 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562662 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562688 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562713 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562744 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562769 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562794 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562817 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562841 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562868 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562892 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562916 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562943 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562968 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.562993 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563019 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563044 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563069 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563074 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563093 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563122 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563146 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563172 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563196 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563221 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563249 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563274 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563299 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563323 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563356 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563381 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563385 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563452 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563427 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563548 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563590 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563663 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563682 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563693 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563764 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563917 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563957 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.563978 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564024 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564083 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564117 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564147 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564174 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564201 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564288 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564316 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564344 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564353 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564370 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564397 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564447 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564473 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564499 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564523 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564550 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564576 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564602 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564625 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564650 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564677 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564701 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564729 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564755 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564770 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564784 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564848 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564951 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564998 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565037 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565078 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565120 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565215 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565258 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565252 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565296 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565340 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565380 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565442 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565481 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565604 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-var-lib-openvswitch\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565644 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-cni-bin\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565680 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-multus-socket-dir-parent\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565727 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565764 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-systemd-units\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565799 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/83ee6659-cfde-4844-bb92-ccb9b26451ea-ovnkube-script-lib\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565842 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-system-cni-dir\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565878 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-host-run-multus-certs\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565912 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-slash\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565944 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-run-systemd\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565978 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-host-var-lib-cni-multus\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566015 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/981403b4-25e5-4ea0-9267-9584719853d6-multus-daemon-config\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566051 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-etc-kubernetes\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566084 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4048b439-3266-46e7-9de0-22377efacc46-rootfs\") pod \"machine-config-daemon-k2n6r\" (UID: \"4048b439-3266-46e7-9de0-22377efacc46\") " pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566118 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-log-socket\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566155 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9a08cee8-66e8-4cf0-b3ea-0c578a2438f7-hosts-file\") pod \"node-resolver-vl9h2\" (UID: \"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\") " pod="openshift-dns/node-resolver-vl9h2" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566186 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/22f8458a-e595-4fe8-98ac-b82d0c64d130-serviceca\") pod \"node-ca-7bhrr\" (UID: \"22f8458a-e595-4fe8-98ac-b82d0c64d130\") " pod="openshift-image-registry/node-ca-7bhrr" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566244 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1b13cdc2-ca84-4da5-9a60-5619987e6cef-cnibin\") pod \"multus-additional-cni-plugins-pwmzl\" (UID: \"1b13cdc2-ca84-4da5-9a60-5619987e6cef\") " pod="openshift-multus/multus-additional-cni-plugins-pwmzl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566285 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566323 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566361 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5pdc\" (UniqueName: \"kubernetes.io/projected/9a08cee8-66e8-4cf0-b3ea-0c578a2438f7-kube-api-access-z5pdc\") pod \"node-resolver-vl9h2\" (UID: \"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\") " pod="openshift-dns/node-resolver-vl9h2" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566433 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4048b439-3266-46e7-9de0-22377efacc46-mcd-auth-proxy-config\") pod \"machine-config-daemon-k2n6r\" (UID: \"4048b439-3266-46e7-9de0-22377efacc46\") " pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566474 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-etc-openvswitch\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566516 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566550 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1b13cdc2-ca84-4da5-9a60-5619987e6cef-system-cni-dir\") pod \"multus-additional-cni-plugins-pwmzl\" (UID: \"1b13cdc2-ca84-4da5-9a60-5619987e6cef\") " pod="openshift-multus/multus-additional-cni-plugins-pwmzl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566611 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566642 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-run-netns\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566699 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-node-log\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566732 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/83ee6659-cfde-4844-bb92-ccb9b26451ea-ovnkube-config\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566763 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/83ee6659-cfde-4844-bb92-ccb9b26451ea-env-overrides\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566804 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566843 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1b13cdc2-ca84-4da5-9a60-5619987e6cef-os-release\") pod \"multus-additional-cni-plugins-pwmzl\" (UID: \"1b13cdc2-ca84-4da5-9a60-5619987e6cef\") " pod="openshift-multus/multus-additional-cni-plugins-pwmzl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566874 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-cnibin\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566906 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-run-ovn-kubernetes\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566936 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-os-release\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566965 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-host-var-lib-cni-bin\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566994 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567019 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-cni-netd\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567044 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/22f8458a-e595-4fe8-98ac-b82d0c64d130-host\") pod \"node-ca-7bhrr\" (UID: \"22f8458a-e595-4fe8-98ac-b82d0c64d130\") " pod="openshift-image-registry/node-ca-7bhrr" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567068 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpmlc\" (UniqueName: \"kubernetes.io/projected/22f8458a-e595-4fe8-98ac-b82d0c64d130-kube-api-access-qpmlc\") pod \"node-ca-7bhrr\" (UID: \"22f8458a-e595-4fe8-98ac-b82d0c64d130\") " pod="openshift-image-registry/node-ca-7bhrr" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567090 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1b13cdc2-ca84-4da5-9a60-5619987e6cef-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pwmzl\" (UID: \"1b13cdc2-ca84-4da5-9a60-5619987e6cef\") " pod="openshift-multus/multus-additional-cni-plugins-pwmzl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567117 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567140 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-run-ovn\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567161 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-multus-cni-dir\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567182 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-host-run-netns\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567203 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-hostroot\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567227 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tkc4\" (UniqueName: \"kubernetes.io/projected/4048b439-3266-46e7-9de0-22377efacc46-kube-api-access-5tkc4\") pod \"machine-config-daemon-k2n6r\" (UID: \"4048b439-3266-46e7-9de0-22377efacc46\") " pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567250 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hn5f\" (UniqueName: \"kubernetes.io/projected/981403b4-25e5-4ea0-9267-9584719853d6-kube-api-access-6hn5f\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567272 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4048b439-3266-46e7-9de0-22377efacc46-proxy-tls\") pod \"machine-config-daemon-k2n6r\" (UID: \"4048b439-3266-46e7-9de0-22377efacc46\") " pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567300 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndb6n\" (UniqueName: \"kubernetes.io/projected/1b13cdc2-ca84-4da5-9a60-5619987e6cef-kube-api-access-ndb6n\") pod \"multus-additional-cni-plugins-pwmzl\" (UID: \"1b13cdc2-ca84-4da5-9a60-5619987e6cef\") " pod="openshift-multus/multus-additional-cni-plugins-pwmzl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567322 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-host-var-lib-kubelet\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567363 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567388 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-kubelet\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567438 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-run-openvswitch\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567476 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8qvm\" (UniqueName: \"kubernetes.io/projected/83ee6659-cfde-4844-bb92-ccb9b26451ea-kube-api-access-l8qvm\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567513 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567546 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567579 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/83ee6659-cfde-4844-bb92-ccb9b26451ea-ovn-node-metrics-cert\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567604 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1b13cdc2-ca84-4da5-9a60-5619987e6cef-cni-binary-copy\") pod \"multus-additional-cni-plugins-pwmzl\" (UID: \"1b13cdc2-ca84-4da5-9a60-5619987e6cef\") " pod="openshift-multus/multus-additional-cni-plugins-pwmzl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567631 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1b13cdc2-ca84-4da5-9a60-5619987e6cef-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pwmzl\" (UID: \"1b13cdc2-ca84-4da5-9a60-5619987e6cef\") " pod="openshift-multus/multus-additional-cni-plugins-pwmzl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567659 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/981403b4-25e5-4ea0-9267-9584719853d6-cni-binary-copy\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567689 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-host-run-k8s-cni-cncf-io\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567718 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-multus-conf-dir\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567747 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567780 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567805 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567830 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567914 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567936 4950 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567952 4950 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567968 4950 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.567985 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.568002 4950 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.568018 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.568034 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.568051 4950 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.568065 4950 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.568081 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.568095 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.568111 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.568127 4950 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.568142 4950 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.568158 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.568173 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.568187 4950 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.568203 4950 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.578048 4950 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.564944 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565056 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.565104 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.566689 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.568691 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.582890 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.582958 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.583061 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.568788 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.569043 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.569057 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.569594 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.569740 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.570045 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.570191 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.570201 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.570550 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.570719 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.571378 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.571608 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.571947 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.572113 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.572258 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.572288 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.573293 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.573394 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.573630 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.573713 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.573767 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.574077 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.574347 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.574622 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.574777 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.574961 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.575032 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.575058 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.575270 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.575266 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.575437 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.575525 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.575606 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.575720 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.575954 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.575983 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.576298 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.576552 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.577147 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.577553 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.577626 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.577864 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.578356 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.578488 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.578632 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.579066 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.579494 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.580726 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.582227 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.582366 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.582488 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.582548 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.582712 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.584019 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.584047 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.584487 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.584535 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.584551 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.584575 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.584681 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.584768 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.584769 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: E0318 20:08:03.584866 4950 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 20:08:03 crc kubenswrapper[4950]: E0318 20:08:03.584924 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:04.084906831 +0000 UTC m=+97.325748699 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.585152 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.585332 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.585978 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.586161 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.586315 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.586580 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.586602 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.586871 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: E0318 20:08:03.586980 4950 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 20:08:03 crc kubenswrapper[4950]: E0318 20:08:03.587025 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:04.087015597 +0000 UTC m=+97.327857465 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.587110 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.587262 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.587680 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.587941 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.588301 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.588562 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.588738 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.589088 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.589763 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.590153 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.590546 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.590711 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.590956 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.592017 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.593751 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.593568 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.594056 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.594173 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.594226 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.594266 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.594362 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.595015 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.600986 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.600583 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.604306 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.604472 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 20:08:03 crc kubenswrapper[4950]: E0318 20:08:03.604621 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 20:08:03 crc kubenswrapper[4950]: E0318 20:08:03.604642 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 20:08:03 crc kubenswrapper[4950]: E0318 20:08:03.604656 4950 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:03 crc kubenswrapper[4950]: E0318 20:08:03.604765 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:04.10471657 +0000 UTC m=+97.345558438 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:03 crc kubenswrapper[4950]: E0318 20:08:03.604869 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 20:08:03 crc kubenswrapper[4950]: E0318 20:08:03.604889 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 20:08:03 crc kubenswrapper[4950]: E0318 20:08:03.604898 4950 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:03 crc kubenswrapper[4950]: E0318 20:08:03.604933 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:04.104924056 +0000 UTC m=+97.345765974 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.604995 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.605598 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.605621 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.605972 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.606050 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.606406 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.606756 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.607031 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.607393 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.607464 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.608185 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.608699 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.608869 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.608915 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.608964 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.609932 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.610217 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.610676 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.610718 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.611065 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.608950 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.612569 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.612797 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.613073 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.613214 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.613546 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.613699 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.613967 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.614094 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.614188 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.614289 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.614572 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.614805 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.615112 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.615237 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.615408 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.615426 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.616900 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.618602 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.619648 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.621240 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.621574 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.621846 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: E0318 20:08:03.621898 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:08:04.121868519 +0000 UTC m=+97.362710407 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.622715 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.622744 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.622819 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.622811 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.624376 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.624994 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.625233 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.625641 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.625942 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.625998 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.626011 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.626639 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.626560 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.627449 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.627950 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.629182 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.630768 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.631836 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.634229 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.634783 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.634812 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.634825 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.634851 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.634864 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:03Z","lastTransitionTime":"2026-03-18T20:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.644257 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.644440 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.650727 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.651503 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.657028 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.668778 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672148 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-cni-netd\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672193 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/22f8458a-e595-4fe8-98ac-b82d0c64d130-host\") pod \"node-ca-7bhrr\" (UID: \"22f8458a-e595-4fe8-98ac-b82d0c64d130\") " pod="openshift-image-registry/node-ca-7bhrr" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672226 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpmlc\" (UniqueName: \"kubernetes.io/projected/22f8458a-e595-4fe8-98ac-b82d0c64d130-kube-api-access-qpmlc\") pod \"node-ca-7bhrr\" (UID: \"22f8458a-e595-4fe8-98ac-b82d0c64d130\") " pod="openshift-image-registry/node-ca-7bhrr" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672260 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1b13cdc2-ca84-4da5-9a60-5619987e6cef-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pwmzl\" (UID: \"1b13cdc2-ca84-4da5-9a60-5619987e6cef\") " pod="openshift-multus/multus-additional-cni-plugins-pwmzl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672309 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-run-ovn\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672335 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-multus-cni-dir\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672359 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-host-run-netns\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672383 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-hostroot\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672469 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tkc4\" (UniqueName: \"kubernetes.io/projected/4048b439-3266-46e7-9de0-22377efacc46-kube-api-access-5tkc4\") pod \"machine-config-daemon-k2n6r\" (UID: \"4048b439-3266-46e7-9de0-22377efacc46\") " pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672496 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hn5f\" (UniqueName: \"kubernetes.io/projected/981403b4-25e5-4ea0-9267-9584719853d6-kube-api-access-6hn5f\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672520 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4048b439-3266-46e7-9de0-22377efacc46-proxy-tls\") pod \"machine-config-daemon-k2n6r\" (UID: \"4048b439-3266-46e7-9de0-22377efacc46\") " pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672540 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndb6n\" (UniqueName: \"kubernetes.io/projected/1b13cdc2-ca84-4da5-9a60-5619987e6cef-kube-api-access-ndb6n\") pod \"multus-additional-cni-plugins-pwmzl\" (UID: \"1b13cdc2-ca84-4da5-9a60-5619987e6cef\") " pod="openshift-multus/multus-additional-cni-plugins-pwmzl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672564 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-host-var-lib-kubelet\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672604 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-kubelet\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672624 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-run-openvswitch\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672647 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8qvm\" (UniqueName: \"kubernetes.io/projected/83ee6659-cfde-4844-bb92-ccb9b26451ea-kube-api-access-l8qvm\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672679 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672704 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/83ee6659-cfde-4844-bb92-ccb9b26451ea-ovn-node-metrics-cert\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672724 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1b13cdc2-ca84-4da5-9a60-5619987e6cef-cni-binary-copy\") pod \"multus-additional-cni-plugins-pwmzl\" (UID: \"1b13cdc2-ca84-4da5-9a60-5619987e6cef\") " pod="openshift-multus/multus-additional-cni-plugins-pwmzl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672763 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1b13cdc2-ca84-4da5-9a60-5619987e6cef-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pwmzl\" (UID: \"1b13cdc2-ca84-4da5-9a60-5619987e6cef\") " pod="openshift-multus/multus-additional-cni-plugins-pwmzl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672789 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/981403b4-25e5-4ea0-9267-9584719853d6-cni-binary-copy\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672816 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-host-run-k8s-cni-cncf-io\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672841 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-multus-conf-dir\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672877 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672901 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672925 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-var-lib-openvswitch\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672951 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-cni-bin\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.672982 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-multus-socket-dir-parent\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673017 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-systemd-units\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673041 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/83ee6659-cfde-4844-bb92-ccb9b26451ea-ovnkube-script-lib\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673067 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-system-cni-dir\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673091 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-host-run-multus-certs\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673117 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-run-systemd\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673144 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-host-var-lib-cni-multus\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673166 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/981403b4-25e5-4ea0-9267-9584719853d6-multus-daemon-config\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673188 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-etc-kubernetes\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673217 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-slash\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673244 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-log-socket\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673272 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4048b439-3266-46e7-9de0-22377efacc46-rootfs\") pod \"machine-config-daemon-k2n6r\" (UID: \"4048b439-3266-46e7-9de0-22377efacc46\") " pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673299 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9a08cee8-66e8-4cf0-b3ea-0c578a2438f7-hosts-file\") pod \"node-resolver-vl9h2\" (UID: \"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\") " pod="openshift-dns/node-resolver-vl9h2" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673333 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/22f8458a-e595-4fe8-98ac-b82d0c64d130-serviceca\") pod \"node-ca-7bhrr\" (UID: \"22f8458a-e595-4fe8-98ac-b82d0c64d130\") " pod="openshift-image-registry/node-ca-7bhrr" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673358 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1b13cdc2-ca84-4da5-9a60-5619987e6cef-cnibin\") pod \"multus-additional-cni-plugins-pwmzl\" (UID: \"1b13cdc2-ca84-4da5-9a60-5619987e6cef\") " pod="openshift-multus/multus-additional-cni-plugins-pwmzl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673399 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5pdc\" (UniqueName: \"kubernetes.io/projected/9a08cee8-66e8-4cf0-b3ea-0c578a2438f7-kube-api-access-z5pdc\") pod \"node-resolver-vl9h2\" (UID: \"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\") " pod="openshift-dns/node-resolver-vl9h2" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673442 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4048b439-3266-46e7-9de0-22377efacc46-mcd-auth-proxy-config\") pod \"machine-config-daemon-k2n6r\" (UID: \"4048b439-3266-46e7-9de0-22377efacc46\") " pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673472 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-etc-openvswitch\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673498 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1b13cdc2-ca84-4da5-9a60-5619987e6cef-system-cni-dir\") pod \"multus-additional-cni-plugins-pwmzl\" (UID: \"1b13cdc2-ca84-4da5-9a60-5619987e6cef\") " pod="openshift-multus/multus-additional-cni-plugins-pwmzl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673523 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-node-log\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673544 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/83ee6659-cfde-4844-bb92-ccb9b26451ea-ovnkube-config\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673569 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/83ee6659-cfde-4844-bb92-ccb9b26451ea-env-overrides\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673594 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1b13cdc2-ca84-4da5-9a60-5619987e6cef-os-release\") pod \"multus-additional-cni-plugins-pwmzl\" (UID: \"1b13cdc2-ca84-4da5-9a60-5619987e6cef\") " pod="openshift-multus/multus-additional-cni-plugins-pwmzl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673622 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-cnibin\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673645 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-run-netns\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673684 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-run-ovn-kubernetes\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673710 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-os-release\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673744 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-host-var-lib-cni-bin\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673875 4950 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673897 4950 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673916 4950 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673930 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673943 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673956 4950 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673971 4950 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673987 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.673999 4950 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674015 4950 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674026 4950 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674038 4950 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674050 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674065 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674077 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674089 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674101 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674116 4950 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674129 4950 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674141 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674153 4950 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674170 4950 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674182 4950 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674195 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674216 4950 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674228 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674240 4950 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674252 4950 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674269 4950 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674290 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674304 4950 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674316 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674332 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674344 4950 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674360 4950 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674372 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674393 4950 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674405 4950 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674436 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674452 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674464 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674482 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674494 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674509 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674521 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674531 4950 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674543 4950 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674558 4950 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674569 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674580 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674595 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674613 4950 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674626 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674638 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674654 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674667 4950 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674682 4950 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674702 4950 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674718 4950 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674730 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674741 4950 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674753 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674769 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674787 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674799 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674813 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674824 4950 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674843 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674854 4950 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674869 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674893 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674906 4950 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674919 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674935 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674947 4950 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674960 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674972 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.674988 4950 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675000 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675012 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675034 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675046 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675062 4950 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675074 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675090 4950 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675106 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675124 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675136 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675152 4950 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675164 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675178 4950 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675193 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675205 4950 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675217 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675230 4950 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675246 4950 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675258 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675270 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675281 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675296 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675308 4950 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675323 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675335 4950 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675351 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675365 4950 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675377 4950 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.675397 4950 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.677135 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.678064 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-multus-socket-dir-parent\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.678189 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-cni-netd\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.678299 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/22f8458a-e595-4fe8-98ac-b82d0c64d130-host\") pod \"node-ca-7bhrr\" (UID: \"22f8458a-e595-4fe8-98ac-b82d0c64d130\") " pod="openshift-image-registry/node-ca-7bhrr" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.678461 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-run-openvswitch\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.678477 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-hostroot\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.678721 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-multus-cni-dir\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.678745 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-host-run-netns\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.679434 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/981403b4-25e5-4ea0-9267-9584719853d6-cni-binary-copy\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.678770 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-host-var-lib-kubelet\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.679558 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-host-run-k8s-cni-cncf-io\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.679630 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.679642 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-multus-conf-dir\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.679861 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.679953 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.680797 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-run-ovn\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.681519 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1b13cdc2-ca84-4da5-9a60-5619987e6cef-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pwmzl\" (UID: \"1b13cdc2-ca84-4da5-9a60-5619987e6cef\") " pod="openshift-multus/multus-additional-cni-plugins-pwmzl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.681616 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-var-lib-openvswitch\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.720916 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-kubelet\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.721483 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1b13cdc2-ca84-4da5-9a60-5619987e6cef-cni-binary-copy\") pod \"multus-additional-cni-plugins-pwmzl\" (UID: \"1b13cdc2-ca84-4da5-9a60-5619987e6cef\") " pod="openshift-multus/multus-additional-cni-plugins-pwmzl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.722057 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.722423 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.722663 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.722797 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.723017 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.723870 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.723358 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.723849 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.723919 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.723115 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.724114 4950 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.724173 4950 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.724231 4950 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.724295 4950 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.724350 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.724421 4950 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.724496 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.724559 4950 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.724619 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.724673 4950 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.724726 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.724790 4950 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.724854 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.724924 4950 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.724982 4950 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725043 4950 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725102 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725166 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725235 4950 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725293 4950 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725323 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725346 4950 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725444 4950 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725455 4950 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725463 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725472 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725481 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725490 4950 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725499 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725508 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725516 4950 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725524 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725532 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725540 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725549 4950 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725558 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725568 4950 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725577 4950 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725586 4950 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725595 4950 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725605 4950 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725615 4950 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725625 4950 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725633 4950 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725642 4950 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725650 4950 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725658 4950 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725666 4950 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725675 4950 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725683 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725692 4950 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725701 4950 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725709 4950 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725717 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725725 4950 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.725131 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1b13cdc2-ca84-4da5-9a60-5619987e6cef-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pwmzl\" (UID: \"1b13cdc2-ca84-4da5-9a60-5619987e6cef\") " pod="openshift-multus/multus-additional-cni-plugins-pwmzl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.724636 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4048b439-3266-46e7-9de0-22377efacc46-proxy-tls\") pod \"machine-config-daemon-k2n6r\" (UID: \"4048b439-3266-46e7-9de0-22377efacc46\") " pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.724690 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.724708 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-cni-bin\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.728527 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.730692 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-systemd-units\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.731170 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/83ee6659-cfde-4844-bb92-ccb9b26451ea-ovnkube-config\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.731229 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1b13cdc2-ca84-4da5-9a60-5619987e6cef-cnibin\") pod \"multus-additional-cni-plugins-pwmzl\" (UID: \"1b13cdc2-ca84-4da5-9a60-5619987e6cef\") " pod="openshift-multus/multus-additional-cni-plugins-pwmzl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.731274 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/83ee6659-cfde-4844-bb92-ccb9b26451ea-ovnkube-script-lib\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.731321 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-system-cni-dir\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.731343 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-host-run-multus-certs\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.731529 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-etc-openvswitch\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.731568 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1b13cdc2-ca84-4da5-9a60-5619987e6cef-system-cni-dir\") pod \"multus-additional-cni-plugins-pwmzl\" (UID: \"1b13cdc2-ca84-4da5-9a60-5619987e6cef\") " pod="openshift-multus/multus-additional-cni-plugins-pwmzl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.731599 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-node-log\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.732095 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4048b439-3266-46e7-9de0-22377efacc46-mcd-auth-proxy-config\") pod \"machine-config-daemon-k2n6r\" (UID: \"4048b439-3266-46e7-9de0-22377efacc46\") " pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.732223 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-run-systemd\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.732240 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-host-var-lib-cni-multus\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.732703 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/981403b4-25e5-4ea0-9267-9584719853d6-multus-daemon-config\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.732738 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-etc-kubernetes\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.732757 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-slash\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.732776 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-log-socket\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.732792 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4048b439-3266-46e7-9de0-22377efacc46-rootfs\") pod \"machine-config-daemon-k2n6r\" (UID: \"4048b439-3266-46e7-9de0-22377efacc46\") " pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.732920 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9a08cee8-66e8-4cf0-b3ea-0c578a2438f7-hosts-file\") pod \"node-resolver-vl9h2\" (UID: \"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\") " pod="openshift-dns/node-resolver-vl9h2" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.732956 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-cnibin\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.733308 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/83ee6659-cfde-4844-bb92-ccb9b26451ea-env-overrides\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.733360 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1b13cdc2-ca84-4da5-9a60-5619987e6cef-os-release\") pod \"multus-additional-cni-plugins-pwmzl\" (UID: \"1b13cdc2-ca84-4da5-9a60-5619987e6cef\") " pod="openshift-multus/multus-additional-cni-plugins-pwmzl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.736892 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/22f8458a-e595-4fe8-98ac-b82d0c64d130-serviceca\") pod \"node-ca-7bhrr\" (UID: \"22f8458a-e595-4fe8-98ac-b82d0c64d130\") " pod="openshift-image-registry/node-ca-7bhrr" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.737495 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-run-netns\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.737541 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-run-ovn-kubernetes\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.737587 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-os-release\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.737620 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/981403b4-25e5-4ea0-9267-9584719853d6-host-var-lib-cni-bin\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.737797 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.743485 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.744784 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpmlc\" (UniqueName: \"kubernetes.io/projected/22f8458a-e595-4fe8-98ac-b82d0c64d130-kube-api-access-qpmlc\") pod \"node-ca-7bhrr\" (UID: \"22f8458a-e595-4fe8-98ac-b82d0c64d130\") " pod="openshift-image-registry/node-ca-7bhrr" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.745782 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hn5f\" (UniqueName: \"kubernetes.io/projected/981403b4-25e5-4ea0-9267-9584719853d6-kube-api-access-6hn5f\") pod \"multus-lqkxm\" (UID: \"981403b4-25e5-4ea0-9267-9584719853d6\") " pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.747654 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8qvm\" (UniqueName: \"kubernetes.io/projected/83ee6659-cfde-4844-bb92-ccb9b26451ea-kube-api-access-l8qvm\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.750551 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.750612 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.750623 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.750639 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.750649 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:03Z","lastTransitionTime":"2026-03-18T20:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.753044 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.756162 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.756882 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndb6n\" (UniqueName: \"kubernetes.io/projected/1b13cdc2-ca84-4da5-9a60-5619987e6cef-kube-api-access-ndb6n\") pod \"multus-additional-cni-plugins-pwmzl\" (UID: \"1b13cdc2-ca84-4da5-9a60-5619987e6cef\") " pod="openshift-multus/multus-additional-cni-plugins-pwmzl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.757616 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/83ee6659-cfde-4844-bb92-ccb9b26451ea-ovn-node-metrics-cert\") pod \"ovnkube-node-p2zdp\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.759012 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tkc4\" (UniqueName: \"kubernetes.io/projected/4048b439-3266-46e7-9de0-22377efacc46-kube-api-access-5tkc4\") pod \"machine-config-daemon-k2n6r\" (UID: \"4048b439-3266-46e7-9de0-22377efacc46\") " pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.763456 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.767031 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5pdc\" (UniqueName: \"kubernetes.io/projected/9a08cee8-66e8-4cf0-b3ea-0c578a2438f7-kube-api-access-z5pdc\") pod \"node-resolver-vl9h2\" (UID: \"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\") " pod="openshift-dns/node-resolver-vl9h2" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.778493 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.782299 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.790734 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:03 crc kubenswrapper[4950]: W0318 20:08:03.797230 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-28a335d746716726fe54048b1a8c3a3ace747d5751bc51c91d8d1e68f20aa72c WatchSource:0}: Error finding container 28a335d746716726fe54048b1a8c3a3ace747d5751bc51c91d8d1e68f20aa72c: Status 404 returned error can't find the container with id 28a335d746716726fe54048b1a8c3a3ace747d5751bc51c91d8d1e68f20aa72c Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.803168 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.815228 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7bhrr" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.826064 4950 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.826087 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.826111 4950 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.826120 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.826129 4950 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.826137 4950 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.826147 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.826155 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.826164 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.826171 4950 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.826181 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.826201 4950 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.826209 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.835093 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.855570 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.855837 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.856592 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.856602 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.856615 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.856623 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:03Z","lastTransitionTime":"2026-03-18T20:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.877178 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-lqkxm" Mar 18 20:08:03 crc kubenswrapper[4950]: W0318 20:08:03.915160 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod981403b4_25e5_4ea0_9267_9584719853d6.slice/crio-6a397045db4e8abfcf7db9e2082641dd38c87abcadf2a23390e7da63693687d3 WatchSource:0}: Error finding container 6a397045db4e8abfcf7db9e2082641dd38c87abcadf2a23390e7da63693687d3: Status 404 returned error can't find the container with id 6a397045db4e8abfcf7db9e2082641dd38c87abcadf2a23390e7da63693687d3 Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.918468 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:03 crc kubenswrapper[4950]: W0318 20:08:03.920911 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b13cdc2_ca84_4da5_9a60_5619987e6cef.slice/crio-a85ec710a4260c5c17f75f8a69e0e898e37b48fc48af0adcfe276c88ec48351a WatchSource:0}: Error finding container a85ec710a4260c5c17f75f8a69e0e898e37b48fc48af0adcfe276c88ec48351a: Status 404 returned error can't find the container with id a85ec710a4260c5c17f75f8a69e0e898e37b48fc48af0adcfe276c88ec48351a Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.936854 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-vl9h2" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.960309 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.960359 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.960371 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.960391 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:03 crc kubenswrapper[4950]: I0318 20:08:03.960421 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:03Z","lastTransitionTime":"2026-03-18T20:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:03 crc kubenswrapper[4950]: W0318 20:08:03.991400 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83ee6659_cfde_4844_bb92_ccb9b26451ea.slice/crio-64ef4fd94ed84bce6bf75deb94b086c35f39f0a5c352ac110845beb6153ade3c WatchSource:0}: Error finding container 64ef4fd94ed84bce6bf75deb94b086c35f39f0a5c352ac110845beb6153ade3c: Status 404 returned error can't find the container with id 64ef4fd94ed84bce6bf75deb94b086c35f39f0a5c352ac110845beb6153ade3c Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.066585 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.066611 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.066619 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.066633 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.066643 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:04Z","lastTransitionTime":"2026-03-18T20:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.110557 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" event={"ID":"1b13cdc2-ca84-4da5-9a60-5619987e6cef","Type":"ContainerStarted","Data":"a85ec710a4260c5c17f75f8a69e0e898e37b48fc48af0adcfe276c88ec48351a"} Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.113655 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerStarted","Data":"64ef4fd94ed84bce6bf75deb94b086c35f39f0a5c352ac110845beb6153ade3c"} Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.116466 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7bhrr" event={"ID":"22f8458a-e595-4fe8-98ac-b82d0c64d130","Type":"ContainerStarted","Data":"588c1c026109ba5cc557f69bbfb710920f98192b2451f84a868da1ec5df9b1d5"} Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.122224 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lqkxm" event={"ID":"981403b4-25e5-4ea0-9267-9584719853d6","Type":"ContainerStarted","Data":"7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3"} Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.122257 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lqkxm" event={"ID":"981403b4-25e5-4ea0-9267-9584719853d6","Type":"ContainerStarted","Data":"6a397045db4e8abfcf7db9e2082641dd38c87abcadf2a23390e7da63693687d3"} Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.124132 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea"} Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.124164 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"79d0593f056dd75477e4db163f62e1494d4a08dc74968cd43053c8850511997a"} Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.125553 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"298513f9c6a12b0c03ab8f376257270d1b6d2f5163cc8d3a390723b500130c93"} Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.127953 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.128048 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.128078 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.128109 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.128137 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:04 crc kubenswrapper[4950]: E0318 20:08:04.128224 4950 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 20:08:04 crc kubenswrapper[4950]: E0318 20:08:04.128272 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:05.128260445 +0000 UTC m=+98.369102313 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 20:08:04 crc kubenswrapper[4950]: E0318 20:08:04.128351 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 20:08:04 crc kubenswrapper[4950]: E0318 20:08:04.128379 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 20:08:04 crc kubenswrapper[4950]: E0318 20:08:04.128396 4950 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:04 crc kubenswrapper[4950]: E0318 20:08:04.128455 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:05.12843904 +0000 UTC m=+98.369280908 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:04 crc kubenswrapper[4950]: E0318 20:08:04.128508 4950 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 20:08:04 crc kubenswrapper[4950]: E0318 20:08:04.128545 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:05.128538732 +0000 UTC m=+98.369380600 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 20:08:04 crc kubenswrapper[4950]: E0318 20:08:04.128554 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 20:08:04 crc kubenswrapper[4950]: E0318 20:08:04.128571 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 20:08:04 crc kubenswrapper[4950]: E0318 20:08:04.128584 4950 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:04 crc kubenswrapper[4950]: E0318 20:08:04.128622 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:05.128612694 +0000 UTC m=+98.369454682 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:04 crc kubenswrapper[4950]: E0318 20:08:04.129234 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:08:05.128870171 +0000 UTC m=+98.369712039 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.129746 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125"} Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.129802 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864"} Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.129814 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"28a335d746716726fe54048b1a8c3a3ace747d5751bc51c91d8d1e68f20aa72c"} Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.131247 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-vl9h2" event={"ID":"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7","Type":"ContainerStarted","Data":"16b6bf505465a74252cfec0d30a20c730d3c3d592e81ec158ec04d76c9f634bf"} Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.135244 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.139930 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerStarted","Data":"a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217"} Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.139981 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerStarted","Data":"d37e8370429940b17fc7d262bce5136bf4a0298e86ff370a93c3ec0de069d005"} Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.153384 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.169921 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.171222 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.171291 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.171304 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.171319 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.171337 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:04Z","lastTransitionTime":"2026-03-18T20:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.181561 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.192291 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.205143 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.228867 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.239507 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.248198 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.259086 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.275943 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.275968 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.275976 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.275991 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.276000 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:04Z","lastTransitionTime":"2026-03-18T20:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.276998 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.283886 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.296594 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.306246 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.317676 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.332023 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.343068 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.353217 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.362234 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.371209 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.378360 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.378387 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.378396 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.378375 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.378427 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.378437 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:04Z","lastTransitionTime":"2026-03-18T20:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.386466 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.394483 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.407894 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.480205 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.480237 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.480246 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.480258 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.480268 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:04Z","lastTransitionTime":"2026-03-18T20:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.582939 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.583231 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.583340 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.583442 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.583536 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:04Z","lastTransitionTime":"2026-03-18T20:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.686121 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.686609 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.686636 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.686666 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.686687 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:04Z","lastTransitionTime":"2026-03-18T20:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.789706 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.790045 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.790264 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.790481 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.790680 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:04Z","lastTransitionTime":"2026-03-18T20:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.893195 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.893254 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.893266 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.893284 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.893295 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:04Z","lastTransitionTime":"2026-03-18T20:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.995632 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.995674 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.995684 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.995699 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:04 crc kubenswrapper[4950]: I0318 20:08:04.995709 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:04Z","lastTransitionTime":"2026-03-18T20:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.097938 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.097970 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.097978 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.097991 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.098032 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:05Z","lastTransitionTime":"2026-03-18T20:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.137478 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.137557 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.137578 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.137598 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.137617 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:05 crc kubenswrapper[4950]: E0318 20:08:05.137672 4950 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 20:08:05 crc kubenswrapper[4950]: E0318 20:08:05.137696 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:08:07.137664606 +0000 UTC m=+100.378506484 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:08:05 crc kubenswrapper[4950]: E0318 20:08:05.137741 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:07.137731738 +0000 UTC m=+100.378573726 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 20:08:05 crc kubenswrapper[4950]: E0318 20:08:05.137930 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 20:08:05 crc kubenswrapper[4950]: E0318 20:08:05.137948 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 20:08:05 crc kubenswrapper[4950]: E0318 20:08:05.138104 4950 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 20:08:05 crc kubenswrapper[4950]: E0318 20:08:05.138120 4950 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:05 crc kubenswrapper[4950]: E0318 20:08:05.138160 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:07.138143179 +0000 UTC m=+100.378985057 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 20:08:05 crc kubenswrapper[4950]: E0318 20:08:05.138034 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 20:08:05 crc kubenswrapper[4950]: E0318 20:08:05.138189 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 20:08:05 crc kubenswrapper[4950]: E0318 20:08:05.138202 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:07.13819045 +0000 UTC m=+100.379032438 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:05 crc kubenswrapper[4950]: E0318 20:08:05.138202 4950 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:05 crc kubenswrapper[4950]: E0318 20:08:05.138244 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:07.138232991 +0000 UTC m=+100.379074869 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.144871 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-vl9h2" event={"ID":"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7","Type":"ContainerStarted","Data":"582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445"} Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.146397 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" event={"ID":"1b13cdc2-ca84-4da5-9a60-5619987e6cef","Type":"ContainerStarted","Data":"e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c"} Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.148941 4950 generic.go:334] "Generic (PLEG): container finished" podID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerID="046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309" exitCode=0 Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.149028 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerDied","Data":"046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309"} Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.151137 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerStarted","Data":"485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7"} Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.155455 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7bhrr" event={"ID":"22f8458a-e595-4fe8-98ac-b82d0c64d130","Type":"ContainerStarted","Data":"6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b"} Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.180520 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.199008 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.201998 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.202029 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.202037 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.202050 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.202058 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:05Z","lastTransitionTime":"2026-03-18T20:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.214515 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.229056 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.241316 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.253261 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.267345 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.282403 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.304762 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.304789 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.304796 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.304811 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.304820 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:05Z","lastTransitionTime":"2026-03-18T20:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.304970 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.341082 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.362531 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.378759 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.396661 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.407472 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.407505 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.407513 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.407526 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.407536 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:05Z","lastTransitionTime":"2026-03-18T20:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.414289 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.428397 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.441252 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.454726 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.466081 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.479021 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.479026 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.479189 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:05 crc kubenswrapper[4950]: E0318 20:08:05.479247 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:05 crc kubenswrapper[4950]: E0318 20:08:05.479160 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:05 crc kubenswrapper[4950]: E0318 20:08:05.479339 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.483242 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.483517 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.484053 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.484937 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.485605 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.486154 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.487600 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.488147 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.489115 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.489802 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.491541 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.492474 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.493247 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.494331 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.495009 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.496143 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.496862 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.498263 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.498855 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.499620 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.503023 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.503229 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.504761 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.505784 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.506376 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.507637 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.508203 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.509487 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.509998 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.510041 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.510052 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.510069 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.510080 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:05Z","lastTransitionTime":"2026-03-18T20:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.510477 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.511905 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.512771 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.513780 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.514331 4950 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.514789 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.517081 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.518192 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.518991 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.520781 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.521906 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.522662 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.523908 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.524778 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.525758 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.526508 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.527635 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.528301 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.532571 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.533243 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.534480 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.535275 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.536824 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.537999 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.538696 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.539119 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.541134 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.541891 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.542973 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.558344 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.576479 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.598773 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.613737 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.613786 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.613797 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.613818 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.613832 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:05Z","lastTransitionTime":"2026-03-18T20:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.716442 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.716487 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.716500 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.716520 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.716533 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:05Z","lastTransitionTime":"2026-03-18T20:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.819686 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.820225 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.820237 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.820258 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.820269 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:05Z","lastTransitionTime":"2026-03-18T20:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.922346 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.922384 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.922396 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.922433 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:05 crc kubenswrapper[4950]: I0318 20:08:05.922448 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:05Z","lastTransitionTime":"2026-03-18T20:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.025230 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.025383 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.025466 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.025533 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.025599 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:06Z","lastTransitionTime":"2026-03-18T20:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.128376 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.128638 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.128746 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.128864 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.128945 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:06Z","lastTransitionTime":"2026-03-18T20:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.160211 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerStarted","Data":"ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8"} Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.160526 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerStarted","Data":"59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f"} Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.160757 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerStarted","Data":"38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8"} Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.160853 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerStarted","Data":"6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1"} Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.160963 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerStarted","Data":"8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61"} Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.161057 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerStarted","Data":"6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2"} Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.164117 4950 generic.go:334] "Generic (PLEG): container finished" podID="1b13cdc2-ca84-4da5-9a60-5619987e6cef" containerID="e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c" exitCode=0 Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.165040 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" event={"ID":"1b13cdc2-ca84-4da5-9a60-5619987e6cef","Type":"ContainerDied","Data":"e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c"} Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.191598 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:06Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.221674 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:06Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.231638 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.231837 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.231924 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.232007 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.232079 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:06Z","lastTransitionTime":"2026-03-18T20:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.238163 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:06Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.253806 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:06Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.267902 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:06Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.287632 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:06Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.313629 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:06Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.329947 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:06Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.334575 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.334615 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.334627 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.334646 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.334657 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:06Z","lastTransitionTime":"2026-03-18T20:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.359455 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:06Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.374834 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:06Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.400691 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:06Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.423477 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:06Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.436548 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.436583 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.436597 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.436612 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.436621 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:06Z","lastTransitionTime":"2026-03-18T20:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.538605 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.538648 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.538658 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.538675 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.538685 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:06Z","lastTransitionTime":"2026-03-18T20:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.645934 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.645973 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.645985 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.646007 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.646018 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:06Z","lastTransitionTime":"2026-03-18T20:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.748837 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.748873 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.748883 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.748899 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.748909 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:06Z","lastTransitionTime":"2026-03-18T20:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.850639 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.850675 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.850746 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.850761 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.850772 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:06Z","lastTransitionTime":"2026-03-18T20:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.952687 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.952735 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.952743 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.952758 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:06 crc kubenswrapper[4950]: I0318 20:08:06.952767 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:06Z","lastTransitionTime":"2026-03-18T20:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.055139 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.055178 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.055190 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.055207 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.055219 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:07Z","lastTransitionTime":"2026-03-18T20:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.156347 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.156542 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:07 crc kubenswrapper[4950]: E0318 20:08:07.156566 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:08:11.156528511 +0000 UTC m=+104.397370439 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.156622 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.156719 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:07 crc kubenswrapper[4950]: E0318 20:08:07.156743 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 20:08:07 crc kubenswrapper[4950]: E0318 20:08:07.156776 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.156779 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:07 crc kubenswrapper[4950]: E0318 20:08:07.156802 4950 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:07 crc kubenswrapper[4950]: E0318 20:08:07.156798 4950 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 20:08:07 crc kubenswrapper[4950]: E0318 20:08:07.156880 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:11.15685452 +0000 UTC m=+104.397696428 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:07 crc kubenswrapper[4950]: E0318 20:08:07.156918 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:11.156901431 +0000 UTC m=+104.397743329 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 20:08:07 crc kubenswrapper[4950]: E0318 20:08:07.156943 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 20:08:07 crc kubenswrapper[4950]: E0318 20:08:07.156972 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 20:08:07 crc kubenswrapper[4950]: E0318 20:08:07.156994 4950 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:07 crc kubenswrapper[4950]: E0318 20:08:07.156996 4950 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 20:08:07 crc kubenswrapper[4950]: E0318 20:08:07.157078 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:11.157055175 +0000 UTC m=+104.397897083 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:07 crc kubenswrapper[4950]: E0318 20:08:07.157111 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:11.157095056 +0000 UTC m=+104.397936954 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.160196 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.160273 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.160312 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.160342 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.160365 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:07Z","lastTransitionTime":"2026-03-18T20:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.171784 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730"} Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.176819 4950 generic.go:334] "Generic (PLEG): container finished" podID="1b13cdc2-ca84-4da5-9a60-5619987e6cef" containerID="d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c" exitCode=0 Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.176869 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" event={"ID":"1b13cdc2-ca84-4da5-9a60-5619987e6cef","Type":"ContainerDied","Data":"d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c"} Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.196854 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.222895 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.242495 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.255012 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.266868 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.266913 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.266926 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.266948 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.266962 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:07Z","lastTransitionTime":"2026-03-18T20:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.271627 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.286730 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.307287 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.325433 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.352681 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.366044 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.372617 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.372656 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.372694 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.372712 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.372723 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:07Z","lastTransitionTime":"2026-03-18T20:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.383348 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.402788 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.416670 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.430724 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.443753 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.452428 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.465199 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.474839 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.474869 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.474879 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.474893 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.474903 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:07Z","lastTransitionTime":"2026-03-18T20:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.474995 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.479184 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.479285 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:07 crc kubenswrapper[4950]: E0318 20:08:07.479333 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:07 crc kubenswrapper[4950]: E0318 20:08:07.479368 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.479398 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:07 crc kubenswrapper[4950]: E0318 20:08:07.479483 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.488622 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.504762 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.524119 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.548884 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.570118 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.578546 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.578608 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.578628 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.578654 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.578672 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:07Z","lastTransitionTime":"2026-03-18T20:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:07 crc kubenswrapper[4950]: E0318 20:08:07.582311 4950 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b13cdc2_ca84_4da5_9a60_5619987e6cef.slice/crio-conmon-f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3.scope\": RecentStats: unable to find data in memory cache]" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.589390 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.606577 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.618919 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.638330 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.651258 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.668575 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.680729 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.680755 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.680763 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.680775 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.680784 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:07Z","lastTransitionTime":"2026-03-18T20:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.702958 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.712942 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.725261 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.735266 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.746648 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.756507 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.769986 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:07Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.782902 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.782934 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.782943 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.782960 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.782971 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:07Z","lastTransitionTime":"2026-03-18T20:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.885990 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.886038 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.886050 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.886068 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.886080 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:07Z","lastTransitionTime":"2026-03-18T20:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.989371 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.989450 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.989463 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.989482 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:07 crc kubenswrapper[4950]: I0318 20:08:07.989496 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:07Z","lastTransitionTime":"2026-03-18T20:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.091665 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.091755 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.091772 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.091827 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.091840 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:08Z","lastTransitionTime":"2026-03-18T20:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.180716 4950 generic.go:334] "Generic (PLEG): container finished" podID="1b13cdc2-ca84-4da5-9a60-5619987e6cef" containerID="f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3" exitCode=0 Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.180774 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" event={"ID":"1b13cdc2-ca84-4da5-9a60-5619987e6cef","Type":"ContainerDied","Data":"f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3"} Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.189523 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerStarted","Data":"50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc"} Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.194089 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.194127 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.194137 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.194154 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.194166 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:08Z","lastTransitionTime":"2026-03-18T20:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.202800 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:08Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.214758 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:08Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.233477 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:08Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.246470 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:08Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.260446 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:08Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.275131 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:08Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.289529 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:08Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.296891 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.296942 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.296959 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.296985 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.297003 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:08Z","lastTransitionTime":"2026-03-18T20:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.309813 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:08Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.322883 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:08Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.336066 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:08Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.353700 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:08Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.369256 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:08Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.398938 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.398982 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.398994 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.399012 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.399025 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:08Z","lastTransitionTime":"2026-03-18T20:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.501311 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.501352 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.501364 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.501383 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.501395 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:08Z","lastTransitionTime":"2026-03-18T20:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.603681 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.603711 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.603722 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.603737 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.603747 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:08Z","lastTransitionTime":"2026-03-18T20:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.710877 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.710912 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.710920 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.710936 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.710944 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:08Z","lastTransitionTime":"2026-03-18T20:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.812614 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.812639 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.812647 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.812659 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.812667 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:08Z","lastTransitionTime":"2026-03-18T20:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.895041 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.895075 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.895085 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.895102 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.895113 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:08Z","lastTransitionTime":"2026-03-18T20:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:08 crc kubenswrapper[4950]: E0318 20:08:08.914352 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:08Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.919843 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.919876 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.919889 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.919928 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.919942 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:08Z","lastTransitionTime":"2026-03-18T20:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:08 crc kubenswrapper[4950]: E0318 20:08:08.945436 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:08Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.949676 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.949709 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.949722 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.949737 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.949750 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:08Z","lastTransitionTime":"2026-03-18T20:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:08 crc kubenswrapper[4950]: E0318 20:08:08.964710 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:08Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.969288 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.969325 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.969337 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.969356 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.969367 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:08Z","lastTransitionTime":"2026-03-18T20:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:08 crc kubenswrapper[4950]: E0318 20:08:08.988638 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:08Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.993617 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.993644 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.993655 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.993672 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:08 crc kubenswrapper[4950]: I0318 20:08:08.993682 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:08Z","lastTransitionTime":"2026-03-18T20:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:09 crc kubenswrapper[4950]: E0318 20:08:09.016895 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: E0318 20:08:09.017160 4950 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.018824 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.018847 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.018857 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.018897 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.018911 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:09Z","lastTransitionTime":"2026-03-18T20:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.045025 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989"] Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.045640 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.049669 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.056142 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.080688 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.081024 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e49a491f-1edc-4a40-b64f-147c0d87481f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-5p989\" (UID: \"e49a491f-1edc-4a40-b64f-147c0d87481f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.081049 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e49a491f-1edc-4a40-b64f-147c0d87481f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-5p989\" (UID: \"e49a491f-1edc-4a40-b64f-147c0d87481f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.081085 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e49a491f-1edc-4a40-b64f-147c0d87481f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-5p989\" (UID: \"e49a491f-1edc-4a40-b64f-147c0d87481f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.081120 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpws6\" (UniqueName: \"kubernetes.io/projected/e49a491f-1edc-4a40-b64f-147c0d87481f-kube-api-access-cpws6\") pod \"ovnkube-control-plane-749d76644c-5p989\" (UID: \"e49a491f-1edc-4a40-b64f-147c0d87481f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.100763 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.121041 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.121068 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.121076 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.121089 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.121098 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:09Z","lastTransitionTime":"2026-03-18T20:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.133517 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.150350 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.165026 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.181004 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.181350 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e49a491f-1edc-4a40-b64f-147c0d87481f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-5p989\" (UID: \"e49a491f-1edc-4a40-b64f-147c0d87481f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.181379 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpws6\" (UniqueName: \"kubernetes.io/projected/e49a491f-1edc-4a40-b64f-147c0d87481f-kube-api-access-cpws6\") pod \"ovnkube-control-plane-749d76644c-5p989\" (UID: \"e49a491f-1edc-4a40-b64f-147c0d87481f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.181439 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e49a491f-1edc-4a40-b64f-147c0d87481f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-5p989\" (UID: \"e49a491f-1edc-4a40-b64f-147c0d87481f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.181459 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e49a491f-1edc-4a40-b64f-147c0d87481f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-5p989\" (UID: \"e49a491f-1edc-4a40-b64f-147c0d87481f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.182029 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e49a491f-1edc-4a40-b64f-147c0d87481f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-5p989\" (UID: \"e49a491f-1edc-4a40-b64f-147c0d87481f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.182065 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e49a491f-1edc-4a40-b64f-147c0d87481f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-5p989\" (UID: \"e49a491f-1edc-4a40-b64f-147c0d87481f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.190195 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e49a491f-1edc-4a40-b64f-147c0d87481f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-5p989\" (UID: \"e49a491f-1edc-4a40-b64f-147c0d87481f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.195828 4950 generic.go:334] "Generic (PLEG): container finished" podID="1b13cdc2-ca84-4da5-9a60-5619987e6cef" containerID="3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330" exitCode=0 Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.195906 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" event={"ID":"1b13cdc2-ca84-4da5-9a60-5619987e6cef","Type":"ContainerDied","Data":"3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330"} Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.202843 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.210807 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpws6\" (UniqueName: \"kubernetes.io/projected/e49a491f-1edc-4a40-b64f-147c0d87481f-kube-api-access-cpws6\") pod \"ovnkube-control-plane-749d76644c-5p989\" (UID: \"e49a491f-1edc-4a40-b64f-147c0d87481f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.217702 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.223384 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.223469 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.223486 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.223510 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.223527 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:09Z","lastTransitionTime":"2026-03-18T20:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.234885 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.245207 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.258355 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.270741 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.282114 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49a491f-1edc-4a40-b64f-147c0d87481f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5p989\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.295066 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.316120 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.326279 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.326317 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.326328 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.326342 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.326353 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:09Z","lastTransitionTime":"2026-03-18T20:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.329313 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.345698 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.357389 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.365574 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.369012 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.382318 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: W0318 20:08:09.389252 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode49a491f_1edc_4a40_b64f_147c0d87481f.slice/crio-7c43151b65ba2dd1ac972a2a9b15bbb49845ca05f6fd97a72542fd1516ef07db WatchSource:0}: Error finding container 7c43151b65ba2dd1ac972a2a9b15bbb49845ca05f6fd97a72542fd1516ef07db: Status 404 returned error can't find the container with id 7c43151b65ba2dd1ac972a2a9b15bbb49845ca05f6fd97a72542fd1516ef07db Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.397423 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.416733 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.428452 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.428486 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.428496 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.428512 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.428523 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:09Z","lastTransitionTime":"2026-03-18T20:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.433366 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.445496 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.457522 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.468909 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49a491f-1edc-4a40-b64f-147c0d87481f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5p989\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:09Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.479961 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:09 crc kubenswrapper[4950]: E0318 20:08:09.480065 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.480303 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:09 crc kubenswrapper[4950]: E0318 20:08:09.480469 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.480645 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:09 crc kubenswrapper[4950]: E0318 20:08:09.480702 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.532861 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.532886 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.532895 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.532909 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.532918 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:09Z","lastTransitionTime":"2026-03-18T20:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.635255 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.635296 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.635307 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.635338 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.635351 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:09Z","lastTransitionTime":"2026-03-18T20:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.738963 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.739467 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.739488 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.739513 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.739535 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:09Z","lastTransitionTime":"2026-03-18T20:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.841978 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.842008 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.842016 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.842028 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.842039 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:09Z","lastTransitionTime":"2026-03-18T20:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.943897 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.943928 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.943937 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.943951 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:09 crc kubenswrapper[4950]: I0318 20:08:09.943959 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:09Z","lastTransitionTime":"2026-03-18T20:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.045824 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.045847 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.045856 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.045870 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.045878 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:10Z","lastTransitionTime":"2026-03-18T20:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.147821 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-l2cjn"] Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.148209 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:10 crc kubenswrapper[4950]: E0318 20:08:10.148260 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.153523 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.153567 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.153581 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.153599 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.153611 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:10Z","lastTransitionTime":"2026-03-18T20:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.189175 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs\") pod \"network-metrics-daemon-l2cjn\" (UID: \"d09a228c-313f-4b55-a298-51ebd404cfd5\") " pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.189241 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q76hg\" (UniqueName: \"kubernetes.io/projected/d09a228c-313f-4b55-a298-51ebd404cfd5-kube-api-access-q76hg\") pod \"network-metrics-daemon-l2cjn\" (UID: \"d09a228c-313f-4b55-a298-51ebd404cfd5\") " pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.200914 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.201923 4950 generic.go:334] "Generic (PLEG): container finished" podID="1b13cdc2-ca84-4da5-9a60-5619987e6cef" containerID="d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1" exitCode=0 Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.201960 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" event={"ID":"1b13cdc2-ca84-4da5-9a60-5619987e6cef","Type":"ContainerDied","Data":"d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1"} Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.241683 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerStarted","Data":"4ef1b45dadc92fc6f9c3a122ca974a406c18cc8259a5fd5597b4e109a9a62afc"} Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.242376 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.242461 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.242492 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.248563 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.254966 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" event={"ID":"e49a491f-1edc-4a40-b64f-147c0d87481f","Type":"ContainerStarted","Data":"122df2a4bd1b73f3890ecaef0f5840c16b127b2b7a8b749ee84f2947334b6aa0"} Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.254998 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" event={"ID":"e49a491f-1edc-4a40-b64f-147c0d87481f","Type":"ContainerStarted","Data":"f1b01d79ecfba0d5984e7415ca4db790ef1635698da9792b6f1e6161a868f243"} Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.255009 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" event={"ID":"e49a491f-1edc-4a40-b64f-147c0d87481f","Type":"ContainerStarted","Data":"7c43151b65ba2dd1ac972a2a9b15bbb49845ca05f6fd97a72542fd1516ef07db"} Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.255426 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.255445 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.255453 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.255481 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.255491 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:10Z","lastTransitionTime":"2026-03-18T20:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.273778 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.277664 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.277838 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.288992 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d09a228c-313f-4b55-a298-51ebd404cfd5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2cjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.290861 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs\") pod \"network-metrics-daemon-l2cjn\" (UID: \"d09a228c-313f-4b55-a298-51ebd404cfd5\") " pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.291009 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q76hg\" (UniqueName: \"kubernetes.io/projected/d09a228c-313f-4b55-a298-51ebd404cfd5-kube-api-access-q76hg\") pod \"network-metrics-daemon-l2cjn\" (UID: \"d09a228c-313f-4b55-a298-51ebd404cfd5\") " pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:10 crc kubenswrapper[4950]: E0318 20:08:10.292099 4950 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 20:08:10 crc kubenswrapper[4950]: E0318 20:08:10.292138 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs podName:d09a228c-313f-4b55-a298-51ebd404cfd5 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:10.792126578 +0000 UTC m=+104.032968446 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs") pod "network-metrics-daemon-l2cjn" (UID: "d09a228c-313f-4b55-a298-51ebd404cfd5") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.302780 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.316515 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.316936 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q76hg\" (UniqueName: \"kubernetes.io/projected/d09a228c-313f-4b55-a298-51ebd404cfd5-kube-api-access-q76hg\") pod \"network-metrics-daemon-l2cjn\" (UID: \"d09a228c-313f-4b55-a298-51ebd404cfd5\") " pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.335215 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.349933 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.357809 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.357845 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.357857 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.357874 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.357886 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:10Z","lastTransitionTime":"2026-03-18T20:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.362662 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49a491f-1edc-4a40-b64f-147c0d87481f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5p989\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.377076 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.395396 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.409391 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.428792 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.456196 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.459657 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.459695 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.459708 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.459722 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.459733 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:10Z","lastTransitionTime":"2026-03-18T20:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.485131 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef1b45dadc92fc6f9c3a122ca974a406c18cc8259a5fd5597b4e109a9a62afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.495682 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.495863 4950 scope.go:117] "RemoveContainer" containerID="7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb" Mar 18 20:08:10 crc kubenswrapper[4950]: E0318 20:08:10.496054 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.505883 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.523469 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.533744 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d09a228c-313f-4b55-a298-51ebd404cfd5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2cjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.545588 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.556876 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.562170 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.562328 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.562395 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.562520 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.562596 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:10Z","lastTransitionTime":"2026-03-18T20:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.572034 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.587077 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.608197 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.623878 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.638561 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.654350 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.665127 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.665368 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.665525 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.665641 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.665740 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:10Z","lastTransitionTime":"2026-03-18T20:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.668512 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.681641 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49a491f-1edc-4a40-b64f-147c0d87481f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b01d79ecfba0d5984e7415ca4db790ef1635698da9792b6f1e6161a868f243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122df2a4bd1b73f3890ecaef0f5840c16b127b2b7a8b749ee84f2947334b6aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5p989\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:10Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.768736 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.769022 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.769148 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.769245 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.769340 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:10Z","lastTransitionTime":"2026-03-18T20:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.796775 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs\") pod \"network-metrics-daemon-l2cjn\" (UID: \"d09a228c-313f-4b55-a298-51ebd404cfd5\") " pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:10 crc kubenswrapper[4950]: E0318 20:08:10.797097 4950 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 20:08:10 crc kubenswrapper[4950]: E0318 20:08:10.797226 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs podName:d09a228c-313f-4b55-a298-51ebd404cfd5 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:11.797212319 +0000 UTC m=+105.038054187 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs") pod "network-metrics-daemon-l2cjn" (UID: "d09a228c-313f-4b55-a298-51ebd404cfd5") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.872103 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.872302 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.872389 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.872512 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.872726 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:10Z","lastTransitionTime":"2026-03-18T20:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.974941 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.975148 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.975213 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.975272 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:10 crc kubenswrapper[4950]: I0318 20:08:10.975326 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:10Z","lastTransitionTime":"2026-03-18T20:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.078493 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.078550 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.078568 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.078593 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.078610 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:11Z","lastTransitionTime":"2026-03-18T20:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.182634 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.182706 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.182727 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.182755 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.182774 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:11Z","lastTransitionTime":"2026-03-18T20:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.201088 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.201221 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.201263 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.201321 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.201359 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:11 crc kubenswrapper[4950]: E0318 20:08:11.201543 4950 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 20:08:11 crc kubenswrapper[4950]: E0318 20:08:11.201645 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:19.20161587 +0000 UTC m=+112.442457778 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 20:08:11 crc kubenswrapper[4950]: E0318 20:08:11.201787 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:08:19.201765934 +0000 UTC m=+112.442607842 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:08:11 crc kubenswrapper[4950]: E0318 20:08:11.201917 4950 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 20:08:11 crc kubenswrapper[4950]: E0318 20:08:11.201989 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:19.201967569 +0000 UTC m=+112.442809477 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 20:08:11 crc kubenswrapper[4950]: E0318 20:08:11.202118 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 20:08:11 crc kubenswrapper[4950]: E0318 20:08:11.202143 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 20:08:11 crc kubenswrapper[4950]: E0318 20:08:11.202163 4950 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:11 crc kubenswrapper[4950]: E0318 20:08:11.202209 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:19.202192725 +0000 UTC m=+112.443034633 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:11 crc kubenswrapper[4950]: E0318 20:08:11.202289 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 20:08:11 crc kubenswrapper[4950]: E0318 20:08:11.202306 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 20:08:11 crc kubenswrapper[4950]: E0318 20:08:11.202320 4950 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:11 crc kubenswrapper[4950]: E0318 20:08:11.202357 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:19.202345149 +0000 UTC m=+112.443187057 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.268098 4950 generic.go:334] "Generic (PLEG): container finished" podID="1b13cdc2-ca84-4da5-9a60-5619987e6cef" containerID="14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f" exitCode=0 Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.268188 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" event={"ID":"1b13cdc2-ca84-4da5-9a60-5619987e6cef","Type":"ContainerDied","Data":"14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f"} Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.269216 4950 scope.go:117] "RemoveContainer" containerID="7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb" Mar 18 20:08:11 crc kubenswrapper[4950]: E0318 20:08:11.269615 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.286569 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.286606 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.286617 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.286631 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.286640 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:11Z","lastTransitionTime":"2026-03-18T20:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.295916 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:11Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.320162 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:11Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.354662 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef1b45dadc92fc6f9c3a122ca974a406c18cc8259a5fd5597b4e109a9a62afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:11Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.381474 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:11Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.389524 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.389562 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.389574 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.389591 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.389602 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:11Z","lastTransitionTime":"2026-03-18T20:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.401079 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:11Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.421309 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:11Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.441018 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d09a228c-313f-4b55-a298-51ebd404cfd5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2cjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:11Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.458883 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"831bf4ca-f84c-48d2-9054-410480eb986b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:07:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 20:07:35.272898 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 20:07:35.273030 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 20:07:35.273947 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2398488620/tls.crt::/tmp/serving-cert-2398488620/tls.key\\\\\\\"\\\\nI0318 20:07:35.779001 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 20:07:35.783097 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 20:07:35.783124 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 20:07:35.783155 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 20:07:35.783174 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 20:07:35.791487 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 20:07:35.791507 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791512 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791517 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 20:07:35.791519 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 20:07:35.791522 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 20:07:35.791525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 20:07:35.791560 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 20:07:35.793966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:07:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:11Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.476277 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:11Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.479955 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.480009 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.480013 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:11 crc kubenswrapper[4950]: E0318 20:08:11.480155 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.480516 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:11 crc kubenswrapper[4950]: E0318 20:08:11.480629 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:11 crc kubenswrapper[4950]: E0318 20:08:11.480862 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:11 crc kubenswrapper[4950]: E0318 20:08:11.480950 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.489647 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:11Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.493365 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.493394 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.493404 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.493437 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.493449 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:11Z","lastTransitionTime":"2026-03-18T20:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.500831 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:11Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.513177 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:11Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.528040 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:11Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.542671 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:11Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.556397 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49a491f-1edc-4a40-b64f-147c0d87481f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b01d79ecfba0d5984e7415ca4db790ef1635698da9792b6f1e6161a868f243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122df2a4bd1b73f3890ecaef0f5840c16b127b2b7a8b749ee84f2947334b6aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5p989\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:11Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.595303 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.595341 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.595351 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.595367 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.595375 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:11Z","lastTransitionTime":"2026-03-18T20:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.699398 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.699481 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.699499 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.699524 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.699542 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:11Z","lastTransitionTime":"2026-03-18T20:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.803117 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.803163 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.803179 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.803203 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.803220 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:11Z","lastTransitionTime":"2026-03-18T20:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.808395 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs\") pod \"network-metrics-daemon-l2cjn\" (UID: \"d09a228c-313f-4b55-a298-51ebd404cfd5\") " pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:11 crc kubenswrapper[4950]: E0318 20:08:11.808606 4950 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 20:08:11 crc kubenswrapper[4950]: E0318 20:08:11.808715 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs podName:d09a228c-313f-4b55-a298-51ebd404cfd5 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:13.808686047 +0000 UTC m=+107.049527945 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs") pod "network-metrics-daemon-l2cjn" (UID: "d09a228c-313f-4b55-a298-51ebd404cfd5") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.906046 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.906079 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.906090 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.906107 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:11 crc kubenswrapper[4950]: I0318 20:08:11.906118 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:11Z","lastTransitionTime":"2026-03-18T20:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.008307 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.008351 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.008363 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.008381 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.008395 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:12Z","lastTransitionTime":"2026-03-18T20:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.112266 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.112320 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.112337 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.112362 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.112382 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:12Z","lastTransitionTime":"2026-03-18T20:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.215661 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.215718 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.215734 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.215758 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.215775 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:12Z","lastTransitionTime":"2026-03-18T20:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.275785 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" event={"ID":"1b13cdc2-ca84-4da5-9a60-5619987e6cef","Type":"ContainerStarted","Data":"52b6aa3dad160c747ebc18d26d2d504c4c72726d958a410ebe4ee008ca4741af"} Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.291696 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:12Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.311465 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52b6aa3dad160c747ebc18d26d2d504c4c72726d958a410ebe4ee008ca4741af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:12Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.318246 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.318591 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.318811 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.319000 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.319130 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:12Z","lastTransitionTime":"2026-03-18T20:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.330446 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d09a228c-313f-4b55-a298-51ebd404cfd5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2cjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:12Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.384628 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:12Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.402089 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:12Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.417646 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:12Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.421055 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.421074 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.421081 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.421094 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.421102 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:12Z","lastTransitionTime":"2026-03-18T20:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.433009 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:12Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.442520 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:12Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.456439 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"831bf4ca-f84c-48d2-9054-410480eb986b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:07:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 20:07:35.272898 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 20:07:35.273030 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 20:07:35.273947 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2398488620/tls.crt::/tmp/serving-cert-2398488620/tls.key\\\\\\\"\\\\nI0318 20:07:35.779001 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 20:07:35.783097 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 20:07:35.783124 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 20:07:35.783155 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 20:07:35.783174 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 20:07:35.791487 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 20:07:35.791507 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791512 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791517 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 20:07:35.791519 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 20:07:35.791522 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 20:07:35.791525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 20:07:35.791560 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 20:07:35.793966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:07:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:12Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.469119 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:12Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.478819 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:12Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.492173 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49a491f-1edc-4a40-b64f-147c0d87481f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b01d79ecfba0d5984e7415ca4db790ef1635698da9792b6f1e6161a868f243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122df2a4bd1b73f3890ecaef0f5840c16b127b2b7a8b749ee84f2947334b6aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5p989\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:12Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.503870 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:12Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.517123 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:12Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.524827 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.524859 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.524869 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.524883 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.524896 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:12Z","lastTransitionTime":"2026-03-18T20:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.539003 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef1b45dadc92fc6f9c3a122ca974a406c18cc8259a5fd5597b4e109a9a62afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:12Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.626932 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.626960 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.626985 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.626999 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.627007 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:12Z","lastTransitionTime":"2026-03-18T20:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.730234 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.730268 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.730278 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.730293 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.730303 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:12Z","lastTransitionTime":"2026-03-18T20:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.832843 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.832895 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.832908 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.832925 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.832938 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:12Z","lastTransitionTime":"2026-03-18T20:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.934893 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.934946 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.934963 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.934984 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:12 crc kubenswrapper[4950]: I0318 20:08:12.934998 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:12Z","lastTransitionTime":"2026-03-18T20:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.038833 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.039175 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.039307 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.039487 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.039645 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:13Z","lastTransitionTime":"2026-03-18T20:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.143812 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.143868 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.143890 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.143920 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.143941 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:13Z","lastTransitionTime":"2026-03-18T20:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.247466 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.247529 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.247550 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.247577 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.247600 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:13Z","lastTransitionTime":"2026-03-18T20:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.282112 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p2zdp_83ee6659-cfde-4844-bb92-ccb9b26451ea/ovnkube-controller/0.log" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.287297 4950 generic.go:334] "Generic (PLEG): container finished" podID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerID="4ef1b45dadc92fc6f9c3a122ca974a406c18cc8259a5fd5597b4e109a9a62afc" exitCode=1 Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.287378 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerDied","Data":"4ef1b45dadc92fc6f9c3a122ca974a406c18cc8259a5fd5597b4e109a9a62afc"} Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.288980 4950 scope.go:117] "RemoveContainer" containerID="4ef1b45dadc92fc6f9c3a122ca974a406c18cc8259a5fd5597b4e109a9a62afc" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.315700 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:13Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.343934 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:13Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.352209 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.352264 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.352278 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.352297 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.352308 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:13Z","lastTransitionTime":"2026-03-18T20:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.369027 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52b6aa3dad160c747ebc18d26d2d504c4c72726d958a410ebe4ee008ca4741af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:13Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.386914 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d09a228c-313f-4b55-a298-51ebd404cfd5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2cjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:13Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.405269 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:13Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.426821 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"831bf4ca-f84c-48d2-9054-410480eb986b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:07:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 20:07:35.272898 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 20:07:35.273030 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 20:07:35.273947 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2398488620/tls.crt::/tmp/serving-cert-2398488620/tls.key\\\\\\\"\\\\nI0318 20:07:35.779001 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 20:07:35.783097 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 20:07:35.783124 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 20:07:35.783155 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 20:07:35.783174 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 20:07:35.791487 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 20:07:35.791507 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791512 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791517 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 20:07:35.791519 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 20:07:35.791522 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 20:07:35.791525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 20:07:35.791560 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 20:07:35.793966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:07:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:13Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.443858 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:13Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.454192 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.454552 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.454723 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.454845 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.454960 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:13Z","lastTransitionTime":"2026-03-18T20:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.461478 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:13Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.479186 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:13 crc kubenswrapper[4950]: E0318 20:08:13.479271 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.479513 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:13 crc kubenswrapper[4950]: E0318 20:08:13.479564 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.479604 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:13 crc kubenswrapper[4950]: E0318 20:08:13.479644 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.479678 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:13 crc kubenswrapper[4950]: E0318 20:08:13.479722 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.522736 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:13Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.535524 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:13Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.549111 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49a491f-1edc-4a40-b64f-147c0d87481f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b01d79ecfba0d5984e7415ca4db790ef1635698da9792b6f1e6161a868f243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122df2a4bd1b73f3890ecaef0f5840c16b127b2b7a8b749ee84f2947334b6aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5p989\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:13Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.557603 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.557634 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.557645 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.557662 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.557674 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:13Z","lastTransitionTime":"2026-03-18T20:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.563676 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:13Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.577529 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:13Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.593571 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:13Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.616100 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef1b45dadc92fc6f9c3a122ca974a406c18cc8259a5fd5597b4e109a9a62afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ef1b45dadc92fc6f9c3a122ca974a406c18cc8259a5fd5597b4e109a9a62afc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"message\\\":\\\"or *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518498 6629 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518644 6629 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 20:08:12.518795 6629 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518869 6629 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518905 6629 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518970 6629 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.520309 6629 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0318 20:08:12.520335 6629 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0318 20:08:12.520354 6629 factory.go:656] Stopping watch factory\\\\nI0318 20:08:12.520365 6629 ovnkube.go:599] Stopped ovnkube\\\\nI0318 20:08:12.520390 6629 handler.go:208] Removed *v1.Node event handler 2\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:13Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.660068 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.660100 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.660111 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.660126 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.660137 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:13Z","lastTransitionTime":"2026-03-18T20:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.762723 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.762768 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.762780 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.762829 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.762842 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:13Z","lastTransitionTime":"2026-03-18T20:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.831136 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs\") pod \"network-metrics-daemon-l2cjn\" (UID: \"d09a228c-313f-4b55-a298-51ebd404cfd5\") " pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:13 crc kubenswrapper[4950]: E0318 20:08:13.831372 4950 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 20:08:13 crc kubenswrapper[4950]: E0318 20:08:13.831573 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs podName:d09a228c-313f-4b55-a298-51ebd404cfd5 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:17.831531008 +0000 UTC m=+111.072372916 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs") pod "network-metrics-daemon-l2cjn" (UID: "d09a228c-313f-4b55-a298-51ebd404cfd5") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.864979 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.865018 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.865028 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.865044 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.865053 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:13Z","lastTransitionTime":"2026-03-18T20:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.966889 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.966948 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.966960 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.966978 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:13 crc kubenswrapper[4950]: I0318 20:08:13.966989 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:13Z","lastTransitionTime":"2026-03-18T20:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.069390 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.069467 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.069483 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.069505 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.069520 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:14Z","lastTransitionTime":"2026-03-18T20:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.171567 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.171627 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.171644 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.171670 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.171686 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:14Z","lastTransitionTime":"2026-03-18T20:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.273513 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.273558 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.273571 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.273591 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.273604 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:14Z","lastTransitionTime":"2026-03-18T20:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.292693 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p2zdp_83ee6659-cfde-4844-bb92-ccb9b26451ea/ovnkube-controller/1.log" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.293463 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p2zdp_83ee6659-cfde-4844-bb92-ccb9b26451ea/ovnkube-controller/0.log" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.296256 4950 generic.go:334] "Generic (PLEG): container finished" podID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerID="a9ff70a20c13a291644f12fd70e8d676e7e7266712039a4bcbacb0d73963d7f6" exitCode=1 Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.296300 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerDied","Data":"a9ff70a20c13a291644f12fd70e8d676e7e7266712039a4bcbacb0d73963d7f6"} Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.296341 4950 scope.go:117] "RemoveContainer" containerID="4ef1b45dadc92fc6f9c3a122ca974a406c18cc8259a5fd5597b4e109a9a62afc" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.297158 4950 scope.go:117] "RemoveContainer" containerID="a9ff70a20c13a291644f12fd70e8d676e7e7266712039a4bcbacb0d73963d7f6" Mar 18 20:08:14 crc kubenswrapper[4950]: E0318 20:08:14.297381 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p2zdp_openshift-ovn-kubernetes(83ee6659-cfde-4844-bb92-ccb9b26451ea)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.317354 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:14Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.331968 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:14Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.343136 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49a491f-1edc-4a40-b64f-147c0d87481f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b01d79ecfba0d5984e7415ca4db790ef1635698da9792b6f1e6161a868f243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122df2a4bd1b73f3890ecaef0f5840c16b127b2b7a8b749ee84f2947334b6aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5p989\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:14Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.361151 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9ff70a20c13a291644f12fd70e8d676e7e7266712039a4bcbacb0d73963d7f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ef1b45dadc92fc6f9c3a122ca974a406c18cc8259a5fd5597b4e109a9a62afc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"message\\\":\\\"or *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518498 6629 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518644 6629 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 20:08:12.518795 6629 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518869 6629 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518905 6629 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518970 6629 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.520309 6629 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0318 20:08:12.520335 6629 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0318 20:08:12.520354 6629 factory.go:656] Stopping watch factory\\\\nI0318 20:08:12.520365 6629 ovnkube.go:599] Stopped ovnkube\\\\nI0318 20:08:12.520390 6629 handler.go:208] Removed *v1.Node event handler 2\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9ff70a20c13a291644f12fd70e8d676e7e7266712039a4bcbacb0d73963d7f6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:14Z\\\",\\\"message\\\":\\\"TrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI0318 20:08:14.221326 6825 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0318 20:08:14.221340 6825 lb_config.go:1031] Cluster endpoints for openshift-etcd-operator/metrics for network=default are: map[]\\\\nI0318 20:08:14.221361 6825 services_controller.go:443] Built service openshift-etcd-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.188\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0318 20:08:14.221380 6825 services_controller.go:444] Built service openshift-etcd-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0318 20:08:14.221390 6825 services_controller.go:445] Built service openshift-etcd-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF0318 20:08:14.221450 6825 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:14Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.373577 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:14Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.376989 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.377062 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.377076 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.377094 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.377107 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:14Z","lastTransitionTime":"2026-03-18T20:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.387053 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:14Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.401884 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d09a228c-313f-4b55-a298-51ebd404cfd5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2cjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:14Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.420112 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:14Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.434253 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:14Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.450818 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52b6aa3dad160c747ebc18d26d2d504c4c72726d958a410ebe4ee008ca4741af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:14Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.465007 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:14Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.477178 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:14Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.479334 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.479402 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.479456 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.479484 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.479502 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:14Z","lastTransitionTime":"2026-03-18T20:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.499266 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"831bf4ca-f84c-48d2-9054-410480eb986b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:07:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 20:07:35.272898 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 20:07:35.273030 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 20:07:35.273947 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2398488620/tls.crt::/tmp/serving-cert-2398488620/tls.key\\\\\\\"\\\\nI0318 20:07:35.779001 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 20:07:35.783097 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 20:07:35.783124 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 20:07:35.783155 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 20:07:35.783174 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 20:07:35.791487 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 20:07:35.791507 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791512 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791517 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 20:07:35.791519 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 20:07:35.791522 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 20:07:35.791525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 20:07:35.791560 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 20:07:35.793966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:07:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:14Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.512868 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:14Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.527471 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:14Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.581385 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.581430 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.581438 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.581454 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.581462 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:14Z","lastTransitionTime":"2026-03-18T20:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.683940 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.683976 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.683985 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.683999 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.684007 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:14Z","lastTransitionTime":"2026-03-18T20:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.787524 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.787587 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.787604 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.787626 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.787645 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:14Z","lastTransitionTime":"2026-03-18T20:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.890304 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.890358 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.890375 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.890397 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.890439 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:14Z","lastTransitionTime":"2026-03-18T20:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.993305 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.993354 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.993370 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.993394 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:14 crc kubenswrapper[4950]: I0318 20:08:14.993437 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:14Z","lastTransitionTime":"2026-03-18T20:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.096628 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.096690 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.096711 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.096733 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.096751 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:15Z","lastTransitionTime":"2026-03-18T20:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.200333 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.200441 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.200469 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.200503 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.200525 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:15Z","lastTransitionTime":"2026-03-18T20:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.302849 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.302923 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.302947 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.302978 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.303001 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:15Z","lastTransitionTime":"2026-03-18T20:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.303628 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p2zdp_83ee6659-cfde-4844-bb92-ccb9b26451ea/ovnkube-controller/1.log" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.406088 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.406194 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.406218 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.406252 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.406279 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:15Z","lastTransitionTime":"2026-03-18T20:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.479252 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.479330 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.479447 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:15 crc kubenswrapper[4950]: E0318 20:08:15.479602 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.480058 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:15 crc kubenswrapper[4950]: E0318 20:08:15.480212 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:15 crc kubenswrapper[4950]: E0318 20:08:15.480333 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:15 crc kubenswrapper[4950]: E0318 20:08:15.480450 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.512167 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.512239 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.512260 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.512287 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.512315 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:15Z","lastTransitionTime":"2026-03-18T20:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.615843 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.615926 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.615952 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.616003 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.616022 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:15Z","lastTransitionTime":"2026-03-18T20:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.719124 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.719242 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.719260 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.719284 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.719301 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:15Z","lastTransitionTime":"2026-03-18T20:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.823004 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.823081 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.823103 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.823135 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.823157 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:15Z","lastTransitionTime":"2026-03-18T20:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.926583 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.926649 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.926670 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.926700 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:15 crc kubenswrapper[4950]: I0318 20:08:15.926721 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:15Z","lastTransitionTime":"2026-03-18T20:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.031156 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.031219 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.031241 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.031270 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.031293 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:16Z","lastTransitionTime":"2026-03-18T20:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.134895 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.134968 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.134992 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.135027 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.135051 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:16Z","lastTransitionTime":"2026-03-18T20:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.238494 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.238616 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.238635 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.238660 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.238678 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:16Z","lastTransitionTime":"2026-03-18T20:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.341316 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.341388 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.341441 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.341480 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.341505 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:16Z","lastTransitionTime":"2026-03-18T20:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.445110 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.445172 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.445189 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.445215 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.445232 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:16Z","lastTransitionTime":"2026-03-18T20:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.548234 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.548290 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.548307 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.548330 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.548347 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:16Z","lastTransitionTime":"2026-03-18T20:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.651583 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.651620 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.651634 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.651653 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.651667 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:16Z","lastTransitionTime":"2026-03-18T20:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.755804 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.755866 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.755879 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.755898 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.755911 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:16Z","lastTransitionTime":"2026-03-18T20:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.859641 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.859709 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.859722 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.859744 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.859784 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:16Z","lastTransitionTime":"2026-03-18T20:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.963296 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.963346 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.963366 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.963385 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:16 crc kubenswrapper[4950]: I0318 20:08:16.963399 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:16Z","lastTransitionTime":"2026-03-18T20:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.065779 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.065828 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.065847 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.065871 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.065888 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:17Z","lastTransitionTime":"2026-03-18T20:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.169624 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.169720 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.169742 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.169789 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.169816 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:17Z","lastTransitionTime":"2026-03-18T20:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.273202 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.273268 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.273289 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.273307 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.273320 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:17Z","lastTransitionTime":"2026-03-18T20:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.377261 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.377326 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.377348 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.377380 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.377404 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:17Z","lastTransitionTime":"2026-03-18T20:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.478835 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.478877 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:17 crc kubenswrapper[4950]: E0318 20:08:17.480341 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.480013 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:17 crc kubenswrapper[4950]: E0318 20:08:17.480498 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.480051 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:17 crc kubenswrapper[4950]: E0318 20:08:17.479518 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:17 crc kubenswrapper[4950]: E0318 20:08:17.481033 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.483014 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.483066 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.483099 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.483125 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.483144 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:17Z","lastTransitionTime":"2026-03-18T20:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.501932 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"831bf4ca-f84c-48d2-9054-410480eb986b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:07:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 20:07:35.272898 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 20:07:35.273030 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 20:07:35.273947 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2398488620/tls.crt::/tmp/serving-cert-2398488620/tls.key\\\\\\\"\\\\nI0318 20:07:35.779001 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 20:07:35.783097 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 20:07:35.783124 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 20:07:35.783155 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 20:07:35.783174 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 20:07:35.791487 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 20:07:35.791507 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791512 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791517 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 20:07:35.791519 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 20:07:35.791522 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 20:07:35.791525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 20:07:35.791560 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 20:07:35.793966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:07:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:17Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.516802 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:17Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.534664 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:17Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.556730 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:17Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.573183 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:17Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.586720 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.586763 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.586775 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.586793 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.586805 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:17Z","lastTransitionTime":"2026-03-18T20:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.592232 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:17Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.609745 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:17Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.628961 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49a491f-1edc-4a40-b64f-147c0d87481f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b01d79ecfba0d5984e7415ca4db790ef1635698da9792b6f1e6161a868f243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122df2a4bd1b73f3890ecaef0f5840c16b127b2b7a8b749ee84f2947334b6aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5p989\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:17Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.646595 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:17Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.668089 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:17Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.689961 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.690024 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.690047 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.690077 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.690100 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:17Z","lastTransitionTime":"2026-03-18T20:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.698622 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9ff70a20c13a291644f12fd70e8d676e7e7266712039a4bcbacb0d73963d7f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ef1b45dadc92fc6f9c3a122ca974a406c18cc8259a5fd5597b4e109a9a62afc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"message\\\":\\\"or *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518498 6629 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518644 6629 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 20:08:12.518795 6629 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518869 6629 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518905 6629 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518970 6629 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.520309 6629 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0318 20:08:12.520335 6629 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0318 20:08:12.520354 6629 factory.go:656] Stopping watch factory\\\\nI0318 20:08:12.520365 6629 ovnkube.go:599] Stopped ovnkube\\\\nI0318 20:08:12.520390 6629 handler.go:208] Removed *v1.Node event handler 2\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9ff70a20c13a291644f12fd70e8d676e7e7266712039a4bcbacb0d73963d7f6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:14Z\\\",\\\"message\\\":\\\"TrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI0318 20:08:14.221326 6825 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0318 20:08:14.221340 6825 lb_config.go:1031] Cluster endpoints for openshift-etcd-operator/metrics for network=default are: map[]\\\\nI0318 20:08:14.221361 6825 services_controller.go:443] Built service openshift-etcd-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.188\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0318 20:08:14.221380 6825 services_controller.go:444] Built service openshift-etcd-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0318 20:08:14.221390 6825 services_controller.go:445] Built service openshift-etcd-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF0318 20:08:14.221450 6825 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:17Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.717815 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:17Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.733194 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:17Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.754514 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52b6aa3dad160c747ebc18d26d2d504c4c72726d958a410ebe4ee008ca4741af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:17Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.770310 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d09a228c-313f-4b55-a298-51ebd404cfd5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2cjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:17Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.792595 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.792660 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.792682 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.792714 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.792737 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:17Z","lastTransitionTime":"2026-03-18T20:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.877759 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs\") pod \"network-metrics-daemon-l2cjn\" (UID: \"d09a228c-313f-4b55-a298-51ebd404cfd5\") " pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:17 crc kubenswrapper[4950]: E0318 20:08:17.877979 4950 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 20:08:17 crc kubenswrapper[4950]: E0318 20:08:17.878086 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs podName:d09a228c-313f-4b55-a298-51ebd404cfd5 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:25.878060045 +0000 UTC m=+119.118901923 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs") pod "network-metrics-daemon-l2cjn" (UID: "d09a228c-313f-4b55-a298-51ebd404cfd5") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.896212 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.896263 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.896279 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.896302 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.896319 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:17Z","lastTransitionTime":"2026-03-18T20:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.999764 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.999825 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:17 crc kubenswrapper[4950]: I0318 20:08:17.999848 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:17.999876 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:17.999900 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:17Z","lastTransitionTime":"2026-03-18T20:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.103221 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.103305 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.103362 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.103385 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.103403 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:18Z","lastTransitionTime":"2026-03-18T20:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.206788 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.206823 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.206832 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.206847 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.206856 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:18Z","lastTransitionTime":"2026-03-18T20:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.309064 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.309112 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.309166 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.309194 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.309210 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:18Z","lastTransitionTime":"2026-03-18T20:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.412219 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.412289 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.412312 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.412339 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.412363 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:18Z","lastTransitionTime":"2026-03-18T20:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.515508 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.515572 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.515589 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.515611 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.515627 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:18Z","lastTransitionTime":"2026-03-18T20:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.618264 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.618313 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.618335 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.618359 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.618376 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:18Z","lastTransitionTime":"2026-03-18T20:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.721170 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.721220 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.721243 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.721272 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.721293 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:18Z","lastTransitionTime":"2026-03-18T20:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.824570 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.824630 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.824650 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.824674 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.824694 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:18Z","lastTransitionTime":"2026-03-18T20:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.928125 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.928211 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.928249 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.928279 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:18 crc kubenswrapper[4950]: I0318 20:08:18.928298 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:18Z","lastTransitionTime":"2026-03-18T20:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.031181 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.031213 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.031225 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.031242 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.031252 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:19Z","lastTransitionTime":"2026-03-18T20:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.134657 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.134689 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.134701 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.134718 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.134730 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:19Z","lastTransitionTime":"2026-03-18T20:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.180777 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.180818 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.180830 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.180845 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.180855 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:19Z","lastTransitionTime":"2026-03-18T20:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:19 crc kubenswrapper[4950]: E0318 20:08:19.193262 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:19Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.198073 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.198132 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.198155 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.198177 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.198193 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:19Z","lastTransitionTime":"2026-03-18T20:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:19 crc kubenswrapper[4950]: E0318 20:08:19.213518 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:19Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.218355 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.218382 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.218392 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.218406 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.218431 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:19Z","lastTransitionTime":"2026-03-18T20:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:19 crc kubenswrapper[4950]: E0318 20:08:19.237859 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:19Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.242324 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.242373 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.242389 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.242436 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.242485 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:19Z","lastTransitionTime":"2026-03-18T20:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:19 crc kubenswrapper[4950]: E0318 20:08:19.257718 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:19Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.261867 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.261921 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.261938 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.261960 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.261976 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:19Z","lastTransitionTime":"2026-03-18T20:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:19 crc kubenswrapper[4950]: E0318 20:08:19.280594 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:19Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:19 crc kubenswrapper[4950]: E0318 20:08:19.280715 4950 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.282359 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.282449 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.282462 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.282501 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.282514 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:19Z","lastTransitionTime":"2026-03-18T20:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.294225 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.294341 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:19 crc kubenswrapper[4950]: E0318 20:08:19.294390 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:08:35.294364783 +0000 UTC m=+128.535206661 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.294474 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:19 crc kubenswrapper[4950]: E0318 20:08:19.294494 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 20:08:19 crc kubenswrapper[4950]: E0318 20:08:19.294513 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 20:08:19 crc kubenswrapper[4950]: E0318 20:08:19.294525 4950 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.294538 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:19 crc kubenswrapper[4950]: E0318 20:08:19.294574 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:35.294558378 +0000 UTC m=+128.535400316 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.294612 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:19 crc kubenswrapper[4950]: E0318 20:08:19.294632 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 20:08:19 crc kubenswrapper[4950]: E0318 20:08:19.294652 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 20:08:19 crc kubenswrapper[4950]: E0318 20:08:19.294664 4950 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:19 crc kubenswrapper[4950]: E0318 20:08:19.294697 4950 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 20:08:19 crc kubenswrapper[4950]: E0318 20:08:19.294718 4950 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 20:08:19 crc kubenswrapper[4950]: E0318 20:08:19.294702 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:35.294689052 +0000 UTC m=+128.535531020 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:19 crc kubenswrapper[4950]: E0318 20:08:19.294755 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:35.294743503 +0000 UTC m=+128.535585371 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 20:08:19 crc kubenswrapper[4950]: E0318 20:08:19.294765 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:35.294759724 +0000 UTC m=+128.535601592 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.385138 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.385207 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.385257 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.385286 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.385308 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:19Z","lastTransitionTime":"2026-03-18T20:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.479254 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.479329 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.479363 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.479328 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:19 crc kubenswrapper[4950]: E0318 20:08:19.479472 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:19 crc kubenswrapper[4950]: E0318 20:08:19.479562 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:19 crc kubenswrapper[4950]: E0318 20:08:19.479662 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:19 crc kubenswrapper[4950]: E0318 20:08:19.479772 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.487294 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.487321 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.487330 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.487341 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.487351 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:19Z","lastTransitionTime":"2026-03-18T20:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.590817 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.590861 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.590877 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.590899 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.590917 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:19Z","lastTransitionTime":"2026-03-18T20:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.693101 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.693137 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.693146 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.693160 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.693170 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:19Z","lastTransitionTime":"2026-03-18T20:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.796831 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.796898 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.796923 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.796953 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.796975 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:19Z","lastTransitionTime":"2026-03-18T20:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.900344 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.900467 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.900485 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.900509 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:19 crc kubenswrapper[4950]: I0318 20:08:19.900526 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:19Z","lastTransitionTime":"2026-03-18T20:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.003772 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.003850 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.003873 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.003906 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.003929 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:20Z","lastTransitionTime":"2026-03-18T20:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.105767 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.105809 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.105822 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.105841 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.105853 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:20Z","lastTransitionTime":"2026-03-18T20:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.208039 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.208070 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.208078 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.208090 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.208099 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:20Z","lastTransitionTime":"2026-03-18T20:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.310740 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.310786 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.310795 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.310808 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.310817 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:20Z","lastTransitionTime":"2026-03-18T20:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.413396 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.413473 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.413492 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.413513 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.413528 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:20Z","lastTransitionTime":"2026-03-18T20:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.516597 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.516629 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.516637 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.516650 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.516659 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:20Z","lastTransitionTime":"2026-03-18T20:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.620292 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.620350 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.620365 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.620389 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.620403 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:20Z","lastTransitionTime":"2026-03-18T20:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.723557 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.723616 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.723632 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.723660 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.723677 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:20Z","lastTransitionTime":"2026-03-18T20:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.827496 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.827568 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.827585 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.827610 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.827627 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:20Z","lastTransitionTime":"2026-03-18T20:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.931686 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.931748 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.931769 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.931792 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:20 crc kubenswrapper[4950]: I0318 20:08:20.931810 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:20Z","lastTransitionTime":"2026-03-18T20:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.035328 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.035405 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.035438 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.035467 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.035486 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:21Z","lastTransitionTime":"2026-03-18T20:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.139250 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.139309 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.139319 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.139341 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.139352 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:21Z","lastTransitionTime":"2026-03-18T20:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.243650 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.243699 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.243709 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.243728 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.243739 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:21Z","lastTransitionTime":"2026-03-18T20:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.347802 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.347886 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.347911 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.347964 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.347989 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:21Z","lastTransitionTime":"2026-03-18T20:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.451551 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.451613 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.451631 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.451659 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.451677 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:21Z","lastTransitionTime":"2026-03-18T20:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.479480 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.479526 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.479616 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.479487 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:21 crc kubenswrapper[4950]: E0318 20:08:21.479764 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:21 crc kubenswrapper[4950]: E0318 20:08:21.479893 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:21 crc kubenswrapper[4950]: E0318 20:08:21.480195 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:21 crc kubenswrapper[4950]: E0318 20:08:21.480458 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.554392 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.554462 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.554473 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.554499 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.554512 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:21Z","lastTransitionTime":"2026-03-18T20:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.657540 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.657597 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.657609 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.657629 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.657643 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:21Z","lastTransitionTime":"2026-03-18T20:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.761270 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.761455 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.761479 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.761504 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.761522 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:21Z","lastTransitionTime":"2026-03-18T20:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.865608 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.865678 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.865701 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.865742 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.865768 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:21Z","lastTransitionTime":"2026-03-18T20:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.968653 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.969063 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.969303 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.969496 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:21 crc kubenswrapper[4950]: I0318 20:08:21.969624 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:21Z","lastTransitionTime":"2026-03-18T20:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.073848 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.073909 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.073926 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.073965 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.073981 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:22Z","lastTransitionTime":"2026-03-18T20:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.177940 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.178010 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.178023 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.178049 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.178066 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:22Z","lastTransitionTime":"2026-03-18T20:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.281844 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.281918 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.281938 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.281969 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.281991 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:22Z","lastTransitionTime":"2026-03-18T20:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.386085 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.386181 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.386206 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.386239 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.386261 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:22Z","lastTransitionTime":"2026-03-18T20:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.490730 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.490850 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.490875 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.490904 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.490923 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:22Z","lastTransitionTime":"2026-03-18T20:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.594777 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.594839 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.594858 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.594888 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.594912 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:22Z","lastTransitionTime":"2026-03-18T20:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.697561 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.697604 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.697613 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.697628 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.697637 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:22Z","lastTransitionTime":"2026-03-18T20:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.800125 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.800210 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.800244 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.800279 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.800304 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:22Z","lastTransitionTime":"2026-03-18T20:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.904530 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.904610 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.904625 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.904651 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:22 crc kubenswrapper[4950]: I0318 20:08:22.904664 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:22Z","lastTransitionTime":"2026-03-18T20:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.008082 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.008167 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.008185 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.008216 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.008237 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:23Z","lastTransitionTime":"2026-03-18T20:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.111579 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.111655 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.111675 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.111704 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.111725 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:23Z","lastTransitionTime":"2026-03-18T20:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.215647 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.215726 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.215750 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.215786 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.215810 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:23Z","lastTransitionTime":"2026-03-18T20:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.318848 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.318925 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.318946 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.318972 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.318992 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:23Z","lastTransitionTime":"2026-03-18T20:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.422967 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.423042 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.423056 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.423084 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.423102 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:23Z","lastTransitionTime":"2026-03-18T20:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.480707 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.480781 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:23 crc kubenswrapper[4950]: E0318 20:08:23.481208 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.481265 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:23 crc kubenswrapper[4950]: E0318 20:08:23.481502 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:23 crc kubenswrapper[4950]: E0318 20:08:23.481669 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.482688 4950 scope.go:117] "RemoveContainer" containerID="7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.482717 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:23 crc kubenswrapper[4950]: E0318 20:08:23.484082 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.527685 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.528099 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.528147 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.528175 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.528194 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:23Z","lastTransitionTime":"2026-03-18T20:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.631721 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.631763 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.631774 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.631789 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.631799 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:23Z","lastTransitionTime":"2026-03-18T20:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.734340 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.734379 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.734388 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.734434 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.734444 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:23Z","lastTransitionTime":"2026-03-18T20:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.839367 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.839461 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.839480 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.839503 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.839518 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:23Z","lastTransitionTime":"2026-03-18T20:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.943729 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.943790 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.943802 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.943819 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:23 crc kubenswrapper[4950]: I0318 20:08:23.943865 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:23Z","lastTransitionTime":"2026-03-18T20:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.046823 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.046857 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.046882 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.046896 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.046906 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:24Z","lastTransitionTime":"2026-03-18T20:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.150463 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.150541 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.150558 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.150611 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.150628 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:24Z","lastTransitionTime":"2026-03-18T20:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.254143 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.254229 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.254246 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.254270 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.254317 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:24Z","lastTransitionTime":"2026-03-18T20:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.343170 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.345053 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a"} Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.346361 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.356878 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.356923 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.356940 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.356962 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.356980 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:24Z","lastTransitionTime":"2026-03-18T20:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.363464 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:24Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.384698 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:24Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.416547 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52b6aa3dad160c747ebc18d26d2d504c4c72726d958a410ebe4ee008ca4741af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:24Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.436237 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d09a228c-313f-4b55-a298-51ebd404cfd5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2cjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:24Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.450455 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:24Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.460818 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.460911 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.460962 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.460990 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.461012 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:24Z","lastTransitionTime":"2026-03-18T20:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.472322 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"831bf4ca-f84c-48d2-9054-410480eb986b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:07:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 20:07:35.272898 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 20:07:35.273030 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 20:07:35.273947 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2398488620/tls.crt::/tmp/serving-cert-2398488620/tls.key\\\\\\\"\\\\nI0318 20:07:35.779001 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 20:07:35.783097 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 20:07:35.783124 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 20:07:35.783155 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 20:07:35.783174 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 20:07:35.791487 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 20:07:35.791507 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791512 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791517 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 20:07:35.791519 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 20:07:35.791522 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 20:07:35.791525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 20:07:35.791560 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 20:07:35.793966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:07:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:24Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.488076 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:24Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.502257 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:24Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.515961 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:24Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.527211 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:24Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.546162 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49a491f-1edc-4a40-b64f-147c0d87481f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b01d79ecfba0d5984e7415ca4db790ef1635698da9792b6f1e6161a868f243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122df2a4bd1b73f3890ecaef0f5840c16b127b2b7a8b749ee84f2947334b6aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5p989\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:24Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.563574 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:24Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.564068 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.564101 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.564114 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.564164 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.564177 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:24Z","lastTransitionTime":"2026-03-18T20:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.577094 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:24Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.590728 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:24Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.610983 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9ff70a20c13a291644f12fd70e8d676e7e7266712039a4bcbacb0d73963d7f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ef1b45dadc92fc6f9c3a122ca974a406c18cc8259a5fd5597b4e109a9a62afc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"message\\\":\\\"or *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518498 6629 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518644 6629 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 20:08:12.518795 6629 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518869 6629 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518905 6629 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518970 6629 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.520309 6629 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0318 20:08:12.520335 6629 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0318 20:08:12.520354 6629 factory.go:656] Stopping watch factory\\\\nI0318 20:08:12.520365 6629 ovnkube.go:599] Stopped ovnkube\\\\nI0318 20:08:12.520390 6629 handler.go:208] Removed *v1.Node event handler 2\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9ff70a20c13a291644f12fd70e8d676e7e7266712039a4bcbacb0d73963d7f6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:14Z\\\",\\\"message\\\":\\\"TrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI0318 20:08:14.221326 6825 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0318 20:08:14.221340 6825 lb_config.go:1031] Cluster endpoints for openshift-etcd-operator/metrics for network=default are: map[]\\\\nI0318 20:08:14.221361 6825 services_controller.go:443] Built service openshift-etcd-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.188\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0318 20:08:14.221380 6825 services_controller.go:444] Built service openshift-etcd-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0318 20:08:14.221390 6825 services_controller.go:445] Built service openshift-etcd-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF0318 20:08:14.221450 6825 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:24Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.666778 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.666857 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.666875 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.666897 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.666950 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:24Z","lastTransitionTime":"2026-03-18T20:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.770077 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.770139 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.770151 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.770169 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.770182 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:24Z","lastTransitionTime":"2026-03-18T20:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.873288 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.873362 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.873386 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.873445 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.873470 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:24Z","lastTransitionTime":"2026-03-18T20:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.975806 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.975851 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.975868 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.975889 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:24 crc kubenswrapper[4950]: I0318 20:08:24.975905 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:24Z","lastTransitionTime":"2026-03-18T20:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.079312 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.079360 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.079372 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.079390 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.079402 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:25Z","lastTransitionTime":"2026-03-18T20:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.182633 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.182694 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.182711 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.182737 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.182756 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:25Z","lastTransitionTime":"2026-03-18T20:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.285526 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.285577 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.285592 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.285612 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.285628 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:25Z","lastTransitionTime":"2026-03-18T20:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.388211 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.388270 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.388289 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.388314 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.388333 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:25Z","lastTransitionTime":"2026-03-18T20:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.479158 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.479182 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.479319 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:25 crc kubenswrapper[4950]: E0318 20:08:25.479538 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.479578 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:25 crc kubenswrapper[4950]: E0318 20:08:25.479748 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:25 crc kubenswrapper[4950]: E0318 20:08:25.479834 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:25 crc kubenswrapper[4950]: E0318 20:08:25.479996 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.490559 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.490624 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.490649 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.490675 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.490697 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:25Z","lastTransitionTime":"2026-03-18T20:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.593455 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.593491 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.593499 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.593512 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.593521 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:25Z","lastTransitionTime":"2026-03-18T20:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.697182 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.697242 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.697259 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.697287 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.697304 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:25Z","lastTransitionTime":"2026-03-18T20:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.800404 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.800465 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.800476 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.800494 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.800506 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:25Z","lastTransitionTime":"2026-03-18T20:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.903728 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.903774 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.903785 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.903801 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.903813 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:25Z","lastTransitionTime":"2026-03-18T20:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:25 crc kubenswrapper[4950]: I0318 20:08:25.975841 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs\") pod \"network-metrics-daemon-l2cjn\" (UID: \"d09a228c-313f-4b55-a298-51ebd404cfd5\") " pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:25 crc kubenswrapper[4950]: E0318 20:08:25.976063 4950 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 20:08:25 crc kubenswrapper[4950]: E0318 20:08:25.976193 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs podName:d09a228c-313f-4b55-a298-51ebd404cfd5 nodeName:}" failed. No retries permitted until 2026-03-18 20:08:41.976164072 +0000 UTC m=+135.217005970 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs") pod "network-metrics-daemon-l2cjn" (UID: "d09a228c-313f-4b55-a298-51ebd404cfd5") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.007120 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.007195 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.007214 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.007264 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.007283 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:26Z","lastTransitionTime":"2026-03-18T20:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.109793 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.109835 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.109845 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.109860 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.109870 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:26Z","lastTransitionTime":"2026-03-18T20:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.213577 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.213671 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.213690 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.213715 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.213732 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:26Z","lastTransitionTime":"2026-03-18T20:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.317445 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.317505 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.317522 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.317548 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.317565 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:26Z","lastTransitionTime":"2026-03-18T20:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.420352 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.420440 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.420458 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.420481 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.420498 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:26Z","lastTransitionTime":"2026-03-18T20:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.523117 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.523148 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.523156 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.523169 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.523178 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:26Z","lastTransitionTime":"2026-03-18T20:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.625902 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.625970 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.625995 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.626026 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.626047 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:26Z","lastTransitionTime":"2026-03-18T20:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.728523 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.728589 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.728612 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.728639 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.728662 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:26Z","lastTransitionTime":"2026-03-18T20:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.831119 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.831148 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.831155 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.831167 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.831177 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:26Z","lastTransitionTime":"2026-03-18T20:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.937907 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.937962 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.937979 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.938002 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:26 crc kubenswrapper[4950]: I0318 20:08:26.938024 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:26Z","lastTransitionTime":"2026-03-18T20:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.040684 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.040747 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.040771 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.040800 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.040824 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:27Z","lastTransitionTime":"2026-03-18T20:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.143554 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.143618 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.143636 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.143660 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.143678 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:27Z","lastTransitionTime":"2026-03-18T20:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.246872 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.246946 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.246967 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.246990 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.247009 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:27Z","lastTransitionTime":"2026-03-18T20:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.349399 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.349518 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.349542 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.349572 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.349598 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:27Z","lastTransitionTime":"2026-03-18T20:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:27 crc kubenswrapper[4950]: E0318 20:08:27.450073 4950 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.479530 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.479620 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.479714 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:27 crc kubenswrapper[4950]: E0318 20:08:27.479907 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.479925 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:27 crc kubenswrapper[4950]: E0318 20:08:27.480095 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:27 crc kubenswrapper[4950]: E0318 20:08:27.480220 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:27 crc kubenswrapper[4950]: E0318 20:08:27.480393 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.502775 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:27Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.523404 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:27Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.550884 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52b6aa3dad160c747ebc18d26d2d504c4c72726d958a410ebe4ee008ca4741af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:27Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:27 crc kubenswrapper[4950]: E0318 20:08:27.553157 4950 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.569340 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d09a228c-313f-4b55-a298-51ebd404cfd5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2cjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:27Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.592124 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"831bf4ca-f84c-48d2-9054-410480eb986b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:07:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 20:07:35.272898 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 20:07:35.273030 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 20:07:35.273947 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2398488620/tls.crt::/tmp/serving-cert-2398488620/tls.key\\\\\\\"\\\\nI0318 20:07:35.779001 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 20:07:35.783097 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 20:07:35.783124 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 20:07:35.783155 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 20:07:35.783174 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 20:07:35.791487 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 20:07:35.791507 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791512 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791517 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 20:07:35.791519 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 20:07:35.791522 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 20:07:35.791525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 20:07:35.791560 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 20:07:35.793966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:07:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:27Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.610109 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:27Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.626372 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:27Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.644760 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:27Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.660232 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:27Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.678213 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49a491f-1edc-4a40-b64f-147c0d87481f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b01d79ecfba0d5984e7415ca4db790ef1635698da9792b6f1e6161a868f243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122df2a4bd1b73f3890ecaef0f5840c16b127b2b7a8b749ee84f2947334b6aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5p989\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:27Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.698562 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:27Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.716185 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:27Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.731933 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:27Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.747952 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:27Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:27 crc kubenswrapper[4950]: I0318 20:08:27.774074 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9ff70a20c13a291644f12fd70e8d676e7e7266712039a4bcbacb0d73963d7f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ef1b45dadc92fc6f9c3a122ca974a406c18cc8259a5fd5597b4e109a9a62afc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"message\\\":\\\"or *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518498 6629 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518644 6629 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0318 20:08:12.518795 6629 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518869 6629 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518905 6629 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.518970 6629 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0318 20:08:12.520309 6629 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0318 20:08:12.520335 6629 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0318 20:08:12.520354 6629 factory.go:656] Stopping watch factory\\\\nI0318 20:08:12.520365 6629 ovnkube.go:599] Stopped ovnkube\\\\nI0318 20:08:12.520390 6629 handler.go:208] Removed *v1.Node event handler 2\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9ff70a20c13a291644f12fd70e8d676e7e7266712039a4bcbacb0d73963d7f6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:14Z\\\",\\\"message\\\":\\\"TrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI0318 20:08:14.221326 6825 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0318 20:08:14.221340 6825 lb_config.go:1031] Cluster endpoints for openshift-etcd-operator/metrics for network=default are: map[]\\\\nI0318 20:08:14.221361 6825 services_controller.go:443] Built service openshift-etcd-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.188\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0318 20:08:14.221380 6825 services_controller.go:444] Built service openshift-etcd-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0318 20:08:14.221390 6825 services_controller.go:445] Built service openshift-etcd-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF0318 20:08:14.221450 6825 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:27Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.406860 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.406939 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.406962 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.406988 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.407008 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:29Z","lastTransitionTime":"2026-03-18T20:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:29 crc kubenswrapper[4950]: E0318 20:08:29.429230 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:29Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.434655 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.434712 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.434721 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.434733 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.434742 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:29Z","lastTransitionTime":"2026-03-18T20:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:29 crc kubenswrapper[4950]: E0318 20:08:29.454637 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:29Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.458834 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.458867 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.458878 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.458893 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.458904 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:29Z","lastTransitionTime":"2026-03-18T20:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:29 crc kubenswrapper[4950]: E0318 20:08:29.472137 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:29Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.475215 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.475241 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.475252 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.475266 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.475277 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:29Z","lastTransitionTime":"2026-03-18T20:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.479513 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:29 crc kubenswrapper[4950]: E0318 20:08:29.479610 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.479516 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.479670 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.479879 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:29 crc kubenswrapper[4950]: E0318 20:08:29.479946 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:29 crc kubenswrapper[4950]: E0318 20:08:29.480056 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:29 crc kubenswrapper[4950]: E0318 20:08:29.480153 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.480264 4950 scope.go:117] "RemoveContainer" containerID="a9ff70a20c13a291644f12fd70e8d676e7e7266712039a4bcbacb0d73963d7f6" Mar 18 20:08:29 crc kubenswrapper[4950]: E0318 20:08:29.492024 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:29Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.496114 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.496158 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.496169 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.496190 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.496203 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:29Z","lastTransitionTime":"2026-03-18T20:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.503268 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:29Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:29 crc kubenswrapper[4950]: E0318 20:08:29.509807 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:29Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:29 crc kubenswrapper[4950]: E0318 20:08:29.509986 4950 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.519550 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:29Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.545541 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9ff70a20c13a291644f12fd70e8d676e7e7266712039a4bcbacb0d73963d7f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9ff70a20c13a291644f12fd70e8d676e7e7266712039a4bcbacb0d73963d7f6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:14Z\\\",\\\"message\\\":\\\"TrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI0318 20:08:14.221326 6825 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0318 20:08:14.221340 6825 lb_config.go:1031] Cluster endpoints for openshift-etcd-operator/metrics for network=default are: map[]\\\\nI0318 20:08:14.221361 6825 services_controller.go:443] Built service openshift-etcd-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.188\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0318 20:08:14.221380 6825 services_controller.go:444] Built service openshift-etcd-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0318 20:08:14.221390 6825 services_controller.go:445] Built service openshift-etcd-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF0318 20:08:14.221450 6825 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p2zdp_openshift-ovn-kubernetes(83ee6659-cfde-4844-bb92-ccb9b26451ea)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:29Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.562546 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:29Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.573387 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:29Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.587976 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52b6aa3dad160c747ebc18d26d2d504c4c72726d958a410ebe4ee008ca4741af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:29Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.596986 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d09a228c-313f-4b55-a298-51ebd404cfd5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2cjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:29Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.613052 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"831bf4ca-f84c-48d2-9054-410480eb986b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:07:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 20:07:35.272898 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 20:07:35.273030 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 20:07:35.273947 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2398488620/tls.crt::/tmp/serving-cert-2398488620/tls.key\\\\\\\"\\\\nI0318 20:07:35.779001 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 20:07:35.783097 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 20:07:35.783124 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 20:07:35.783155 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 20:07:35.783174 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 20:07:35.791487 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 20:07:35.791507 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791512 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791517 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 20:07:35.791519 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 20:07:35.791522 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 20:07:35.791525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 20:07:35.791560 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 20:07:35.793966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:07:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:29Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.624219 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:29Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.639541 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:29Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.653765 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:29Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.667166 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:29Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.681138 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:29Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.697017 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:29Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:29 crc kubenswrapper[4950]: I0318 20:08:29.707128 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49a491f-1edc-4a40-b64f-147c0d87481f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b01d79ecfba0d5984e7415ca4db790ef1635698da9792b6f1e6161a868f243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122df2a4bd1b73f3890ecaef0f5840c16b127b2b7a8b749ee84f2947334b6aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5p989\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:29Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:30 crc kubenswrapper[4950]: I0318 20:08:30.368634 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p2zdp_83ee6659-cfde-4844-bb92-ccb9b26451ea/ovnkube-controller/2.log" Mar 18 20:08:30 crc kubenswrapper[4950]: I0318 20:08:30.369568 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p2zdp_83ee6659-cfde-4844-bb92-ccb9b26451ea/ovnkube-controller/1.log" Mar 18 20:08:30 crc kubenswrapper[4950]: I0318 20:08:30.374125 4950 generic.go:334] "Generic (PLEG): container finished" podID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerID="88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7" exitCode=1 Mar 18 20:08:30 crc kubenswrapper[4950]: I0318 20:08:30.374193 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerDied","Data":"88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7"} Mar 18 20:08:30 crc kubenswrapper[4950]: I0318 20:08:30.374245 4950 scope.go:117] "RemoveContainer" containerID="a9ff70a20c13a291644f12fd70e8d676e7e7266712039a4bcbacb0d73963d7f6" Mar 18 20:08:30 crc kubenswrapper[4950]: I0318 20:08:30.375493 4950 scope.go:117] "RemoveContainer" containerID="88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7" Mar 18 20:08:30 crc kubenswrapper[4950]: E0318 20:08:30.375756 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p2zdp_openshift-ovn-kubernetes(83ee6659-cfde-4844-bb92-ccb9b26451ea)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" Mar 18 20:08:30 crc kubenswrapper[4950]: I0318 20:08:30.418282 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9ff70a20c13a291644f12fd70e8d676e7e7266712039a4bcbacb0d73963d7f6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:14Z\\\",\\\"message\\\":\\\"TrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI0318 20:08:14.221326 6825 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0318 20:08:14.221340 6825 lb_config.go:1031] Cluster endpoints for openshift-etcd-operator/metrics for network=default are: map[]\\\\nI0318 20:08:14.221361 6825 services_controller.go:443] Built service openshift-etcd-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.188\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0318 20:08:14.221380 6825 services_controller.go:444] Built service openshift-etcd-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0318 20:08:14.221390 6825 services_controller.go:445] Built service openshift-etcd-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF0318 20:08:14.221450 6825 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:30Z\\\",\\\"message\\\":\\\"election_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0318 20:08:30.254102 7035 services_controller.go:360] Finished syncing service networking-console-plugin on namespace openshift-network-console for network=default : 3.081202ms\\\\nI0318 20:08:30.251407 7035 handler.go:208] Removed *v1.Node event handler 2\\\\nI0318 20:08:30.254135 7035 services_controller.go:360] Finished syncing service cluster-autoscaler-operator on namespace openshift-machine-api for network=default : 3.284578ms\\\\nI0318 20:08:30.254263 7035 services_controller.go:356] Processing sync for service openshift-apiserver-operator/metrics for network=default\\\\nI0318 20:08:30.254332 7035 port_cache.go:96] port-cache(openshift-network-console_networking-console-plugin-85b44fc459-gdk6g): added port \\\\u0026{name:openshift-network-console_networking-console-plugin-85b44fc459-gdk6g uuid:c94130be-172c-477c-88c4-40cc7eba30fe logicalSwitch:crc ips:[0xc007ee7980] mac:[10 88 10 217 0 92] expires:{wall:0 ext:0 loc:\\\\u003cnil\\\\u003e}} with IP: [10.217.0.92/23] and MAC: 0a:58:0a:d9:00:5c\\\\nI0318 20:08:30.254353 7035 pods.go:252] [openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] addLogicalPort took \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:30Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:30 crc kubenswrapper[4950]: I0318 20:08:30.441364 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:30Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:30 crc kubenswrapper[4950]: I0318 20:08:30.462116 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:30Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:30 crc kubenswrapper[4950]: I0318 20:08:30.476920 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d09a228c-313f-4b55-a298-51ebd404cfd5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2cjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:30Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:30 crc kubenswrapper[4950]: I0318 20:08:30.491856 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:30Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:30 crc kubenswrapper[4950]: I0318 20:08:30.506542 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:30Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:30 crc kubenswrapper[4950]: I0318 20:08:30.521840 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52b6aa3dad160c747ebc18d26d2d504c4c72726d958a410ebe4ee008ca4741af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:30Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:30 crc kubenswrapper[4950]: I0318 20:08:30.533201 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:30Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:30 crc kubenswrapper[4950]: I0318 20:08:30.542988 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:30Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:30 crc kubenswrapper[4950]: I0318 20:08:30.558468 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"831bf4ca-f84c-48d2-9054-410480eb986b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:07:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 20:07:35.272898 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 20:07:35.273030 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 20:07:35.273947 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2398488620/tls.crt::/tmp/serving-cert-2398488620/tls.key\\\\\\\"\\\\nI0318 20:07:35.779001 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 20:07:35.783097 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 20:07:35.783124 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 20:07:35.783155 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 20:07:35.783174 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 20:07:35.791487 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 20:07:35.791507 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791512 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791517 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 20:07:35.791519 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 20:07:35.791522 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 20:07:35.791525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 20:07:35.791560 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 20:07:35.793966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:07:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:30Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:30 crc kubenswrapper[4950]: I0318 20:08:30.574293 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:30Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:30 crc kubenswrapper[4950]: I0318 20:08:30.590347 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:30Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:30 crc kubenswrapper[4950]: I0318 20:08:30.605375 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:30Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:30 crc kubenswrapper[4950]: I0318 20:08:30.616436 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:30Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:30 crc kubenswrapper[4950]: I0318 20:08:30.629737 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49a491f-1edc-4a40-b64f-147c0d87481f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b01d79ecfba0d5984e7415ca4db790ef1635698da9792b6f1e6161a868f243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122df2a4bd1b73f3890ecaef0f5840c16b127b2b7a8b749ee84f2947334b6aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5p989\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:30Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:31 crc kubenswrapper[4950]: I0318 20:08:31.381850 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p2zdp_83ee6659-cfde-4844-bb92-ccb9b26451ea/ovnkube-controller/2.log" Mar 18 20:08:31 crc kubenswrapper[4950]: I0318 20:08:31.479243 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:31 crc kubenswrapper[4950]: I0318 20:08:31.479245 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:31 crc kubenswrapper[4950]: I0318 20:08:31.479531 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:31 crc kubenswrapper[4950]: E0318 20:08:31.479448 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:31 crc kubenswrapper[4950]: E0318 20:08:31.479691 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:31 crc kubenswrapper[4950]: E0318 20:08:31.479809 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:31 crc kubenswrapper[4950]: I0318 20:08:31.479276 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:31 crc kubenswrapper[4950]: E0318 20:08:31.481047 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:32 crc kubenswrapper[4950]: I0318 20:08:32.494389 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 18 20:08:32 crc kubenswrapper[4950]: E0318 20:08:32.554737 4950 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 20:08:33 crc kubenswrapper[4950]: I0318 20:08:33.479122 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:33 crc kubenswrapper[4950]: E0318 20:08:33.479308 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:33 crc kubenswrapper[4950]: I0318 20:08:33.479671 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:33 crc kubenswrapper[4950]: I0318 20:08:33.479743 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:33 crc kubenswrapper[4950]: E0318 20:08:33.480585 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:33 crc kubenswrapper[4950]: I0318 20:08:33.480260 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:33 crc kubenswrapper[4950]: E0318 20:08:33.480955 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:33 crc kubenswrapper[4950]: E0318 20:08:33.481047 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:33 crc kubenswrapper[4950]: I0318 20:08:33.918659 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:08:33 crc kubenswrapper[4950]: I0318 20:08:33.919810 4950 scope.go:117] "RemoveContainer" containerID="88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7" Mar 18 20:08:33 crc kubenswrapper[4950]: E0318 20:08:33.920083 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p2zdp_openshift-ovn-kubernetes(83ee6659-cfde-4844-bb92-ccb9b26451ea)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" Mar 18 20:08:33 crc kubenswrapper[4950]: I0318 20:08:33.942854 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36099937-1da3-4e7b-abf2-171aef148364\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c38d8f163e6271ccc589aac12d588aabc2ed0963ec5d13976b3a56e94c64a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973ac70ac0e653b6a420f5bfb070f6304dc2d828022a97a5804ac3c050187888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4987ceacf9ceda23f6c9f9b3b79cc2984e0e088b89d99ba7c2b0560baf5f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea96d85a7524dbdb86733ec82ee52e98ad1de83fc26e1651d039637e489523f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea96d85a7524dbdb86733ec82ee52e98ad1de83fc26e1651d039637e489523f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:33Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:33 crc kubenswrapper[4950]: I0318 20:08:33.957709 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:33Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:33 crc kubenswrapper[4950]: I0318 20:08:33.980688 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:33Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:34 crc kubenswrapper[4950]: I0318 20:08:34.011753 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:30Z\\\",\\\"message\\\":\\\"election_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0318 20:08:30.254102 7035 services_controller.go:360] Finished syncing service networking-console-plugin on namespace openshift-network-console for network=default : 3.081202ms\\\\nI0318 20:08:30.251407 7035 handler.go:208] Removed *v1.Node event handler 2\\\\nI0318 20:08:30.254135 7035 services_controller.go:360] Finished syncing service cluster-autoscaler-operator on namespace openshift-machine-api for network=default : 3.284578ms\\\\nI0318 20:08:30.254263 7035 services_controller.go:356] Processing sync for service openshift-apiserver-operator/metrics for network=default\\\\nI0318 20:08:30.254332 7035 port_cache.go:96] port-cache(openshift-network-console_networking-console-plugin-85b44fc459-gdk6g): added port \\\\u0026{name:openshift-network-console_networking-console-plugin-85b44fc459-gdk6g uuid:c94130be-172c-477c-88c4-40cc7eba30fe logicalSwitch:crc ips:[0xc007ee7980] mac:[10 88 10 217 0 92] expires:{wall:0 ext:0 loc:\\\\u003cnil\\\\u003e}} with IP: [10.217.0.92/23] and MAC: 0a:58:0a:d9:00:5c\\\\nI0318 20:08:30.254353 7035 pods.go:252] [openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] addLogicalPort took \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p2zdp_openshift-ovn-kubernetes(83ee6659-cfde-4844-bb92-ccb9b26451ea)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:34Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:34 crc kubenswrapper[4950]: I0318 20:08:34.034402 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:34Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:34 crc kubenswrapper[4950]: I0318 20:08:34.064364 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:34Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:34 crc kubenswrapper[4950]: I0318 20:08:34.080400 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52b6aa3dad160c747ebc18d26d2d504c4c72726d958a410ebe4ee008ca4741af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:34Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:34 crc kubenswrapper[4950]: I0318 20:08:34.094283 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d09a228c-313f-4b55-a298-51ebd404cfd5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2cjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:34Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:34 crc kubenswrapper[4950]: I0318 20:08:34.103602 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:34Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:34 crc kubenswrapper[4950]: I0318 20:08:34.115560 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"831bf4ca-f84c-48d2-9054-410480eb986b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:07:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 20:07:35.272898 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 20:07:35.273030 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 20:07:35.273947 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2398488620/tls.crt::/tmp/serving-cert-2398488620/tls.key\\\\\\\"\\\\nI0318 20:07:35.779001 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 20:07:35.783097 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 20:07:35.783124 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 20:07:35.783155 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 20:07:35.783174 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 20:07:35.791487 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 20:07:35.791507 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791512 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791517 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 20:07:35.791519 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 20:07:35.791522 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 20:07:35.791525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 20:07:35.791560 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 20:07:35.793966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:07:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:34Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:34 crc kubenswrapper[4950]: I0318 20:08:34.128348 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:34Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:34 crc kubenswrapper[4950]: I0318 20:08:34.141172 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:34Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:34 crc kubenswrapper[4950]: I0318 20:08:34.152643 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:34Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:34 crc kubenswrapper[4950]: I0318 20:08:34.162691 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:34Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:34 crc kubenswrapper[4950]: I0318 20:08:34.174089 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49a491f-1edc-4a40-b64f-147c0d87481f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b01d79ecfba0d5984e7415ca4db790ef1635698da9792b6f1e6161a868f243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122df2a4bd1b73f3890ecaef0f5840c16b127b2b7a8b749ee84f2947334b6aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5p989\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:34Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:34 crc kubenswrapper[4950]: I0318 20:08:34.187403 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:34Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:35 crc kubenswrapper[4950]: I0318 20:08:35.391917 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:08:35 crc kubenswrapper[4950]: E0318 20:08:35.392202 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:09:07.392157465 +0000 UTC m=+160.632999373 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:08:35 crc kubenswrapper[4950]: E0318 20:08:35.392354 4950 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 20:08:35 crc kubenswrapper[4950]: E0318 20:08:35.392506 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 20:09:07.392475164 +0000 UTC m=+160.633317082 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 20:08:35 crc kubenswrapper[4950]: I0318 20:08:35.392163 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:35 crc kubenswrapper[4950]: I0318 20:08:35.392812 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:35 crc kubenswrapper[4950]: E0318 20:08:35.392959 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 20:08:35 crc kubenswrapper[4950]: E0318 20:08:35.392983 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 20:08:35 crc kubenswrapper[4950]: E0318 20:08:35.393003 4950 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:35 crc kubenswrapper[4950]: E0318 20:08:35.393064 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 20:09:07.393048979 +0000 UTC m=+160.633890887 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:35 crc kubenswrapper[4950]: I0318 20:08:35.393647 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:35 crc kubenswrapper[4950]: E0318 20:08:35.393782 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 20:08:35 crc kubenswrapper[4950]: E0318 20:08:35.393805 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 20:08:35 crc kubenswrapper[4950]: E0318 20:08:35.393823 4950 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:35 crc kubenswrapper[4950]: E0318 20:08:35.393880 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 20:09:07.393864491 +0000 UTC m=+160.634706399 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:08:35 crc kubenswrapper[4950]: I0318 20:08:35.393939 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:35 crc kubenswrapper[4950]: E0318 20:08:35.394048 4950 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 20:08:35 crc kubenswrapper[4950]: E0318 20:08:35.394137 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 20:09:07.394115647 +0000 UTC m=+160.634957555 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 20:08:35 crc kubenswrapper[4950]: I0318 20:08:35.479768 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:35 crc kubenswrapper[4950]: I0318 20:08:35.479821 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:35 crc kubenswrapper[4950]: E0318 20:08:35.480213 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:35 crc kubenswrapper[4950]: I0318 20:08:35.479853 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:35 crc kubenswrapper[4950]: I0318 20:08:35.479822 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:35 crc kubenswrapper[4950]: E0318 20:08:35.480395 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:35 crc kubenswrapper[4950]: E0318 20:08:35.480511 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:35 crc kubenswrapper[4950]: E0318 20:08:35.480623 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:37 crc kubenswrapper[4950]: I0318 20:08:37.478669 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:37 crc kubenswrapper[4950]: I0318 20:08:37.478736 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:37 crc kubenswrapper[4950]: E0318 20:08:37.478846 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:37 crc kubenswrapper[4950]: I0318 20:08:37.478887 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:37 crc kubenswrapper[4950]: I0318 20:08:37.478933 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:37 crc kubenswrapper[4950]: E0318 20:08:37.479090 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:37 crc kubenswrapper[4950]: E0318 20:08:37.479224 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:37 crc kubenswrapper[4950]: E0318 20:08:37.479306 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:37 crc kubenswrapper[4950]: I0318 20:08:37.501212 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:37Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:37 crc kubenswrapper[4950]: I0318 20:08:37.524076 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52b6aa3dad160c747ebc18d26d2d504c4c72726d958a410ebe4ee008ca4741af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:37Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:37 crc kubenswrapper[4950]: I0318 20:08:37.544356 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d09a228c-313f-4b55-a298-51ebd404cfd5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2cjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:37Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:37 crc kubenswrapper[4950]: E0318 20:08:37.555631 4950 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 20:08:37 crc kubenswrapper[4950]: I0318 20:08:37.578014 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:37Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:37 crc kubenswrapper[4950]: I0318 20:08:37.596023 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:37Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:37 crc kubenswrapper[4950]: I0318 20:08:37.616554 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:37Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:37 crc kubenswrapper[4950]: I0318 20:08:37.633019 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:37Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:37 crc kubenswrapper[4950]: I0318 20:08:37.646242 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:37Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:37 crc kubenswrapper[4950]: I0318 20:08:37.663332 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"831bf4ca-f84c-48d2-9054-410480eb986b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:07:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 20:07:35.272898 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 20:07:35.273030 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 20:07:35.273947 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2398488620/tls.crt::/tmp/serving-cert-2398488620/tls.key\\\\\\\"\\\\nI0318 20:07:35.779001 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 20:07:35.783097 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 20:07:35.783124 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 20:07:35.783155 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 20:07:35.783174 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 20:07:35.791487 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 20:07:35.791507 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791512 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791517 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 20:07:35.791519 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 20:07:35.791522 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 20:07:35.791525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 20:07:35.791560 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 20:07:35.793966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:07:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:37Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:37 crc kubenswrapper[4950]: I0318 20:08:37.676253 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:37Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:37 crc kubenswrapper[4950]: I0318 20:08:37.685395 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:37Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:37 crc kubenswrapper[4950]: I0318 20:08:37.694824 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49a491f-1edc-4a40-b64f-147c0d87481f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b01d79ecfba0d5984e7415ca4db790ef1635698da9792b6f1e6161a868f243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122df2a4bd1b73f3890ecaef0f5840c16b127b2b7a8b749ee84f2947334b6aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5p989\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:37Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:37 crc kubenswrapper[4950]: I0318 20:08:37.704100 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:37Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:37 crc kubenswrapper[4950]: I0318 20:08:37.723391 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:37Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:37 crc kubenswrapper[4950]: I0318 20:08:37.747371 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:30Z\\\",\\\"message\\\":\\\"election_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0318 20:08:30.254102 7035 services_controller.go:360] Finished syncing service networking-console-plugin on namespace openshift-network-console for network=default : 3.081202ms\\\\nI0318 20:08:30.251407 7035 handler.go:208] Removed *v1.Node event handler 2\\\\nI0318 20:08:30.254135 7035 services_controller.go:360] Finished syncing service cluster-autoscaler-operator on namespace openshift-machine-api for network=default : 3.284578ms\\\\nI0318 20:08:30.254263 7035 services_controller.go:356] Processing sync for service openshift-apiserver-operator/metrics for network=default\\\\nI0318 20:08:30.254332 7035 port_cache.go:96] port-cache(openshift-network-console_networking-console-plugin-85b44fc459-gdk6g): added port \\\\u0026{name:openshift-network-console_networking-console-plugin-85b44fc459-gdk6g uuid:c94130be-172c-477c-88c4-40cc7eba30fe logicalSwitch:crc ips:[0xc007ee7980] mac:[10 88 10 217 0 92] expires:{wall:0 ext:0 loc:\\\\u003cnil\\\\u003e}} with IP: [10.217.0.92/23] and MAC: 0a:58:0a:d9:00:5c\\\\nI0318 20:08:30.254353 7035 pods.go:252] [openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] addLogicalPort took \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p2zdp_openshift-ovn-kubernetes(83ee6659-cfde-4844-bb92-ccb9b26451ea)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:37Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:37 crc kubenswrapper[4950]: I0318 20:08:37.760110 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36099937-1da3-4e7b-abf2-171aef148364\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c38d8f163e6271ccc589aac12d588aabc2ed0963ec5d13976b3a56e94c64a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973ac70ac0e653b6a420f5bfb070f6304dc2d828022a97a5804ac3c050187888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4987ceacf9ceda23f6c9f9b3b79cc2984e0e088b89d99ba7c2b0560baf5f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea96d85a7524dbdb86733ec82ee52e98ad1de83fc26e1651d039637e489523f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea96d85a7524dbdb86733ec82ee52e98ad1de83fc26e1651d039637e489523f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:37Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:38 crc kubenswrapper[4950]: I0318 20:08:38.696178 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:08:38 crc kubenswrapper[4950]: I0318 20:08:38.718466 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36099937-1da3-4e7b-abf2-171aef148364\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c38d8f163e6271ccc589aac12d588aabc2ed0963ec5d13976b3a56e94c64a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973ac70ac0e653b6a420f5bfb070f6304dc2d828022a97a5804ac3c050187888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4987ceacf9ceda23f6c9f9b3b79cc2984e0e088b89d99ba7c2b0560baf5f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea96d85a7524dbdb86733ec82ee52e98ad1de83fc26e1651d039637e489523f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea96d85a7524dbdb86733ec82ee52e98ad1de83fc26e1651d039637e489523f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:38Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:38 crc kubenswrapper[4950]: I0318 20:08:38.738089 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:38Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:38 crc kubenswrapper[4950]: I0318 20:08:38.757496 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:38Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:38 crc kubenswrapper[4950]: I0318 20:08:38.786312 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:30Z\\\",\\\"message\\\":\\\"election_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0318 20:08:30.254102 7035 services_controller.go:360] Finished syncing service networking-console-plugin on namespace openshift-network-console for network=default : 3.081202ms\\\\nI0318 20:08:30.251407 7035 handler.go:208] Removed *v1.Node event handler 2\\\\nI0318 20:08:30.254135 7035 services_controller.go:360] Finished syncing service cluster-autoscaler-operator on namespace openshift-machine-api for network=default : 3.284578ms\\\\nI0318 20:08:30.254263 7035 services_controller.go:356] Processing sync for service openshift-apiserver-operator/metrics for network=default\\\\nI0318 20:08:30.254332 7035 port_cache.go:96] port-cache(openshift-network-console_networking-console-plugin-85b44fc459-gdk6g): added port \\\\u0026{name:openshift-network-console_networking-console-plugin-85b44fc459-gdk6g uuid:c94130be-172c-477c-88c4-40cc7eba30fe logicalSwitch:crc ips:[0xc007ee7980] mac:[10 88 10 217 0 92] expires:{wall:0 ext:0 loc:\\\\u003cnil\\\\u003e}} with IP: [10.217.0.92/23] and MAC: 0a:58:0a:d9:00:5c\\\\nI0318 20:08:30.254353 7035 pods.go:252] [openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] addLogicalPort took \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p2zdp_openshift-ovn-kubernetes(83ee6659-cfde-4844-bb92-ccb9b26451ea)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:38Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:38 crc kubenswrapper[4950]: I0318 20:08:38.807242 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:38Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:38 crc kubenswrapper[4950]: I0318 20:08:38.845001 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:38Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:38 crc kubenswrapper[4950]: I0318 20:08:38.868961 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52b6aa3dad160c747ebc18d26d2d504c4c72726d958a410ebe4ee008ca4741af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:38Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:38 crc kubenswrapper[4950]: I0318 20:08:38.887323 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d09a228c-313f-4b55-a298-51ebd404cfd5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2cjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:38Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:38 crc kubenswrapper[4950]: I0318 20:08:38.906640 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"831bf4ca-f84c-48d2-9054-410480eb986b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:07:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 20:07:35.272898 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 20:07:35.273030 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 20:07:35.273947 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2398488620/tls.crt::/tmp/serving-cert-2398488620/tls.key\\\\\\\"\\\\nI0318 20:07:35.779001 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 20:07:35.783097 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 20:07:35.783124 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 20:07:35.783155 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 20:07:35.783174 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 20:07:35.791487 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 20:07:35.791507 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791512 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791517 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 20:07:35.791519 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 20:07:35.791522 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 20:07:35.791525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 20:07:35.791560 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 20:07:35.793966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:07:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:38Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:38 crc kubenswrapper[4950]: I0318 20:08:38.926144 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:38Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:38 crc kubenswrapper[4950]: I0318 20:08:38.944844 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:38Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:38 crc kubenswrapper[4950]: I0318 20:08:38.964204 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:38Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:38 crc kubenswrapper[4950]: I0318 20:08:38.980831 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:38Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.002203 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:38Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.019732 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:39Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.039086 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49a491f-1edc-4a40-b64f-147c0d87481f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b01d79ecfba0d5984e7415ca4db790ef1635698da9792b6f1e6161a868f243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122df2a4bd1b73f3890ecaef0f5840c16b127b2b7a8b749ee84f2947334b6aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5p989\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:39Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.479502 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.479567 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.479584 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:39 crc kubenswrapper[4950]: E0318 20:08:39.479675 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.479946 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:39 crc kubenswrapper[4950]: E0318 20:08:39.480043 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:39 crc kubenswrapper[4950]: E0318 20:08:39.480278 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:39 crc kubenswrapper[4950]: E0318 20:08:39.480803 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.611100 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.611468 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.611702 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.611877 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.612026 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:39Z","lastTransitionTime":"2026-03-18T20:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:39 crc kubenswrapper[4950]: E0318 20:08:39.634170 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:39Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.639721 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.639782 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.639805 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.639835 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.639856 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:39Z","lastTransitionTime":"2026-03-18T20:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:39 crc kubenswrapper[4950]: E0318 20:08:39.660644 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:39Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.665982 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.666050 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.666074 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.666104 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.666128 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:39Z","lastTransitionTime":"2026-03-18T20:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:39 crc kubenswrapper[4950]: E0318 20:08:39.689262 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:39Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.694926 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.694989 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.695007 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.695032 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.695050 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:39Z","lastTransitionTime":"2026-03-18T20:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:39 crc kubenswrapper[4950]: E0318 20:08:39.715697 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:39Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.721095 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.721311 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.721496 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.721724 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:39 crc kubenswrapper[4950]: I0318 20:08:39.721869 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:39Z","lastTransitionTime":"2026-03-18T20:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:39 crc kubenswrapper[4950]: E0318 20:08:39.742677 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:39Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:39 crc kubenswrapper[4950]: E0318 20:08:39.743302 4950 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 20:08:41 crc kubenswrapper[4950]: I0318 20:08:41.479380 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:41 crc kubenswrapper[4950]: I0318 20:08:41.479476 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:41 crc kubenswrapper[4950]: E0318 20:08:41.479584 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:41 crc kubenswrapper[4950]: I0318 20:08:41.479386 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:41 crc kubenswrapper[4950]: E0318 20:08:41.479705 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:41 crc kubenswrapper[4950]: I0318 20:08:41.479785 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:41 crc kubenswrapper[4950]: E0318 20:08:41.479798 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:41 crc kubenswrapper[4950]: E0318 20:08:41.479986 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:42 crc kubenswrapper[4950]: I0318 20:08:42.067324 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs\") pod \"network-metrics-daemon-l2cjn\" (UID: \"d09a228c-313f-4b55-a298-51ebd404cfd5\") " pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:42 crc kubenswrapper[4950]: E0318 20:08:42.067491 4950 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 20:08:42 crc kubenswrapper[4950]: E0318 20:08:42.067543 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs podName:d09a228c-313f-4b55-a298-51ebd404cfd5 nodeName:}" failed. No retries permitted until 2026-03-18 20:09:14.067529028 +0000 UTC m=+167.308370896 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs") pod "network-metrics-daemon-l2cjn" (UID: "d09a228c-313f-4b55-a298-51ebd404cfd5") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 20:08:42 crc kubenswrapper[4950]: I0318 20:08:42.495877 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 18 20:08:42 crc kubenswrapper[4950]: E0318 20:08:42.557064 4950 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 20:08:43 crc kubenswrapper[4950]: I0318 20:08:43.479129 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:43 crc kubenswrapper[4950]: I0318 20:08:43.479214 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:43 crc kubenswrapper[4950]: E0318 20:08:43.479279 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:43 crc kubenswrapper[4950]: I0318 20:08:43.479313 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:43 crc kubenswrapper[4950]: E0318 20:08:43.479721 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:43 crc kubenswrapper[4950]: E0318 20:08:43.479724 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:43 crc kubenswrapper[4950]: I0318 20:08:43.479883 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:43 crc kubenswrapper[4950]: E0318 20:08:43.480002 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:45 crc kubenswrapper[4950]: I0318 20:08:45.478670 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:45 crc kubenswrapper[4950]: I0318 20:08:45.478728 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:45 crc kubenswrapper[4950]: E0318 20:08:45.478773 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:45 crc kubenswrapper[4950]: I0318 20:08:45.478826 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:45 crc kubenswrapper[4950]: I0318 20:08:45.478744 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:45 crc kubenswrapper[4950]: E0318 20:08:45.478957 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:45 crc kubenswrapper[4950]: E0318 20:08:45.479012 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:45 crc kubenswrapper[4950]: E0318 20:08:45.479129 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:47 crc kubenswrapper[4950]: I0318 20:08:47.479530 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:47 crc kubenswrapper[4950]: E0318 20:08:47.480965 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:47 crc kubenswrapper[4950]: I0318 20:08:47.479845 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:47 crc kubenswrapper[4950]: E0318 20:08:47.481756 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:47 crc kubenswrapper[4950]: I0318 20:08:47.479888 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:47 crc kubenswrapper[4950]: E0318 20:08:47.482572 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:47 crc kubenswrapper[4950]: I0318 20:08:47.479786 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:47 crc kubenswrapper[4950]: E0318 20:08:47.483835 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:47 crc kubenswrapper[4950]: I0318 20:08:47.502553 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:47Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:47 crc kubenswrapper[4950]: I0318 20:08:47.523079 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:47Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:47 crc kubenswrapper[4950]: I0318 20:08:47.548226 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52b6aa3dad160c747ebc18d26d2d504c4c72726d958a410ebe4ee008ca4741af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:47Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:47 crc kubenswrapper[4950]: E0318 20:08:47.558518 4950 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 20:08:47 crc kubenswrapper[4950]: I0318 20:08:47.567624 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d09a228c-313f-4b55-a298-51ebd404cfd5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2cjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:47Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:47 crc kubenswrapper[4950]: I0318 20:08:47.586228 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:47Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:47 crc kubenswrapper[4950]: I0318 20:08:47.609370 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"831bf4ca-f84c-48d2-9054-410480eb986b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:07:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 20:07:35.272898 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 20:07:35.273030 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 20:07:35.273947 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2398488620/tls.crt::/tmp/serving-cert-2398488620/tls.key\\\\\\\"\\\\nI0318 20:07:35.779001 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 20:07:35.783097 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 20:07:35.783124 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 20:07:35.783155 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 20:07:35.783174 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 20:07:35.791487 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 20:07:35.791507 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791512 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791517 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 20:07:35.791519 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 20:07:35.791522 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 20:07:35.791525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 20:07:35.791560 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 20:07:35.793966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:07:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:47Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:47 crc kubenswrapper[4950]: I0318 20:08:47.628088 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:47Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:47 crc kubenswrapper[4950]: I0318 20:08:47.647597 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:47Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:47 crc kubenswrapper[4950]: I0318 20:08:47.665550 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:47Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:47 crc kubenswrapper[4950]: I0318 20:08:47.684144 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:47Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:47 crc kubenswrapper[4950]: I0318 20:08:47.703319 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49a491f-1edc-4a40-b64f-147c0d87481f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b01d79ecfba0d5984e7415ca4db790ef1635698da9792b6f1e6161a868f243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122df2a4bd1b73f3890ecaef0f5840c16b127b2b7a8b749ee84f2947334b6aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5p989\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:47Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:47 crc kubenswrapper[4950]: I0318 20:08:47.726595 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de376a3-1738-4cd5-8b36-2e4636215912\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ced76909ac5d1fd3962332517244a448367b89662e932e93daa0de34f70b4a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c09d46a4d315127ff42d2eb02cf657f6fbe2bdc23eb05addee4f8da9d736498\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:06:57Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 20:06:29.551493 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 20:06:29.554866 1 observer_polling.go:159] Starting file observer\\\\nI0318 20:06:29.586793 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 20:06:29.591515 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0318 20:06:57.967959 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0318 20:06:57.968124 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:57Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e662b624a83fcb0b0eab1048193fd00ce4a744f1a4653d776e0584aec856c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e5979160660396987502257fbb0f04771de4c0de21fe3d1be16deb1da3aacf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e89899cf4b5cbc8566cda6ebfc9a858c7bb0f743bc4ed17b63de28c47fcfa30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:47Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:47 crc kubenswrapper[4950]: I0318 20:08:47.751910 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:47Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:47 crc kubenswrapper[4950]: I0318 20:08:47.772799 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36099937-1da3-4e7b-abf2-171aef148364\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c38d8f163e6271ccc589aac12d588aabc2ed0963ec5d13976b3a56e94c64a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973ac70ac0e653b6a420f5bfb070f6304dc2d828022a97a5804ac3c050187888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4987ceacf9ceda23f6c9f9b3b79cc2984e0e088b89d99ba7c2b0560baf5f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea96d85a7524dbdb86733ec82ee52e98ad1de83fc26e1651d039637e489523f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea96d85a7524dbdb86733ec82ee52e98ad1de83fc26e1651d039637e489523f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:47Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:47 crc kubenswrapper[4950]: I0318 20:08:47.791864 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:47Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:47 crc kubenswrapper[4950]: I0318 20:08:47.812946 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:47Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:47 crc kubenswrapper[4950]: I0318 20:08:47.853630 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:30Z\\\",\\\"message\\\":\\\"election_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0318 20:08:30.254102 7035 services_controller.go:360] Finished syncing service networking-console-plugin on namespace openshift-network-console for network=default : 3.081202ms\\\\nI0318 20:08:30.251407 7035 handler.go:208] Removed *v1.Node event handler 2\\\\nI0318 20:08:30.254135 7035 services_controller.go:360] Finished syncing service cluster-autoscaler-operator on namespace openshift-machine-api for network=default : 3.284578ms\\\\nI0318 20:08:30.254263 7035 services_controller.go:356] Processing sync for service openshift-apiserver-operator/metrics for network=default\\\\nI0318 20:08:30.254332 7035 port_cache.go:96] port-cache(openshift-network-console_networking-console-plugin-85b44fc459-gdk6g): added port \\\\u0026{name:openshift-network-console_networking-console-plugin-85b44fc459-gdk6g uuid:c94130be-172c-477c-88c4-40cc7eba30fe logicalSwitch:crc ips:[0xc007ee7980] mac:[10 88 10 217 0 92] expires:{wall:0 ext:0 loc:\\\\u003cnil\\\\u003e}} with IP: [10.217.0.92/23] and MAC: 0a:58:0a:d9:00:5c\\\\nI0318 20:08:30.254353 7035 pods.go:252] [openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] addLogicalPort took \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p2zdp_openshift-ovn-kubernetes(83ee6659-cfde-4844-bb92-ccb9b26451ea)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:47Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:48 crc kubenswrapper[4950]: I0318 20:08:48.480584 4950 scope.go:117] "RemoveContainer" containerID="88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7" Mar 18 20:08:48 crc kubenswrapper[4950]: E0318 20:08:48.480817 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p2zdp_openshift-ovn-kubernetes(83ee6659-cfde-4844-bb92-ccb9b26451ea)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" Mar 18 20:08:49 crc kubenswrapper[4950]: I0318 20:08:49.479247 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:49 crc kubenswrapper[4950]: E0318 20:08:49.479877 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:49 crc kubenswrapper[4950]: I0318 20:08:49.479394 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:49 crc kubenswrapper[4950]: I0318 20:08:49.479310 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:49 crc kubenswrapper[4950]: E0318 20:08:49.480067 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:49 crc kubenswrapper[4950]: E0318 20:08:49.480260 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:49 crc kubenswrapper[4950]: I0318 20:08:49.479316 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:49 crc kubenswrapper[4950]: E0318 20:08:49.480827 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:49 crc kubenswrapper[4950]: I0318 20:08:49.987319 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:49 crc kubenswrapper[4950]: I0318 20:08:49.987381 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:49 crc kubenswrapper[4950]: I0318 20:08:49.987398 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:49 crc kubenswrapper[4950]: I0318 20:08:49.987450 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:49 crc kubenswrapper[4950]: I0318 20:08:49.987468 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:49Z","lastTransitionTime":"2026-03-18T20:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:50 crc kubenswrapper[4950]: E0318 20:08:50.011641 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:50Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:50 crc kubenswrapper[4950]: I0318 20:08:50.020246 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:50 crc kubenswrapper[4950]: I0318 20:08:50.020317 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:50 crc kubenswrapper[4950]: I0318 20:08:50.020341 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:50 crc kubenswrapper[4950]: I0318 20:08:50.020368 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:50 crc kubenswrapper[4950]: I0318 20:08:50.020390 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:50Z","lastTransitionTime":"2026-03-18T20:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:50 crc kubenswrapper[4950]: E0318 20:08:50.043282 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:50Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:50 crc kubenswrapper[4950]: I0318 20:08:50.048683 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:50 crc kubenswrapper[4950]: I0318 20:08:50.048753 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:50 crc kubenswrapper[4950]: I0318 20:08:50.048771 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:50 crc kubenswrapper[4950]: I0318 20:08:50.048797 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:50 crc kubenswrapper[4950]: I0318 20:08:50.048814 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:50Z","lastTransitionTime":"2026-03-18T20:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:50 crc kubenswrapper[4950]: E0318 20:08:50.070107 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:50Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:50 crc kubenswrapper[4950]: I0318 20:08:50.073523 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:50 crc kubenswrapper[4950]: I0318 20:08:50.073590 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:50 crc kubenswrapper[4950]: I0318 20:08:50.073609 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:50 crc kubenswrapper[4950]: I0318 20:08:50.073634 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:50 crc kubenswrapper[4950]: I0318 20:08:50.073653 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:50Z","lastTransitionTime":"2026-03-18T20:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:50 crc kubenswrapper[4950]: E0318 20:08:50.087953 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:50Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:50 crc kubenswrapper[4950]: I0318 20:08:50.091892 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:08:50 crc kubenswrapper[4950]: I0318 20:08:50.091928 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:08:50 crc kubenswrapper[4950]: I0318 20:08:50.091939 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:08:50 crc kubenswrapper[4950]: I0318 20:08:50.091955 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:08:50 crc kubenswrapper[4950]: I0318 20:08:50.091964 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:08:50Z","lastTransitionTime":"2026-03-18T20:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:08:50 crc kubenswrapper[4950]: E0318 20:08:50.104544 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:50Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:50 crc kubenswrapper[4950]: E0318 20:08:50.104673 4950 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.466860 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lqkxm_981403b4-25e5-4ea0-9267-9584719853d6/kube-multus/0.log" Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.466901 4950 generic.go:334] "Generic (PLEG): container finished" podID="981403b4-25e5-4ea0-9267-9584719853d6" containerID="7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3" exitCode=1 Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.466924 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lqkxm" event={"ID":"981403b4-25e5-4ea0-9267-9584719853d6","Type":"ContainerDied","Data":"7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3"} Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.467231 4950 scope.go:117] "RemoveContainer" containerID="7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3" Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.478872 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.478886 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.478883 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:51 crc kubenswrapper[4950]: E0318 20:08:51.479287 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:51 crc kubenswrapper[4950]: E0318 20:08:51.479144 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.478926 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:51 crc kubenswrapper[4950]: E0318 20:08:51.479382 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:51 crc kubenswrapper[4950]: E0318 20:08:51.479483 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.487781 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d09a228c-313f-4b55-a298-51ebd404cfd5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2cjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:51Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.506386 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:51Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.526090 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:51Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.549380 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52b6aa3dad160c747ebc18d26d2d504c4c72726d958a410ebe4ee008ca4741af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:51Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.567016 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:51Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.582202 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:51Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.607026 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"831bf4ca-f84c-48d2-9054-410480eb986b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:07:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 20:07:35.272898 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 20:07:35.273030 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 20:07:35.273947 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2398488620/tls.crt::/tmp/serving-cert-2398488620/tls.key\\\\\\\"\\\\nI0318 20:07:35.779001 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 20:07:35.783097 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 20:07:35.783124 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 20:07:35.783155 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 20:07:35.783174 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 20:07:35.791487 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 20:07:35.791507 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791512 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791517 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 20:07:35.791519 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 20:07:35.791522 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 20:07:35.791525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 20:07:35.791560 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 20:07:35.793966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:07:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:51Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.624393 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:51Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.642265 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:51Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.659867 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:51Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.674932 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:51Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.694198 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49a491f-1edc-4a40-b64f-147c0d87481f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b01d79ecfba0d5984e7415ca4db790ef1635698da9792b6f1e6161a868f243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122df2a4bd1b73f3890ecaef0f5840c16b127b2b7a8b749ee84f2947334b6aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5p989\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:51Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.712826 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de376a3-1738-4cd5-8b36-2e4636215912\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ced76909ac5d1fd3962332517244a448367b89662e932e93daa0de34f70b4a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c09d46a4d315127ff42d2eb02cf657f6fbe2bdc23eb05addee4f8da9d736498\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:06:57Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 20:06:29.551493 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 20:06:29.554866 1 observer_polling.go:159] Starting file observer\\\\nI0318 20:06:29.586793 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 20:06:29.591515 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0318 20:06:57.967959 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0318 20:06:57.968124 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:57Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e662b624a83fcb0b0eab1048193fd00ce4a744f1a4653d776e0584aec856c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e5979160660396987502257fbb0f04771de4c0de21fe3d1be16deb1da3aacf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e89899cf4b5cbc8566cda6ebfc9a858c7bb0f743bc4ed17b63de28c47fcfa30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:51Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.738916 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:30Z\\\",\\\"message\\\":\\\"election_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0318 20:08:30.254102 7035 services_controller.go:360] Finished syncing service networking-console-plugin on namespace openshift-network-console for network=default : 3.081202ms\\\\nI0318 20:08:30.251407 7035 handler.go:208] Removed *v1.Node event handler 2\\\\nI0318 20:08:30.254135 7035 services_controller.go:360] Finished syncing service cluster-autoscaler-operator on namespace openshift-machine-api for network=default : 3.284578ms\\\\nI0318 20:08:30.254263 7035 services_controller.go:356] Processing sync for service openshift-apiserver-operator/metrics for network=default\\\\nI0318 20:08:30.254332 7035 port_cache.go:96] port-cache(openshift-network-console_networking-console-plugin-85b44fc459-gdk6g): added port \\\\u0026{name:openshift-network-console_networking-console-plugin-85b44fc459-gdk6g uuid:c94130be-172c-477c-88c4-40cc7eba30fe logicalSwitch:crc ips:[0xc007ee7980] mac:[10 88 10 217 0 92] expires:{wall:0 ext:0 loc:\\\\u003cnil\\\\u003e}} with IP: [10.217.0.92/23] and MAC: 0a:58:0a:d9:00:5c\\\\nI0318 20:08:30.254353 7035 pods.go:252] [openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] addLogicalPort took \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p2zdp_openshift-ovn-kubernetes(83ee6659-cfde-4844-bb92-ccb9b26451ea)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:51Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.756477 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36099937-1da3-4e7b-abf2-171aef148364\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c38d8f163e6271ccc589aac12d588aabc2ed0963ec5d13976b3a56e94c64a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973ac70ac0e653b6a420f5bfb070f6304dc2d828022a97a5804ac3c050187888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4987ceacf9ceda23f6c9f9b3b79cc2984e0e088b89d99ba7c2b0560baf5f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea96d85a7524dbdb86733ec82ee52e98ad1de83fc26e1651d039637e489523f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea96d85a7524dbdb86733ec82ee52e98ad1de83fc26e1651d039637e489523f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:51Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.774152 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:51Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:51 crc kubenswrapper[4950]: I0318 20:08:51.798457 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:51Z\\\",\\\"message\\\":\\\"2026-03-18T20:08:05+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9214f383-12a6-4ea7-9936-fb82712c11ab\\\\n2026-03-18T20:08:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9214f383-12a6-4ea7-9936-fb82712c11ab to /host/opt/cni/bin/\\\\n2026-03-18T20:08:06Z [verbose] multus-daemon started\\\\n2026-03-18T20:08:06Z [verbose] Readiness Indicator file check\\\\n2026-03-18T20:08:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:51Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:52 crc kubenswrapper[4950]: I0318 20:08:52.472521 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lqkxm_981403b4-25e5-4ea0-9267-9584719853d6/kube-multus/0.log" Mar 18 20:08:52 crc kubenswrapper[4950]: I0318 20:08:52.473555 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lqkxm" event={"ID":"981403b4-25e5-4ea0-9267-9584719853d6","Type":"ContainerStarted","Data":"8b42f10f95ebda7760ce304ba8c5dfbf0e891059750a239f5ea8d0c96b673165"} Mar 18 20:08:52 crc kubenswrapper[4950]: I0318 20:08:52.492627 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36099937-1da3-4e7b-abf2-171aef148364\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c38d8f163e6271ccc589aac12d588aabc2ed0963ec5d13976b3a56e94c64a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973ac70ac0e653b6a420f5bfb070f6304dc2d828022a97a5804ac3c050187888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4987ceacf9ceda23f6c9f9b3b79cc2984e0e088b89d99ba7c2b0560baf5f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea96d85a7524dbdb86733ec82ee52e98ad1de83fc26e1651d039637e489523f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea96d85a7524dbdb86733ec82ee52e98ad1de83fc26e1651d039637e489523f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:52Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:52 crc kubenswrapper[4950]: I0318 20:08:52.514178 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:52Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:52 crc kubenswrapper[4950]: I0318 20:08:52.533779 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b42f10f95ebda7760ce304ba8c5dfbf0e891059750a239f5ea8d0c96b673165\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:51Z\\\",\\\"message\\\":\\\"2026-03-18T20:08:05+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9214f383-12a6-4ea7-9936-fb82712c11ab\\\\n2026-03-18T20:08:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9214f383-12a6-4ea7-9936-fb82712c11ab to /host/opt/cni/bin/\\\\n2026-03-18T20:08:06Z [verbose] multus-daemon started\\\\n2026-03-18T20:08:06Z [verbose] Readiness Indicator file check\\\\n2026-03-18T20:08:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:52Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:52 crc kubenswrapper[4950]: E0318 20:08:52.560046 4950 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 20:08:52 crc kubenswrapper[4950]: I0318 20:08:52.563280 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:30Z\\\",\\\"message\\\":\\\"election_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0318 20:08:30.254102 7035 services_controller.go:360] Finished syncing service networking-console-plugin on namespace openshift-network-console for network=default : 3.081202ms\\\\nI0318 20:08:30.251407 7035 handler.go:208] Removed *v1.Node event handler 2\\\\nI0318 20:08:30.254135 7035 services_controller.go:360] Finished syncing service cluster-autoscaler-operator on namespace openshift-machine-api for network=default : 3.284578ms\\\\nI0318 20:08:30.254263 7035 services_controller.go:356] Processing sync for service openshift-apiserver-operator/metrics for network=default\\\\nI0318 20:08:30.254332 7035 port_cache.go:96] port-cache(openshift-network-console_networking-console-plugin-85b44fc459-gdk6g): added port \\\\u0026{name:openshift-network-console_networking-console-plugin-85b44fc459-gdk6g uuid:c94130be-172c-477c-88c4-40cc7eba30fe logicalSwitch:crc ips:[0xc007ee7980] mac:[10 88 10 217 0 92] expires:{wall:0 ext:0 loc:\\\\u003cnil\\\\u003e}} with IP: [10.217.0.92/23] and MAC: 0a:58:0a:d9:00:5c\\\\nI0318 20:08:30.254353 7035 pods.go:252] [openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] addLogicalPort took \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p2zdp_openshift-ovn-kubernetes(83ee6659-cfde-4844-bb92-ccb9b26451ea)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:52Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:52 crc kubenswrapper[4950]: I0318 20:08:52.584126 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:52Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:52 crc kubenswrapper[4950]: I0318 20:08:52.605618 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:52Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:52 crc kubenswrapper[4950]: I0318 20:08:52.630849 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52b6aa3dad160c747ebc18d26d2d504c4c72726d958a410ebe4ee008ca4741af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:52Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:52 crc kubenswrapper[4950]: I0318 20:08:52.644177 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d09a228c-313f-4b55-a298-51ebd404cfd5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2cjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:52Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:52 crc kubenswrapper[4950]: I0318 20:08:52.662586 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"831bf4ca-f84c-48d2-9054-410480eb986b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:07:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 20:07:35.272898 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 20:07:35.273030 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 20:07:35.273947 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2398488620/tls.crt::/tmp/serving-cert-2398488620/tls.key\\\\\\\"\\\\nI0318 20:07:35.779001 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 20:07:35.783097 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 20:07:35.783124 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 20:07:35.783155 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 20:07:35.783174 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 20:07:35.791487 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 20:07:35.791507 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791512 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791517 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 20:07:35.791519 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 20:07:35.791522 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 20:07:35.791525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 20:07:35.791560 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 20:07:35.793966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:07:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:52Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:52 crc kubenswrapper[4950]: I0318 20:08:52.679196 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:52Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:52 crc kubenswrapper[4950]: I0318 20:08:52.691209 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:52Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:52 crc kubenswrapper[4950]: I0318 20:08:52.704868 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:52Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:52 crc kubenswrapper[4950]: I0318 20:08:52.715128 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:52Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:52 crc kubenswrapper[4950]: I0318 20:08:52.730717 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49a491f-1edc-4a40-b64f-147c0d87481f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b01d79ecfba0d5984e7415ca4db790ef1635698da9792b6f1e6161a868f243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122df2a4bd1b73f3890ecaef0f5840c16b127b2b7a8b749ee84f2947334b6aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5p989\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:52Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:52 crc kubenswrapper[4950]: I0318 20:08:52.746212 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de376a3-1738-4cd5-8b36-2e4636215912\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ced76909ac5d1fd3962332517244a448367b89662e932e93daa0de34f70b4a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c09d46a4d315127ff42d2eb02cf657f6fbe2bdc23eb05addee4f8da9d736498\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:06:57Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 20:06:29.551493 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 20:06:29.554866 1 observer_polling.go:159] Starting file observer\\\\nI0318 20:06:29.586793 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 20:06:29.591515 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0318 20:06:57.967959 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0318 20:06:57.968124 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:57Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e662b624a83fcb0b0eab1048193fd00ce4a744f1a4653d776e0584aec856c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e5979160660396987502257fbb0f04771de4c0de21fe3d1be16deb1da3aacf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e89899cf4b5cbc8566cda6ebfc9a858c7bb0f743bc4ed17b63de28c47fcfa30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:52Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:52 crc kubenswrapper[4950]: I0318 20:08:52.769269 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:52Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:52 crc kubenswrapper[4950]: I0318 20:08:52.783731 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:52Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:53 crc kubenswrapper[4950]: I0318 20:08:53.479521 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:53 crc kubenswrapper[4950]: I0318 20:08:53.479611 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:53 crc kubenswrapper[4950]: I0318 20:08:53.479529 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:53 crc kubenswrapper[4950]: I0318 20:08:53.479526 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:53 crc kubenswrapper[4950]: E0318 20:08:53.479814 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:53 crc kubenswrapper[4950]: E0318 20:08:53.479916 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:53 crc kubenswrapper[4950]: E0318 20:08:53.480358 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:53 crc kubenswrapper[4950]: E0318 20:08:53.480089 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:55 crc kubenswrapper[4950]: I0318 20:08:55.479372 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:55 crc kubenswrapper[4950]: I0318 20:08:55.479494 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:55 crc kubenswrapper[4950]: I0318 20:08:55.479523 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:55 crc kubenswrapper[4950]: E0318 20:08:55.479657 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:55 crc kubenswrapper[4950]: E0318 20:08:55.479772 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:55 crc kubenswrapper[4950]: E0318 20:08:55.479888 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:55 crc kubenswrapper[4950]: I0318 20:08:55.480758 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:55 crc kubenswrapper[4950]: E0318 20:08:55.481123 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:57 crc kubenswrapper[4950]: I0318 20:08:57.479580 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:57 crc kubenswrapper[4950]: E0318 20:08:57.479734 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:57 crc kubenswrapper[4950]: I0318 20:08:57.480056 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:57 crc kubenswrapper[4950]: E0318 20:08:57.480180 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:57 crc kubenswrapper[4950]: I0318 20:08:57.480491 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:57 crc kubenswrapper[4950]: E0318 20:08:57.480616 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:57 crc kubenswrapper[4950]: I0318 20:08:57.482132 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:57 crc kubenswrapper[4950]: E0318 20:08:57.482272 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:57 crc kubenswrapper[4950]: I0318 20:08:57.498157 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b42f10f95ebda7760ce304ba8c5dfbf0e891059750a239f5ea8d0c96b673165\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:51Z\\\",\\\"message\\\":\\\"2026-03-18T20:08:05+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9214f383-12a6-4ea7-9936-fb82712c11ab\\\\n2026-03-18T20:08:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9214f383-12a6-4ea7-9936-fb82712c11ab to /host/opt/cni/bin/\\\\n2026-03-18T20:08:06Z [verbose] multus-daemon started\\\\n2026-03-18T20:08:06Z [verbose] Readiness Indicator file check\\\\n2026-03-18T20:08:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:57Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:57 crc kubenswrapper[4950]: I0318 20:08:57.519401 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:30Z\\\",\\\"message\\\":\\\"election_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0318 20:08:30.254102 7035 services_controller.go:360] Finished syncing service networking-console-plugin on namespace openshift-network-console for network=default : 3.081202ms\\\\nI0318 20:08:30.251407 7035 handler.go:208] Removed *v1.Node event handler 2\\\\nI0318 20:08:30.254135 7035 services_controller.go:360] Finished syncing service cluster-autoscaler-operator on namespace openshift-machine-api for network=default : 3.284578ms\\\\nI0318 20:08:30.254263 7035 services_controller.go:356] Processing sync for service openshift-apiserver-operator/metrics for network=default\\\\nI0318 20:08:30.254332 7035 port_cache.go:96] port-cache(openshift-network-console_networking-console-plugin-85b44fc459-gdk6g): added port \\\\u0026{name:openshift-network-console_networking-console-plugin-85b44fc459-gdk6g uuid:c94130be-172c-477c-88c4-40cc7eba30fe logicalSwitch:crc ips:[0xc007ee7980] mac:[10 88 10 217 0 92] expires:{wall:0 ext:0 loc:\\\\u003cnil\\\\u003e}} with IP: [10.217.0.92/23] and MAC: 0a:58:0a:d9:00:5c\\\\nI0318 20:08:30.254353 7035 pods.go:252] [openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] addLogicalPort took \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p2zdp_openshift-ovn-kubernetes(83ee6659-cfde-4844-bb92-ccb9b26451ea)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:57Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:57 crc kubenswrapper[4950]: I0318 20:08:57.531193 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36099937-1da3-4e7b-abf2-171aef148364\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c38d8f163e6271ccc589aac12d588aabc2ed0963ec5d13976b3a56e94c64a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973ac70ac0e653b6a420f5bfb070f6304dc2d828022a97a5804ac3c050187888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4987ceacf9ceda23f6c9f9b3b79cc2984e0e088b89d99ba7c2b0560baf5f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea96d85a7524dbdb86733ec82ee52e98ad1de83fc26e1651d039637e489523f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea96d85a7524dbdb86733ec82ee52e98ad1de83fc26e1651d039637e489523f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:57Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:57 crc kubenswrapper[4950]: I0318 20:08:57.542240 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:57Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:57 crc kubenswrapper[4950]: I0318 20:08:57.555940 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52b6aa3dad160c747ebc18d26d2d504c4c72726d958a410ebe4ee008ca4741af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:57Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:57 crc kubenswrapper[4950]: E0318 20:08:57.560652 4950 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 20:08:57 crc kubenswrapper[4950]: I0318 20:08:57.570210 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d09a228c-313f-4b55-a298-51ebd404cfd5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2cjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:57Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:57 crc kubenswrapper[4950]: I0318 20:08:57.583295 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:57Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:57 crc kubenswrapper[4950]: I0318 20:08:57.600914 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:57Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:57 crc kubenswrapper[4950]: I0318 20:08:57.614113 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:57Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:57 crc kubenswrapper[4950]: I0318 20:08:57.629175 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:57Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:57 crc kubenswrapper[4950]: I0318 20:08:57.641031 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:57Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:57 crc kubenswrapper[4950]: I0318 20:08:57.658822 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"831bf4ca-f84c-48d2-9054-410480eb986b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:07:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 20:07:35.272898 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 20:07:35.273030 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 20:07:35.273947 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2398488620/tls.crt::/tmp/serving-cert-2398488620/tls.key\\\\\\\"\\\\nI0318 20:07:35.779001 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 20:07:35.783097 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 20:07:35.783124 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 20:07:35.783155 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 20:07:35.783174 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 20:07:35.791487 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 20:07:35.791507 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791512 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791517 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 20:07:35.791519 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 20:07:35.791522 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 20:07:35.791525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 20:07:35.791560 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 20:07:35.793966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:07:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:57Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:57 crc kubenswrapper[4950]: I0318 20:08:57.671240 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:57Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:57 crc kubenswrapper[4950]: I0318 20:08:57.684705 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:57Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:57 crc kubenswrapper[4950]: I0318 20:08:57.696694 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:57Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:57 crc kubenswrapper[4950]: I0318 20:08:57.710981 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49a491f-1edc-4a40-b64f-147c0d87481f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b01d79ecfba0d5984e7415ca4db790ef1635698da9792b6f1e6161a868f243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122df2a4bd1b73f3890ecaef0f5840c16b127b2b7a8b749ee84f2947334b6aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5p989\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:57Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:57 crc kubenswrapper[4950]: I0318 20:08:57.726301 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de376a3-1738-4cd5-8b36-2e4636215912\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ced76909ac5d1fd3962332517244a448367b89662e932e93daa0de34f70b4a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c09d46a4d315127ff42d2eb02cf657f6fbe2bdc23eb05addee4f8da9d736498\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:06:57Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 20:06:29.551493 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 20:06:29.554866 1 observer_polling.go:159] Starting file observer\\\\nI0318 20:06:29.586793 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 20:06:29.591515 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0318 20:06:57.967959 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0318 20:06:57.968124 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:57Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e662b624a83fcb0b0eab1048193fd00ce4a744f1a4653d776e0584aec856c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e5979160660396987502257fbb0f04771de4c0de21fe3d1be16deb1da3aacf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e89899cf4b5cbc8566cda6ebfc9a858c7bb0f743bc4ed17b63de28c47fcfa30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:08:57Z is after 2025-08-24T17:21:41Z" Mar 18 20:08:59 crc kubenswrapper[4950]: I0318 20:08:59.478859 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:08:59 crc kubenswrapper[4950]: I0318 20:08:59.479019 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:08:59 crc kubenswrapper[4950]: I0318 20:08:59.479101 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:08:59 crc kubenswrapper[4950]: E0318 20:08:59.479705 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:08:59 crc kubenswrapper[4950]: E0318 20:08:59.479791 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:08:59 crc kubenswrapper[4950]: E0318 20:08:59.479848 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:08:59 crc kubenswrapper[4950]: I0318 20:08:59.480208 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:08:59 crc kubenswrapper[4950]: E0318 20:08:59.480529 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:08:59 crc kubenswrapper[4950]: I0318 20:08:59.497368 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.463187 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.463269 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.463294 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.463325 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.463348 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:09:00Z","lastTransitionTime":"2026-03-18T20:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:09:00 crc kubenswrapper[4950]: E0318 20:09:00.486822 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:00Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.492120 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.492186 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.492209 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.492240 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.492264 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:09:00Z","lastTransitionTime":"2026-03-18T20:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:09:00 crc kubenswrapper[4950]: E0318 20:09:00.515149 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:00Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.520186 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.520246 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.520270 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.520300 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.520326 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:09:00Z","lastTransitionTime":"2026-03-18T20:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:09:00 crc kubenswrapper[4950]: E0318 20:09:00.539991 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:00Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.544474 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.544535 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.544553 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.544579 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.544595 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:09:00Z","lastTransitionTime":"2026-03-18T20:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:09:00 crc kubenswrapper[4950]: E0318 20:09:00.559536 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:00Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.563017 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.563117 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.563130 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.563145 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:09:00 crc kubenswrapper[4950]: I0318 20:09:00.563153 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:09:00Z","lastTransitionTime":"2026-03-18T20:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:09:00 crc kubenswrapper[4950]: E0318 20:09:00.580244 4950 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-18T20:09:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"28057439-b38c-49e6-908d-4b8eb91daf63\\\",\\\"systemUUID\\\":\\\"96c0961c-2d17-44f9-aece-16ddcd69eb85\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:00Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:00 crc kubenswrapper[4950]: E0318 20:09:00.580364 4950 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 18 20:09:01 crc kubenswrapper[4950]: I0318 20:09:01.479545 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:01 crc kubenswrapper[4950]: E0318 20:09:01.479674 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:01 crc kubenswrapper[4950]: I0318 20:09:01.479783 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:01 crc kubenswrapper[4950]: I0318 20:09:01.479861 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:01 crc kubenswrapper[4950]: E0318 20:09:01.479996 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:01 crc kubenswrapper[4950]: E0318 20:09:01.480032 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:01 crc kubenswrapper[4950]: I0318 20:09:01.480389 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:01 crc kubenswrapper[4950]: E0318 20:09:01.480667 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:02 crc kubenswrapper[4950]: I0318 20:09:02.480899 4950 scope.go:117] "RemoveContainer" containerID="88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7" Mar 18 20:09:02 crc kubenswrapper[4950]: E0318 20:09:02.562565 4950 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.479718 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.479770 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:03 crc kubenswrapper[4950]: E0318 20:09:03.479850 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:03 crc kubenswrapper[4950]: E0318 20:09:03.479967 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.480048 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:03 crc kubenswrapper[4950]: E0318 20:09:03.480136 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.480274 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:03 crc kubenswrapper[4950]: E0318 20:09:03.480539 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.517740 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p2zdp_83ee6659-cfde-4844-bb92-ccb9b26451ea/ovnkube-controller/2.log" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.521144 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerStarted","Data":"544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4"} Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.521633 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.539024 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:03Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.555563 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"831bf4ca-f84c-48d2-9054-410480eb986b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:07:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 20:07:35.272898 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 20:07:35.273030 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 20:07:35.273947 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2398488620/tls.crt::/tmp/serving-cert-2398488620/tls.key\\\\\\\"\\\\nI0318 20:07:35.779001 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 20:07:35.783097 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 20:07:35.783124 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 20:07:35.783155 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 20:07:35.783174 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 20:07:35.791487 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 20:07:35.791507 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791512 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791517 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 20:07:35.791519 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 20:07:35.791522 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 20:07:35.791525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 20:07:35.791560 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 20:07:35.793966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:07:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:03Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.573621 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:03Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.590684 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:03Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.603899 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:03Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.615166 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:03Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.628466 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49a491f-1edc-4a40-b64f-147c0d87481f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b01d79ecfba0d5984e7415ca4db790ef1635698da9792b6f1e6161a868f243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122df2a4bd1b73f3890ecaef0f5840c16b127b2b7a8b749ee84f2947334b6aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5p989\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:03Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.643096 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de376a3-1738-4cd5-8b36-2e4636215912\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ced76909ac5d1fd3962332517244a448367b89662e932e93daa0de34f70b4a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c09d46a4d315127ff42d2eb02cf657f6fbe2bdc23eb05addee4f8da9d736498\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:06:57Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 20:06:29.551493 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 20:06:29.554866 1 observer_polling.go:159] Starting file observer\\\\nI0318 20:06:29.586793 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 20:06:29.591515 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0318 20:06:57.967959 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0318 20:06:57.968124 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:57Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e662b624a83fcb0b0eab1048193fd00ce4a744f1a4653d776e0584aec856c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e5979160660396987502257fbb0f04771de4c0de21fe3d1be16deb1da3aacf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e89899cf4b5cbc8566cda6ebfc9a858c7bb0f743bc4ed17b63de28c47fcfa30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:03Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.657871 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee46e2ea-d8f4-4db4-bb8e-b970b6258216\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88a63d65bec16e87319bce4a3265538d83f89e35ce181fe8013917a17c4c88ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e66d6bfce75c340eebe5f45e62b451508a52e5b83caf85af2c318bb87e47d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e66d6bfce75c340eebe5f45e62b451508a52e5b83caf85af2c318bb87e47d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:03Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.675694 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:03Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.691887 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36099937-1da3-4e7b-abf2-171aef148364\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c38d8f163e6271ccc589aac12d588aabc2ed0963ec5d13976b3a56e94c64a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973ac70ac0e653b6a420f5bfb070f6304dc2d828022a97a5804ac3c050187888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4987ceacf9ceda23f6c9f9b3b79cc2984e0e088b89d99ba7c2b0560baf5f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea96d85a7524dbdb86733ec82ee52e98ad1de83fc26e1651d039637e489523f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea96d85a7524dbdb86733ec82ee52e98ad1de83fc26e1651d039637e489523f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:03Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.709229 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:03Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.724331 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b42f10f95ebda7760ce304ba8c5dfbf0e891059750a239f5ea8d0c96b673165\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:51Z\\\",\\\"message\\\":\\\"2026-03-18T20:08:05+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9214f383-12a6-4ea7-9936-fb82712c11ab\\\\n2026-03-18T20:08:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9214f383-12a6-4ea7-9936-fb82712c11ab to /host/opt/cni/bin/\\\\n2026-03-18T20:08:06Z [verbose] multus-daemon started\\\\n2026-03-18T20:08:06Z [verbose] Readiness Indicator file check\\\\n2026-03-18T20:08:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:03Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.752166 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:30Z\\\",\\\"message\\\":\\\"election_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0318 20:08:30.254102 7035 services_controller.go:360] Finished syncing service networking-console-plugin on namespace openshift-network-console for network=default : 3.081202ms\\\\nI0318 20:08:30.251407 7035 handler.go:208] Removed *v1.Node event handler 2\\\\nI0318 20:08:30.254135 7035 services_controller.go:360] Finished syncing service cluster-autoscaler-operator on namespace openshift-machine-api for network=default : 3.284578ms\\\\nI0318 20:08:30.254263 7035 services_controller.go:356] Processing sync for service openshift-apiserver-operator/metrics for network=default\\\\nI0318 20:08:30.254332 7035 port_cache.go:96] port-cache(openshift-network-console_networking-console-plugin-85b44fc459-gdk6g): added port \\\\u0026{name:openshift-network-console_networking-console-plugin-85b44fc459-gdk6g uuid:c94130be-172c-477c-88c4-40cc7eba30fe logicalSwitch:crc ips:[0xc007ee7980] mac:[10 88 10 217 0 92] expires:{wall:0 ext:0 loc:\\\\u003cnil\\\\u003e}} with IP: [10.217.0.92/23] and MAC: 0a:58:0a:d9:00:5c\\\\nI0318 20:08:30.254353 7035 pods.go:252] [openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] addLogicalPort took \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:03Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.770166 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:03Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.788651 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:03Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.805388 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52b6aa3dad160c747ebc18d26d2d504c4c72726d958a410ebe4ee008ca4741af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:03Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:03 crc kubenswrapper[4950]: I0318 20:09:03.819301 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d09a228c-313f-4b55-a298-51ebd404cfd5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2cjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:03Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.528518 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p2zdp_83ee6659-cfde-4844-bb92-ccb9b26451ea/ovnkube-controller/3.log" Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.530099 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p2zdp_83ee6659-cfde-4844-bb92-ccb9b26451ea/ovnkube-controller/2.log" Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.534498 4950 generic.go:334] "Generic (PLEG): container finished" podID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerID="544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4" exitCode=1 Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.534556 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerDied","Data":"544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4"} Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.534604 4950 scope.go:117] "RemoveContainer" containerID="88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7" Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.535754 4950 scope.go:117] "RemoveContainer" containerID="544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4" Mar 18 20:09:04 crc kubenswrapper[4950]: E0318 20:09:04.536084 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p2zdp_openshift-ovn-kubernetes(83ee6659-cfde-4844-bb92-ccb9b26451ea)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.558468 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d09a228c-313f-4b55-a298-51ebd404cfd5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2cjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:04Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.576760 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:04Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.592107 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:04Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.612445 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52b6aa3dad160c747ebc18d26d2d504c4c72726d958a410ebe4ee008ca4741af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:04Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.626554 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:04Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.639132 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:04Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.820616 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"831bf4ca-f84c-48d2-9054-410480eb986b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:07:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 20:07:35.272898 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 20:07:35.273030 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 20:07:35.273947 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2398488620/tls.crt::/tmp/serving-cert-2398488620/tls.key\\\\\\\"\\\\nI0318 20:07:35.779001 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 20:07:35.783097 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 20:07:35.783124 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 20:07:35.783155 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 20:07:35.783174 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 20:07:35.791487 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 20:07:35.791507 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791512 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791517 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 20:07:35.791519 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 20:07:35.791522 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 20:07:35.791525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 20:07:35.791560 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 20:07:35.793966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:07:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:04Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.834722 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:04Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.848369 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:04Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.863239 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:04Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.874037 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:04Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.886330 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49a491f-1edc-4a40-b64f-147c0d87481f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b01d79ecfba0d5984e7415ca4db790ef1635698da9792b6f1e6161a868f243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122df2a4bd1b73f3890ecaef0f5840c16b127b2b7a8b749ee84f2947334b6aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5p989\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:04Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.902521 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de376a3-1738-4cd5-8b36-2e4636215912\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ced76909ac5d1fd3962332517244a448367b89662e932e93daa0de34f70b4a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c09d46a4d315127ff42d2eb02cf657f6fbe2bdc23eb05addee4f8da9d736498\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:06:57Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 20:06:29.551493 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 20:06:29.554866 1 observer_polling.go:159] Starting file observer\\\\nI0318 20:06:29.586793 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 20:06:29.591515 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0318 20:06:57.967959 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0318 20:06:57.968124 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:57Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e662b624a83fcb0b0eab1048193fd00ce4a744f1a4653d776e0584aec856c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e5979160660396987502257fbb0f04771de4c0de21fe3d1be16deb1da3aacf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e89899cf4b5cbc8566cda6ebfc9a858c7bb0f743bc4ed17b63de28c47fcfa30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:04Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.918351 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee46e2ea-d8f4-4db4-bb8e-b970b6258216\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88a63d65bec16e87319bce4a3265538d83f89e35ce181fe8013917a17c4c88ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e66d6bfce75c340eebe5f45e62b451508a52e5b83caf85af2c318bb87e47d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e66d6bfce75c340eebe5f45e62b451508a52e5b83caf85af2c318bb87e47d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:04Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.951292 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://88ee6aece00ac25f2c4d6fdade45eee05973a8ee89637c69a6d4846ee54e10a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:30Z\\\",\\\"message\\\":\\\"election_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0318 20:08:30.254102 7035 services_controller.go:360] Finished syncing service networking-console-plugin on namespace openshift-network-console for network=default : 3.081202ms\\\\nI0318 20:08:30.251407 7035 handler.go:208] Removed *v1.Node event handler 2\\\\nI0318 20:08:30.254135 7035 services_controller.go:360] Finished syncing service cluster-autoscaler-operator on namespace openshift-machine-api for network=default : 3.284578ms\\\\nI0318 20:08:30.254263 7035 services_controller.go:356] Processing sync for service openshift-apiserver-operator/metrics for network=default\\\\nI0318 20:08:30.254332 7035 port_cache.go:96] port-cache(openshift-network-console_networking-console-plugin-85b44fc459-gdk6g): added port \\\\u0026{name:openshift-network-console_networking-console-plugin-85b44fc459-gdk6g uuid:c94130be-172c-477c-88c4-40cc7eba30fe logicalSwitch:crc ips:[0xc007ee7980] mac:[10 88 10 217 0 92] expires:{wall:0 ext:0 loc:\\\\u003cnil\\\\u003e}} with IP: [10.217.0.92/23] and MAC: 0a:58:0a:d9:00:5c\\\\nI0318 20:08:30.254353 7035 pods.go:252] [openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] addLogicalPort took \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:09:03Z\\\",\\\"message\\\":\\\"vices_controller.go:360] Finished syncing service openshift on namespace default for network=default : 8.321µs\\\\nI0318 20:09:03.564354 7354 services_controller.go:356] Processing sync for service openshift-etcd-operator/metrics for network=default\\\\nI0318 20:09:03.564389 7354 factory.go:656] Stopping watch factory\\\\nI0318 20:09:03.564510 7354 ovnkube_controller.go:1292] Config duration recorder: kind/namespace/name service/openshift-ovn-kubernetes/ovn-kubernetes-control-plane. OVN-Kubernetes controller took 0.141373592 seconds. No OVN measurement.\\\\nI0318 20:09:03.564578 7354 ovnkube.go:599] Stopped ovnkube\\\\nI0318 20:09:03.564142 7354 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nI0318 20:09:03.564693 7354 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0318 20:09:03.564704 7354 obj_retry.go:409] Going to retry *v1.Pod resource setup for 5 objects: [openshift-multus/network-metrics-daemon-l2cjn openshift-multus/multus-lqkxm openshift-dns/node-resolver-vl9h2 openshift-multus/multus-additional-cni-plugins-pwmzl openshift-network-console/networking-console-plugin-85b44fc459-gdk6g]\\\\nF0318 20:09:03.564802 7354 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:04Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.969059 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36099937-1da3-4e7b-abf2-171aef148364\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c38d8f163e6271ccc589aac12d588aabc2ed0963ec5d13976b3a56e94c64a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973ac70ac0e653b6a420f5bfb070f6304dc2d828022a97a5804ac3c050187888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4987ceacf9ceda23f6c9f9b3b79cc2984e0e088b89d99ba7c2b0560baf5f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea96d85a7524dbdb86733ec82ee52e98ad1de83fc26e1651d039637e489523f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea96d85a7524dbdb86733ec82ee52e98ad1de83fc26e1651d039637e489523f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:04Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.982352 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:04Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:04 crc kubenswrapper[4950]: I0318 20:09:04.999506 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b42f10f95ebda7760ce304ba8c5dfbf0e891059750a239f5ea8d0c96b673165\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:51Z\\\",\\\"message\\\":\\\"2026-03-18T20:08:05+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9214f383-12a6-4ea7-9936-fb82712c11ab\\\\n2026-03-18T20:08:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9214f383-12a6-4ea7-9936-fb82712c11ab to /host/opt/cni/bin/\\\\n2026-03-18T20:08:06Z [verbose] multus-daemon started\\\\n2026-03-18T20:08:06Z [verbose] Readiness Indicator file check\\\\n2026-03-18T20:08:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:04Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.479501 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.479634 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:05 crc kubenswrapper[4950]: E0318 20:09:05.479727 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.479917 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.479946 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:05 crc kubenswrapper[4950]: E0318 20:09:05.480005 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:05 crc kubenswrapper[4950]: E0318 20:09:05.480152 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:05 crc kubenswrapper[4950]: E0318 20:09:05.480300 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.538795 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p2zdp_83ee6659-cfde-4844-bb92-ccb9b26451ea/ovnkube-controller/3.log" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.542798 4950 scope.go:117] "RemoveContainer" containerID="544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4" Mar 18 20:09:05 crc kubenswrapper[4950]: E0318 20:09:05.543016 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p2zdp_openshift-ovn-kubernetes(83ee6659-cfde-4844-bb92-ccb9b26451ea)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.560331 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36099937-1da3-4e7b-abf2-171aef148364\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c38d8f163e6271ccc589aac12d588aabc2ed0963ec5d13976b3a56e94c64a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://973ac70ac0e653b6a420f5bfb070f6304dc2d828022a97a5804ac3c050187888\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c4987ceacf9ceda23f6c9f9b3b79cc2984e0e088b89d99ba7c2b0560baf5f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea96d85a7524dbdb86733ec82ee52e98ad1de83fc26e1651d039637e489523f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea96d85a7524dbdb86733ec82ee52e98ad1de83fc26e1651d039637e489523f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.578450 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.594277 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-lqkxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"981403b4-25e5-4ea0-9267-9584719853d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b42f10f95ebda7760ce304ba8c5dfbf0e891059750a239f5ea8d0c96b673165\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:08:51Z\\\",\\\"message\\\":\\\"2026-03-18T20:08:05+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9214f383-12a6-4ea7-9936-fb82712c11ab\\\\n2026-03-18T20:08:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9214f383-12a6-4ea7-9936-fb82712c11ab to /host/opt/cni/bin/\\\\n2026-03-18T20:08:06Z [verbose] multus-daemon started\\\\n2026-03-18T20:08:06Z [verbose] Readiness Indicator file check\\\\n2026-03-18T20:08:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hn5f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-lqkxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.618668 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83ee6659-cfde-4844-bb92-ccb9b26451ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-18T20:09:03Z\\\",\\\"message\\\":\\\"vices_controller.go:360] Finished syncing service openshift on namespace default for network=default : 8.321µs\\\\nI0318 20:09:03.564354 7354 services_controller.go:356] Processing sync for service openshift-etcd-operator/metrics for network=default\\\\nI0318 20:09:03.564389 7354 factory.go:656] Stopping watch factory\\\\nI0318 20:09:03.564510 7354 ovnkube_controller.go:1292] Config duration recorder: kind/namespace/name service/openshift-ovn-kubernetes/ovn-kubernetes-control-plane. OVN-Kubernetes controller took 0.141373592 seconds. No OVN measurement.\\\\nI0318 20:09:03.564578 7354 ovnkube.go:599] Stopped ovnkube\\\\nI0318 20:09:03.564142 7354 obj_retry.go:434] periodicallyRetryResources: Retry channel got triggered: retrying failed objects of type *v1.Pod\\\\nI0318 20:09:03.564693 7354 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0318 20:09:03.564704 7354 obj_retry.go:409] Going to retry *v1.Pod resource setup for 5 objects: [openshift-multus/network-metrics-daemon-l2cjn openshift-multus/multus-lqkxm openshift-dns/node-resolver-vl9h2 openshift-multus/multus-additional-cni-plugins-pwmzl openshift-network-console/networking-console-plugin-85b44fc459-gdk6g]\\\\nF0318 20:09:03.564802 7354 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:09:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p2zdp_openshift-ovn-kubernetes(83ee6659-cfde-4844-bb92-ccb9b26451ea)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l8qvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p2zdp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.637927 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f964e794e9fbff7a678c76cacb1a110810add97a36da111927a0123220886125\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f0d82e58ddf836f9f2078d8ed6a961bffdffb414b9fbeed9cb1e0513bb9864\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.657264 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.681366 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b13cdc2-ca84-4da5-9a60-5619987e6cef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52b6aa3dad160c747ebc18d26d2d504c4c72726d958a410ebe4ee008ca4741af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e4b7553a16fcecbbbf58213a0ec3aed66c3523aa6e5467276488d05c0241161c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b774cc2149c3d7ee9602eab046044327f1af515b51b2182a8b4772c5ca678c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f12401c3b8d27af65d74b71be8e78bdea8650c2aae9b75a54d40bab379e0e1b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f16caefd1eae55cc585ba175c9aea41852c960f2a1371fa857a78f58dea2330\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75684cd67d61468d5caf3f1b4d14597f45c273e25aef79a0747ee24d41975d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14dec570499ecb9eceb21c0c3653d78f4d81ff038386497fb001056b1d53174f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ndb6n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pwmzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.698624 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d09a228c-313f-4b55-a298-51ebd404cfd5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q76hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l2cjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.723612 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"831bf4ca-f84c-48d2-9054-410480eb986b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:07:35Z\\\",\\\"message\\\":\\\"le observer\\\\nW0318 20:07:35.272898 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0318 20:07:35.273030 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0318 20:07:35.273947 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2398488620/tls.crt::/tmp/serving-cert-2398488620/tls.key\\\\\\\"\\\\nI0318 20:07:35.779001 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0318 20:07:35.783097 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0318 20:07:35.783124 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0318 20:07:35.783155 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0318 20:07:35.783174 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0318 20:07:35.791487 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0318 20:07:35.791507 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791512 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0318 20:07:35.791517 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0318 20:07:35.791519 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0318 20:07:35.791522 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0318 20:07:35.791525 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0318 20:07:35.791560 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0318 20:07:35.793966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:07:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.745372 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.763961 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08bd350d08fb0a8d15c060bce29ee1dabe4de6289e09ed8ec358168297e57730\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.782739 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4048b439-3266-46e7-9de0-22377efacc46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://485c45c866dfdcaf52a61806c736923335433363d6c73262eba430083f1280d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tkc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k2n6r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.800649 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vl9h2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a08cee8-66e8-4cf0-b3ea-0c578a2438f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://582c87d2359948c1df7bfb9b78864c972cbe4574a5030faba78ed21cfe45b445\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z5pdc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vl9h2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.819705 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49a491f-1edc-4a40-b64f-147c0d87481f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1b01d79ecfba0d5984e7415ca4db790ef1635698da9792b6f1e6161a868f243\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://122df2a4bd1b73f3890ecaef0f5840c16b127b2b7a8b749ee84f2947334b6aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cpws6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5p989\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.840623 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8de376a3-1738-4cd5-8b36-2e4636215912\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ced76909ac5d1fd3962332517244a448367b89662e932e93daa0de34f70b4a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c09d46a4d315127ff42d2eb02cf657f6fbe2bdc23eb05addee4f8da9d736498\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-18T20:06:57Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0318 20:06:29.551493 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0318 20:06:29.554866 1 observer_polling.go:159] Starting file observer\\\\nI0318 20:06:29.586793 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0318 20:06:29.591515 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0318 20:06:57.967959 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0318 20:06:57.968124 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:06:57Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e662b624a83fcb0b0eab1048193fd00ce4a744f1a4653d776e0584aec856c49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e5979160660396987502257fbb0f04771de4c0de21fe3d1be16deb1da3aacf2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e89899cf4b5cbc8566cda6ebfc9a858c7bb0f743bc4ed17b63de28c47fcfa30\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.856822 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ee46e2ea-d8f4-4db4-bb8e-b970b6258216\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88a63d65bec16e87319bce4a3265538d83f89e35ce181fe8013917a17c4c88ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:06:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e66d6bfce75c340eebe5f45e62b451508a52e5b83caf85af2c318bb87e47d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e66d6bfce75c340eebe5f45e62b451508a52e5b83caf85af2c318bb87e47d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-18T20:06:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-18T20:06:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:06:27Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.879932 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca0ebd73fbd23dad6a37e05a799ab471d54ecd0ef67725713cae675b6be804ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:05 crc kubenswrapper[4950]: I0318 20:09:05.898094 4950 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7bhrr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22f8458a-e595-4fe8-98ac-b82d0c64d130\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-18T20:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c547cbcc08d6a427116fddb6cfc96e4358657c732c6ab76fabaa0ebf1b4526b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-18T20:08:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qpmlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-18T20:08:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7bhrr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-18T20:09:05Z is after 2025-08-24T17:21:41Z" Mar 18 20:09:07 crc kubenswrapper[4950]: I0318 20:09:07.441086 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:09:07 crc kubenswrapper[4950]: I0318 20:09:07.441235 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:07 crc kubenswrapper[4950]: I0318 20:09:07.441267 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:07 crc kubenswrapper[4950]: I0318 20:09:07.441305 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:07 crc kubenswrapper[4950]: I0318 20:09:07.441328 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:07 crc kubenswrapper[4950]: E0318 20:09:07.441490 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 20:09:07 crc kubenswrapper[4950]: E0318 20:09:07.441510 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 20:09:07 crc kubenswrapper[4950]: E0318 20:09:07.441523 4950 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:09:07 crc kubenswrapper[4950]: E0318 20:09:07.441580 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-18 20:10:11.441564126 +0000 UTC m=+224.682406014 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:09:07 crc kubenswrapper[4950]: E0318 20:09:07.441599 4950 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 20:09:07 crc kubenswrapper[4950]: E0318 20:09:07.441616 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 18 20:09:07 crc kubenswrapper[4950]: E0318 20:09:07.441671 4950 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 18 20:09:07 crc kubenswrapper[4950]: E0318 20:09:07.441691 4950 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:09:07 crc kubenswrapper[4950]: E0318 20:09:07.441711 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 20:10:11.441680228 +0000 UTC m=+224.682522126 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 18 20:09:07 crc kubenswrapper[4950]: E0318 20:09:07.441756 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-18 20:10:11.44173611 +0000 UTC m=+224.682578008 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 18 20:09:07 crc kubenswrapper[4950]: E0318 20:09:07.442110 4950 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 20:09:07 crc kubenswrapper[4950]: E0318 20:09:07.442163 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-18 20:10:11.44214766 +0000 UTC m=+224.682989568 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 18 20:09:07 crc kubenswrapper[4950]: E0318 20:09:07.442534 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:11.442521289 +0000 UTC m=+224.683363167 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:09:07 crc kubenswrapper[4950]: I0318 20:09:07.482185 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:07 crc kubenswrapper[4950]: E0318 20:09:07.482293 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:07 crc kubenswrapper[4950]: I0318 20:09:07.482561 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:07 crc kubenswrapper[4950]: E0318 20:09:07.482630 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:07 crc kubenswrapper[4950]: I0318 20:09:07.482750 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:07 crc kubenswrapper[4950]: E0318 20:09:07.482802 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:07 crc kubenswrapper[4950]: I0318 20:09:07.482912 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:07 crc kubenswrapper[4950]: E0318 20:09:07.482980 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:07 crc kubenswrapper[4950]: E0318 20:09:07.563772 4950 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 20:09:07 crc kubenswrapper[4950]: I0318 20:09:07.585582 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-7bhrr" podStartSLOduration=112.585562633 podStartE2EDuration="1m52.585562633s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:09:07.544475978 +0000 UTC m=+160.785317906" watchObservedRunningTime="2026-03-18 20:09:07.585562633 +0000 UTC m=+160.826404511" Mar 18 20:09:07 crc kubenswrapper[4950]: I0318 20:09:07.603813 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5p989" podStartSLOduration=112.603792543 podStartE2EDuration="1m52.603792543s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:09:07.585779108 +0000 UTC m=+160.826620986" watchObservedRunningTime="2026-03-18 20:09:07.603792543 +0000 UTC m=+160.844634421" Mar 18 20:09:07 crc kubenswrapper[4950]: I0318 20:09:07.604020 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=25.604015709 podStartE2EDuration="25.604015709s" podCreationTimestamp="2026-03-18 20:08:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:09:07.603084916 +0000 UTC m=+160.843926804" watchObservedRunningTime="2026-03-18 20:09:07.604015709 +0000 UTC m=+160.844857587" Mar 18 20:09:07 crc kubenswrapper[4950]: I0318 20:09:07.615840 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=8.615819609999999 podStartE2EDuration="8.61581961s" podCreationTimestamp="2026-03-18 20:08:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:09:07.615457422 +0000 UTC m=+160.856299320" watchObservedRunningTime="2026-03-18 20:09:07.61581961 +0000 UTC m=+160.856661488" Mar 18 20:09:07 crc kubenswrapper[4950]: I0318 20:09:07.633299 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-lqkxm" podStartSLOduration=112.633273272 podStartE2EDuration="1m52.633273272s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:09:07.633263281 +0000 UTC m=+160.874105169" watchObservedRunningTime="2026-03-18 20:09:07.633273272 +0000 UTC m=+160.874115150" Mar 18 20:09:07 crc kubenswrapper[4950]: I0318 20:09:07.693094 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=35.693072679 podStartE2EDuration="35.693072679s" podCreationTimestamp="2026-03-18 20:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:09:07.678979931 +0000 UTC m=+160.919821819" watchObservedRunningTime="2026-03-18 20:09:07.693072679 +0000 UTC m=+160.933914547" Mar 18 20:09:07 crc kubenswrapper[4950]: I0318 20:09:07.710153 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-pwmzl" podStartSLOduration=112.710138461 podStartE2EDuration="1m52.710138461s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:09:07.709242399 +0000 UTC m=+160.950084267" watchObservedRunningTime="2026-03-18 20:09:07.710138461 +0000 UTC m=+160.950980329" Mar 18 20:09:07 crc kubenswrapper[4950]: I0318 20:09:07.797215 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podStartSLOduration=112.797193061 podStartE2EDuration="1m52.797193061s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:09:07.79713783 +0000 UTC m=+161.037979698" watchObservedRunningTime="2026-03-18 20:09:07.797193061 +0000 UTC m=+161.038034929" Mar 18 20:09:07 crc kubenswrapper[4950]: I0318 20:09:07.816628 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-vl9h2" podStartSLOduration=112.816610621 podStartE2EDuration="1m52.816610621s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:09:07.816594381 +0000 UTC m=+161.057436259" watchObservedRunningTime="2026-03-18 20:09:07.816610621 +0000 UTC m=+161.057452489" Mar 18 20:09:07 crc kubenswrapper[4950]: I0318 20:09:07.846947 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=57.84693251 podStartE2EDuration="57.84693251s" podCreationTimestamp="2026-03-18 20:08:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:09:07.846087789 +0000 UTC m=+161.086929667" watchObservedRunningTime="2026-03-18 20:09:07.84693251 +0000 UTC m=+161.087774378" Mar 18 20:09:09 crc kubenswrapper[4950]: I0318 20:09:09.480838 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:09 crc kubenswrapper[4950]: I0318 20:09:09.480903 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:09 crc kubenswrapper[4950]: E0318 20:09:09.481805 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:09 crc kubenswrapper[4950]: I0318 20:09:09.480965 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:09 crc kubenswrapper[4950]: E0318 20:09:09.481928 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:09 crc kubenswrapper[4950]: I0318 20:09:09.480941 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:09 crc kubenswrapper[4950]: E0318 20:09:09.482028 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:09 crc kubenswrapper[4950]: E0318 20:09:09.482261 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.646731 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.646812 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.646834 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.646861 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.646879 4950 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-18T20:09:10Z","lastTransitionTime":"2026-03-18T20:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.713713 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-kktxv"] Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.714386 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kktxv" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.716764 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.719217 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.719662 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.721438 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.786030 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e0477c9f-05ed-4250-a387-55e3fdd58db5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-kktxv\" (UID: \"e0477c9f-05ed-4250-a387-55e3fdd58db5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kktxv" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.786094 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e0477c9f-05ed-4250-a387-55e3fdd58db5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-kktxv\" (UID: \"e0477c9f-05ed-4250-a387-55e3fdd58db5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kktxv" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.786134 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0477c9f-05ed-4250-a387-55e3fdd58db5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-kktxv\" (UID: \"e0477c9f-05ed-4250-a387-55e3fdd58db5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kktxv" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.786170 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0477c9f-05ed-4250-a387-55e3fdd58db5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-kktxv\" (UID: \"e0477c9f-05ed-4250-a387-55e3fdd58db5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kktxv" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.786245 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e0477c9f-05ed-4250-a387-55e3fdd58db5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-kktxv\" (UID: \"e0477c9f-05ed-4250-a387-55e3fdd58db5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kktxv" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.887290 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0477c9f-05ed-4250-a387-55e3fdd58db5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-kktxv\" (UID: \"e0477c9f-05ed-4250-a387-55e3fdd58db5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kktxv" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.887456 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0477c9f-05ed-4250-a387-55e3fdd58db5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-kktxv\" (UID: \"e0477c9f-05ed-4250-a387-55e3fdd58db5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kktxv" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.887546 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e0477c9f-05ed-4250-a387-55e3fdd58db5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-kktxv\" (UID: \"e0477c9f-05ed-4250-a387-55e3fdd58db5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kktxv" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.887630 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e0477c9f-05ed-4250-a387-55e3fdd58db5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-kktxv\" (UID: \"e0477c9f-05ed-4250-a387-55e3fdd58db5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kktxv" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.887701 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e0477c9f-05ed-4250-a387-55e3fdd58db5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-kktxv\" (UID: \"e0477c9f-05ed-4250-a387-55e3fdd58db5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kktxv" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.887816 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e0477c9f-05ed-4250-a387-55e3fdd58db5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-kktxv\" (UID: \"e0477c9f-05ed-4250-a387-55e3fdd58db5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kktxv" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.888673 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e0477c9f-05ed-4250-a387-55e3fdd58db5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-kktxv\" (UID: \"e0477c9f-05ed-4250-a387-55e3fdd58db5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kktxv" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.889624 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e0477c9f-05ed-4250-a387-55e3fdd58db5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-kktxv\" (UID: \"e0477c9f-05ed-4250-a387-55e3fdd58db5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kktxv" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.898863 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0477c9f-05ed-4250-a387-55e3fdd58db5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-kktxv\" (UID: \"e0477c9f-05ed-4250-a387-55e3fdd58db5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kktxv" Mar 18 20:09:10 crc kubenswrapper[4950]: I0318 20:09:10.918407 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0477c9f-05ed-4250-a387-55e3fdd58db5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-kktxv\" (UID: \"e0477c9f-05ed-4250-a387-55e3fdd58db5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kktxv" Mar 18 20:09:11 crc kubenswrapper[4950]: I0318 20:09:11.031181 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kktxv" Mar 18 20:09:11 crc kubenswrapper[4950]: W0318 20:09:11.054713 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0477c9f_05ed_4250_a387_55e3fdd58db5.slice/crio-b72b29d4993ff78564bafc2c1f2d8eb89b0836e7dac003077107d66cc3368827 WatchSource:0}: Error finding container b72b29d4993ff78564bafc2c1f2d8eb89b0836e7dac003077107d66cc3368827: Status 404 returned error can't find the container with id b72b29d4993ff78564bafc2c1f2d8eb89b0836e7dac003077107d66cc3368827 Mar 18 20:09:11 crc kubenswrapper[4950]: I0318 20:09:11.476395 4950 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 18 20:09:11 crc kubenswrapper[4950]: I0318 20:09:11.479561 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:11 crc kubenswrapper[4950]: I0318 20:09:11.479642 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:11 crc kubenswrapper[4950]: I0318 20:09:11.479569 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:11 crc kubenswrapper[4950]: E0318 20:09:11.479834 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:11 crc kubenswrapper[4950]: E0318 20:09:11.480016 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:11 crc kubenswrapper[4950]: E0318 20:09:11.480931 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:11 crc kubenswrapper[4950]: I0318 20:09:11.481034 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:11 crc kubenswrapper[4950]: E0318 20:09:11.481263 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:11 crc kubenswrapper[4950]: I0318 20:09:11.484345 4950 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 18 20:09:11 crc kubenswrapper[4950]: I0318 20:09:11.566925 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kktxv" event={"ID":"e0477c9f-05ed-4250-a387-55e3fdd58db5","Type":"ContainerStarted","Data":"cbcf309ed5d4024a64c88fa4526fdb977ef2f901aedea1920e645881e2d8a4a1"} Mar 18 20:09:11 crc kubenswrapper[4950]: I0318 20:09:11.567017 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kktxv" event={"ID":"e0477c9f-05ed-4250-a387-55e3fdd58db5","Type":"ContainerStarted","Data":"b72b29d4993ff78564bafc2c1f2d8eb89b0836e7dac003077107d66cc3368827"} Mar 18 20:09:12 crc kubenswrapper[4950]: E0318 20:09:12.565306 4950 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 20:09:13 crc kubenswrapper[4950]: I0318 20:09:13.478896 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:13 crc kubenswrapper[4950]: I0318 20:09:13.479029 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:13 crc kubenswrapper[4950]: I0318 20:09:13.479096 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:13 crc kubenswrapper[4950]: I0318 20:09:13.479113 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:13 crc kubenswrapper[4950]: E0318 20:09:13.479306 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:13 crc kubenswrapper[4950]: E0318 20:09:13.479514 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:13 crc kubenswrapper[4950]: E0318 20:09:13.479953 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:13 crc kubenswrapper[4950]: E0318 20:09:13.480077 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:14 crc kubenswrapper[4950]: I0318 20:09:14.128509 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs\") pod \"network-metrics-daemon-l2cjn\" (UID: \"d09a228c-313f-4b55-a298-51ebd404cfd5\") " pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:14 crc kubenswrapper[4950]: E0318 20:09:14.128735 4950 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 20:09:14 crc kubenswrapper[4950]: E0318 20:09:14.128888 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs podName:d09a228c-313f-4b55-a298-51ebd404cfd5 nodeName:}" failed. No retries permitted until 2026-03-18 20:10:18.128852534 +0000 UTC m=+231.369694472 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs") pod "network-metrics-daemon-l2cjn" (UID: "d09a228c-313f-4b55-a298-51ebd404cfd5") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 18 20:09:15 crc kubenswrapper[4950]: I0318 20:09:15.479377 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:15 crc kubenswrapper[4950]: E0318 20:09:15.479503 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:15 crc kubenswrapper[4950]: I0318 20:09:15.479376 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:15 crc kubenswrapper[4950]: I0318 20:09:15.479661 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:15 crc kubenswrapper[4950]: E0318 20:09:15.479710 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:15 crc kubenswrapper[4950]: E0318 20:09:15.479915 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:15 crc kubenswrapper[4950]: I0318 20:09:15.480617 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:15 crc kubenswrapper[4950]: E0318 20:09:15.480904 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:17 crc kubenswrapper[4950]: I0318 20:09:17.479710 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:17 crc kubenswrapper[4950]: I0318 20:09:17.479792 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:17 crc kubenswrapper[4950]: I0318 20:09:17.479823 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:17 crc kubenswrapper[4950]: E0318 20:09:17.481199 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:17 crc kubenswrapper[4950]: I0318 20:09:17.481244 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:17 crc kubenswrapper[4950]: E0318 20:09:17.481555 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:17 crc kubenswrapper[4950]: E0318 20:09:17.481700 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:17 crc kubenswrapper[4950]: E0318 20:09:17.482164 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:17 crc kubenswrapper[4950]: I0318 20:09:17.483099 4950 scope.go:117] "RemoveContainer" containerID="544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4" Mar 18 20:09:17 crc kubenswrapper[4950]: E0318 20:09:17.483576 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p2zdp_openshift-ovn-kubernetes(83ee6659-cfde-4844-bb92-ccb9b26451ea)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" Mar 18 20:09:17 crc kubenswrapper[4950]: E0318 20:09:17.566177 4950 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 20:09:18 crc kubenswrapper[4950]: I0318 20:09:18.501518 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kktxv" podStartSLOduration=123.501483509 podStartE2EDuration="2m3.501483509s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:09:11.587788508 +0000 UTC m=+164.828630446" watchObservedRunningTime="2026-03-18 20:09:18.501483509 +0000 UTC m=+171.742325457" Mar 18 20:09:18 crc kubenswrapper[4950]: I0318 20:09:18.502516 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 18 20:09:19 crc kubenswrapper[4950]: I0318 20:09:19.479715 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:19 crc kubenswrapper[4950]: E0318 20:09:19.479837 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:19 crc kubenswrapper[4950]: I0318 20:09:19.479942 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:19 crc kubenswrapper[4950]: I0318 20:09:19.480000 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:19 crc kubenswrapper[4950]: I0318 20:09:19.479950 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:19 crc kubenswrapper[4950]: E0318 20:09:19.480203 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:19 crc kubenswrapper[4950]: E0318 20:09:19.480296 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:19 crc kubenswrapper[4950]: E0318 20:09:19.480458 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:21 crc kubenswrapper[4950]: I0318 20:09:21.479265 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:21 crc kubenswrapper[4950]: I0318 20:09:21.479358 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:21 crc kubenswrapper[4950]: E0318 20:09:21.479450 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:21 crc kubenswrapper[4950]: I0318 20:09:21.479756 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:21 crc kubenswrapper[4950]: E0318 20:09:21.479742 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:21 crc kubenswrapper[4950]: E0318 20:09:21.479817 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:21 crc kubenswrapper[4950]: I0318 20:09:21.480452 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:21 crc kubenswrapper[4950]: E0318 20:09:21.480544 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:22 crc kubenswrapper[4950]: E0318 20:09:22.567716 4950 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 20:09:23 crc kubenswrapper[4950]: I0318 20:09:23.479570 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:23 crc kubenswrapper[4950]: I0318 20:09:23.479849 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:23 crc kubenswrapper[4950]: I0318 20:09:23.479706 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:23 crc kubenswrapper[4950]: I0318 20:09:23.479650 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:23 crc kubenswrapper[4950]: E0318 20:09:23.480021 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:23 crc kubenswrapper[4950]: E0318 20:09:23.480186 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:23 crc kubenswrapper[4950]: E0318 20:09:23.480340 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:23 crc kubenswrapper[4950]: E0318 20:09:23.480527 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:25 crc kubenswrapper[4950]: I0318 20:09:25.479118 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:25 crc kubenswrapper[4950]: I0318 20:09:25.479157 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:25 crc kubenswrapper[4950]: I0318 20:09:25.479198 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:25 crc kubenswrapper[4950]: I0318 20:09:25.479119 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:25 crc kubenswrapper[4950]: E0318 20:09:25.479292 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:25 crc kubenswrapper[4950]: E0318 20:09:25.479399 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:25 crc kubenswrapper[4950]: E0318 20:09:25.479577 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:25 crc kubenswrapper[4950]: E0318 20:09:25.479707 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:27 crc kubenswrapper[4950]: I0318 20:09:27.478713 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:27 crc kubenswrapper[4950]: I0318 20:09:27.478719 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:27 crc kubenswrapper[4950]: I0318 20:09:27.479784 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:27 crc kubenswrapper[4950]: I0318 20:09:27.479924 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:27 crc kubenswrapper[4950]: E0318 20:09:27.483844 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:27 crc kubenswrapper[4950]: E0318 20:09:27.484453 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:27 crc kubenswrapper[4950]: E0318 20:09:27.484770 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:27 crc kubenswrapper[4950]: E0318 20:09:27.484907 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:27 crc kubenswrapper[4950]: I0318 20:09:27.534728 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=9.534705355 podStartE2EDuration="9.534705355s" podCreationTimestamp="2026-03-18 20:09:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:09:27.534041739 +0000 UTC m=+180.774883667" watchObservedRunningTime="2026-03-18 20:09:27.534705355 +0000 UTC m=+180.775547263" Mar 18 20:09:27 crc kubenswrapper[4950]: E0318 20:09:27.568517 4950 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 20:09:29 crc kubenswrapper[4950]: I0318 20:09:29.479272 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:29 crc kubenswrapper[4950]: E0318 20:09:29.479874 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:29 crc kubenswrapper[4950]: I0318 20:09:29.479523 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:29 crc kubenswrapper[4950]: E0318 20:09:29.480002 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:29 crc kubenswrapper[4950]: I0318 20:09:29.479568 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:29 crc kubenswrapper[4950]: E0318 20:09:29.480130 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:29 crc kubenswrapper[4950]: I0318 20:09:29.479398 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:29 crc kubenswrapper[4950]: E0318 20:09:29.480232 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:31 crc kubenswrapper[4950]: I0318 20:09:31.479593 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:31 crc kubenswrapper[4950]: I0318 20:09:31.479640 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:31 crc kubenswrapper[4950]: I0318 20:09:31.479590 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:31 crc kubenswrapper[4950]: E0318 20:09:31.479728 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:31 crc kubenswrapper[4950]: E0318 20:09:31.479858 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:31 crc kubenswrapper[4950]: E0318 20:09:31.479971 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:31 crc kubenswrapper[4950]: I0318 20:09:31.480114 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:31 crc kubenswrapper[4950]: E0318 20:09:31.480201 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:32 crc kubenswrapper[4950]: I0318 20:09:32.480404 4950 scope.go:117] "RemoveContainer" containerID="544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4" Mar 18 20:09:32 crc kubenswrapper[4950]: E0318 20:09:32.480698 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p2zdp_openshift-ovn-kubernetes(83ee6659-cfde-4844-bb92-ccb9b26451ea)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" Mar 18 20:09:32 crc kubenswrapper[4950]: E0318 20:09:32.570030 4950 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 20:09:33 crc kubenswrapper[4950]: I0318 20:09:33.479284 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:33 crc kubenswrapper[4950]: I0318 20:09:33.479298 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:33 crc kubenswrapper[4950]: I0318 20:09:33.479382 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:33 crc kubenswrapper[4950]: I0318 20:09:33.480033 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:33 crc kubenswrapper[4950]: E0318 20:09:33.480214 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:33 crc kubenswrapper[4950]: E0318 20:09:33.480473 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:33 crc kubenswrapper[4950]: E0318 20:09:33.481495 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:33 crc kubenswrapper[4950]: E0318 20:09:33.481673 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:35 crc kubenswrapper[4950]: I0318 20:09:35.479609 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:35 crc kubenswrapper[4950]: I0318 20:09:35.479639 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:35 crc kubenswrapper[4950]: I0318 20:09:35.479609 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:35 crc kubenswrapper[4950]: I0318 20:09:35.479688 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:35 crc kubenswrapper[4950]: E0318 20:09:35.479786 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:35 crc kubenswrapper[4950]: E0318 20:09:35.479949 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:35 crc kubenswrapper[4950]: E0318 20:09:35.480057 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:35 crc kubenswrapper[4950]: E0318 20:09:35.480205 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:37 crc kubenswrapper[4950]: I0318 20:09:37.480725 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:37 crc kubenswrapper[4950]: I0318 20:09:37.480775 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:37 crc kubenswrapper[4950]: I0318 20:09:37.480796 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:37 crc kubenswrapper[4950]: E0318 20:09:37.487724 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:37 crc kubenswrapper[4950]: I0318 20:09:37.488364 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:37 crc kubenswrapper[4950]: E0318 20:09:37.488598 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:37 crc kubenswrapper[4950]: E0318 20:09:37.488883 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:37 crc kubenswrapper[4950]: E0318 20:09:37.489194 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:37 crc kubenswrapper[4950]: E0318 20:09:37.570707 4950 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 20:09:37 crc kubenswrapper[4950]: I0318 20:09:37.674698 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lqkxm_981403b4-25e5-4ea0-9267-9584719853d6/kube-multus/1.log" Mar 18 20:09:37 crc kubenswrapper[4950]: I0318 20:09:37.675450 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lqkxm_981403b4-25e5-4ea0-9267-9584719853d6/kube-multus/0.log" Mar 18 20:09:37 crc kubenswrapper[4950]: I0318 20:09:37.675506 4950 generic.go:334] "Generic (PLEG): container finished" podID="981403b4-25e5-4ea0-9267-9584719853d6" containerID="8b42f10f95ebda7760ce304ba8c5dfbf0e891059750a239f5ea8d0c96b673165" exitCode=1 Mar 18 20:09:37 crc kubenswrapper[4950]: I0318 20:09:37.675551 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lqkxm" event={"ID":"981403b4-25e5-4ea0-9267-9584719853d6","Type":"ContainerDied","Data":"8b42f10f95ebda7760ce304ba8c5dfbf0e891059750a239f5ea8d0c96b673165"} Mar 18 20:09:37 crc kubenswrapper[4950]: I0318 20:09:37.675599 4950 scope.go:117] "RemoveContainer" containerID="7a2cc1274fc4427a3f48b7c2628ec9b922abe931a4d13e0897761d35bf6057c3" Mar 18 20:09:37 crc kubenswrapper[4950]: I0318 20:09:37.676614 4950 scope.go:117] "RemoveContainer" containerID="8b42f10f95ebda7760ce304ba8c5dfbf0e891059750a239f5ea8d0c96b673165" Mar 18 20:09:37 crc kubenswrapper[4950]: E0318 20:09:37.677278 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-lqkxm_openshift-multus(981403b4-25e5-4ea0-9267-9584719853d6)\"" pod="openshift-multus/multus-lqkxm" podUID="981403b4-25e5-4ea0-9267-9584719853d6" Mar 18 20:09:38 crc kubenswrapper[4950]: I0318 20:09:38.682026 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lqkxm_981403b4-25e5-4ea0-9267-9584719853d6/kube-multus/1.log" Mar 18 20:09:39 crc kubenswrapper[4950]: I0318 20:09:39.478805 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:39 crc kubenswrapper[4950]: I0318 20:09:39.478872 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:39 crc kubenswrapper[4950]: I0318 20:09:39.478966 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:39 crc kubenswrapper[4950]: I0318 20:09:39.479043 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:39 crc kubenswrapper[4950]: E0318 20:09:39.479043 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:39 crc kubenswrapper[4950]: E0318 20:09:39.479206 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:39 crc kubenswrapper[4950]: E0318 20:09:39.479403 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:39 crc kubenswrapper[4950]: E0318 20:09:39.479600 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:41 crc kubenswrapper[4950]: I0318 20:09:41.479468 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:41 crc kubenswrapper[4950]: I0318 20:09:41.479512 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:41 crc kubenswrapper[4950]: E0318 20:09:41.479701 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:41 crc kubenswrapper[4950]: I0318 20:09:41.479770 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:41 crc kubenswrapper[4950]: I0318 20:09:41.479757 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:41 crc kubenswrapper[4950]: E0318 20:09:41.479922 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:41 crc kubenswrapper[4950]: E0318 20:09:41.480036 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:41 crc kubenswrapper[4950]: E0318 20:09:41.480084 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:42 crc kubenswrapper[4950]: E0318 20:09:42.572552 4950 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 20:09:43 crc kubenswrapper[4950]: I0318 20:09:43.479696 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:43 crc kubenswrapper[4950]: I0318 20:09:43.479825 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:43 crc kubenswrapper[4950]: I0318 20:09:43.479901 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:43 crc kubenswrapper[4950]: E0318 20:09:43.479906 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:43 crc kubenswrapper[4950]: E0318 20:09:43.480051 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:43 crc kubenswrapper[4950]: E0318 20:09:43.480205 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:43 crc kubenswrapper[4950]: I0318 20:09:43.480298 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:43 crc kubenswrapper[4950]: E0318 20:09:43.481525 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:44 crc kubenswrapper[4950]: I0318 20:09:44.480618 4950 scope.go:117] "RemoveContainer" containerID="544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4" Mar 18 20:09:44 crc kubenswrapper[4950]: I0318 20:09:44.709519 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p2zdp_83ee6659-cfde-4844-bb92-ccb9b26451ea/ovnkube-controller/3.log" Mar 18 20:09:44 crc kubenswrapper[4950]: I0318 20:09:44.719985 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerStarted","Data":"c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d"} Mar 18 20:09:44 crc kubenswrapper[4950]: I0318 20:09:44.720486 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:09:45 crc kubenswrapper[4950]: I0318 20:09:45.485335 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:45 crc kubenswrapper[4950]: I0318 20:09:45.485348 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:45 crc kubenswrapper[4950]: E0318 20:09:45.485940 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:45 crc kubenswrapper[4950]: I0318 20:09:45.485396 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:45 crc kubenswrapper[4950]: E0318 20:09:45.486030 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:45 crc kubenswrapper[4950]: I0318 20:09:45.485350 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:45 crc kubenswrapper[4950]: E0318 20:09:45.486077 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:45 crc kubenswrapper[4950]: E0318 20:09:45.486104 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:45 crc kubenswrapper[4950]: I0318 20:09:45.501596 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" podStartSLOduration=150.501567465 podStartE2EDuration="2m30.501567465s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:09:44.759203734 +0000 UTC m=+198.000045622" watchObservedRunningTime="2026-03-18 20:09:45.501567465 +0000 UTC m=+198.742409393" Mar 18 20:09:45 crc kubenswrapper[4950]: I0318 20:09:45.503222 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-l2cjn"] Mar 18 20:09:45 crc kubenswrapper[4950]: I0318 20:09:45.723153 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:45 crc kubenswrapper[4950]: E0318 20:09:45.723323 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:47 crc kubenswrapper[4950]: I0318 20:09:47.479257 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:47 crc kubenswrapper[4950]: E0318 20:09:47.481289 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:47 crc kubenswrapper[4950]: I0318 20:09:47.481349 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:47 crc kubenswrapper[4950]: I0318 20:09:47.481392 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:47 crc kubenswrapper[4950]: I0318 20:09:47.481557 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:47 crc kubenswrapper[4950]: E0318 20:09:47.481629 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:47 crc kubenswrapper[4950]: E0318 20:09:47.481836 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:47 crc kubenswrapper[4950]: E0318 20:09:47.481904 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:47 crc kubenswrapper[4950]: E0318 20:09:47.573381 4950 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 20:09:49 crc kubenswrapper[4950]: I0318 20:09:49.479284 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:49 crc kubenswrapper[4950]: I0318 20:09:49.479365 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:49 crc kubenswrapper[4950]: I0318 20:09:49.479462 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:49 crc kubenswrapper[4950]: E0318 20:09:49.479533 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:49 crc kubenswrapper[4950]: I0318 20:09:49.479578 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:49 crc kubenswrapper[4950]: E0318 20:09:49.479697 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:49 crc kubenswrapper[4950]: E0318 20:09:49.479834 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:49 crc kubenswrapper[4950]: E0318 20:09:49.479872 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:51 crc kubenswrapper[4950]: I0318 20:09:51.479822 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:51 crc kubenswrapper[4950]: I0318 20:09:51.479999 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:51 crc kubenswrapper[4950]: E0318 20:09:51.480463 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:51 crc kubenswrapper[4950]: I0318 20:09:51.480498 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:51 crc kubenswrapper[4950]: I0318 20:09:51.480500 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:51 crc kubenswrapper[4950]: E0318 20:09:51.480627 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:51 crc kubenswrapper[4950]: E0318 20:09:51.480754 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:51 crc kubenswrapper[4950]: E0318 20:09:51.480902 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:52 crc kubenswrapper[4950]: I0318 20:09:52.479213 4950 scope.go:117] "RemoveContainer" containerID="8b42f10f95ebda7760ce304ba8c5dfbf0e891059750a239f5ea8d0c96b673165" Mar 18 20:09:52 crc kubenswrapper[4950]: E0318 20:09:52.574442 4950 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 18 20:09:52 crc kubenswrapper[4950]: I0318 20:09:52.751347 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lqkxm_981403b4-25e5-4ea0-9267-9584719853d6/kube-multus/1.log" Mar 18 20:09:52 crc kubenswrapper[4950]: I0318 20:09:52.751455 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lqkxm" event={"ID":"981403b4-25e5-4ea0-9267-9584719853d6","Type":"ContainerStarted","Data":"25e6ec3b99e4fad5368ef138e2764c1c0dab606df101d76849a96e45014cf62c"} Mar 18 20:09:53 crc kubenswrapper[4950]: I0318 20:09:53.478905 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:53 crc kubenswrapper[4950]: I0318 20:09:53.478989 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:53 crc kubenswrapper[4950]: I0318 20:09:53.478930 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:53 crc kubenswrapper[4950]: E0318 20:09:53.479124 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:53 crc kubenswrapper[4950]: I0318 20:09:53.479186 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:53 crc kubenswrapper[4950]: E0318 20:09:53.479321 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:53 crc kubenswrapper[4950]: E0318 20:09:53.479524 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:53 crc kubenswrapper[4950]: E0318 20:09:53.479756 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:55 crc kubenswrapper[4950]: I0318 20:09:55.482386 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:55 crc kubenswrapper[4950]: I0318 20:09:55.482563 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:55 crc kubenswrapper[4950]: E0318 20:09:55.482750 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:55 crc kubenswrapper[4950]: I0318 20:09:55.482782 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:55 crc kubenswrapper[4950]: I0318 20:09:55.482808 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:55 crc kubenswrapper[4950]: E0318 20:09:55.482905 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:55 crc kubenswrapper[4950]: E0318 20:09:55.483034 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:55 crc kubenswrapper[4950]: E0318 20:09:55.483104 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:57 crc kubenswrapper[4950]: I0318 20:09:57.479298 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:57 crc kubenswrapper[4950]: I0318 20:09:57.479461 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:57 crc kubenswrapper[4950]: E0318 20:09:57.481208 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 18 20:09:57 crc kubenswrapper[4950]: I0318 20:09:57.481299 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:57 crc kubenswrapper[4950]: I0318 20:09:57.481245 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:57 crc kubenswrapper[4950]: E0318 20:09:57.481549 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 18 20:09:57 crc kubenswrapper[4950]: E0318 20:09:57.481732 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 18 20:09:57 crc kubenswrapper[4950]: E0318 20:09:57.481888 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l2cjn" podUID="d09a228c-313f-4b55-a298-51ebd404cfd5" Mar 18 20:09:59 crc kubenswrapper[4950]: I0318 20:09:59.479103 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:09:59 crc kubenswrapper[4950]: I0318 20:09:59.479181 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:09:59 crc kubenswrapper[4950]: I0318 20:09:59.479261 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:09:59 crc kubenswrapper[4950]: I0318 20:09:59.479293 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:09:59 crc kubenswrapper[4950]: I0318 20:09:59.483701 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 18 20:09:59 crc kubenswrapper[4950]: I0318 20:09:59.483785 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 18 20:09:59 crc kubenswrapper[4950]: I0318 20:09:59.484011 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 18 20:09:59 crc kubenswrapper[4950]: I0318 20:09:59.484079 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 18 20:09:59 crc kubenswrapper[4950]: I0318 20:09:59.484339 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 18 20:09:59 crc kubenswrapper[4950]: I0318 20:09:59.484341 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.207566 4950 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.260328 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-27jdq"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.261231 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.264008 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m7qjx"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.264562 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m7qjx" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.267668 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-kjb6z"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.268272 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-kjb6z" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.271698 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.272269 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.274287 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.274636 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.275005 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.275246 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.280881 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.281247 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.287781 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.288121 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.288589 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.289090 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.289397 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.289542 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.289707 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.289749 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.290005 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.290187 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.290646 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.290922 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-p9pk8"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.291723 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-s8cr9"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.293384 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p9pk8" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.293608 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-s8cr9" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.294701 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9l9mq"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.294995 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wv7kr"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.300003 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.305493 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.305729 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.307770 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.310078 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.314701 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.315486 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.320452 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.320484 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.320663 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.320677 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.321046 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.321177 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.321362 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.321525 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.321547 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.322747 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-kjb6z"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.322785 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m7qjx"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.323457 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.324301 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.324549 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.324741 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.326709 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.327519 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-27jdq"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.327653 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.327663 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.327795 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.327917 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.327931 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.328018 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.328328 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.328432 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.330244 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8jvmc"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.330754 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8jvmc" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.334390 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.340993 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.341287 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.341385 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.341496 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.341588 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.341639 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.341693 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.341718 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.341779 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.341814 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.341911 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.342001 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.342542 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.342644 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.342736 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.342751 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.344457 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.344562 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.345033 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.345129 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.345200 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d5rvd"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.345276 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.345359 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.345454 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.345521 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.345684 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-g2f49"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.346025 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-g2f49" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.346242 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d5rvd" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.350793 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.351459 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.367310 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9pgcg"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.367889 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9pgcg" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.368630 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.368755 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.368880 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.368978 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.369083 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.369167 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.369253 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.370349 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2mqb"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.370725 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2mqb" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.371192 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-46hjn"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.371443 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.371552 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.371598 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.371634 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.373370 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2kz74"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.373986 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.386710 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-47kcd"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.387280 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-47kcd" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.391622 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/792583b6-449b-4b55-b138-cf51971e2a17-serving-cert\") pod \"authentication-operator-69f744f599-s8cr9\" (UID: \"792583b6-449b-4b55-b138-cf51971e2a17\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8cr9" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.391647 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-encryption-config\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.391665 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/94495808-8600-45f6-a7fe-74ded4bd3ec4-client-ca\") pod \"route-controller-manager-6576b87f9c-wcd54\" (UID: \"94495808-8600-45f6-a7fe-74ded4bd3ec4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.391681 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5qm4\" (UniqueName: \"kubernetes.io/projected/a0e597f4-b43a-47db-9f22-a38272060004-kube-api-access-s5qm4\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.391696 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-etcd-client\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.391711 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ff73cd60-bdf9-477a-8ede-1e781aaffd22-image-import-ca\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.391735 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27a9fc13-1b12-40f1-9de4-fc04f3810df9-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-m7qjx\" (UID: \"27a9fc13-1b12-40f1-9de4-fc04f3810df9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m7qjx" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.391752 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxb2j\" (UniqueName: \"kubernetes.io/projected/792583b6-449b-4b55-b138-cf51971e2a17-kube-api-access-qxb2j\") pod \"authentication-operator-69f744f599-s8cr9\" (UID: \"792583b6-449b-4b55-b138-cf51971e2a17\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8cr9" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.391768 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs286\" (UniqueName: \"kubernetes.io/projected/dc9fd238-6c7a-46db-bb03-3782b9cfc481-kube-api-access-rs286\") pod \"cluster-image-registry-operator-dc59b4c8b-8jvmc\" (UID: \"dc9fd238-6c7a-46db-bb03-3782b9cfc481\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8jvmc" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.391782 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bfd29e9b-fbda-41d0-8300-a0d9337e0e6d-images\") pod \"machine-api-operator-5694c8668f-kjb6z\" (UID: \"bfd29e9b-fbda-41d0-8300-a0d9337e0e6d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-kjb6z" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.391799 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfd25023-4dd6-4527-bcde-836258333c37-config\") pod \"machine-approver-56656f9798-p9pk8\" (UID: \"cfd25023-4dd6-4527-bcde-836258333c37\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p9pk8" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.391812 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zggbj\" (UniqueName: \"kubernetes.io/projected/ff73cd60-bdf9-477a-8ede-1e781aaffd22-kube-api-access-zggbj\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.391827 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/792583b6-449b-4b55-b138-cf51971e2a17-service-ca-bundle\") pod \"authentication-operator-69f744f599-s8cr9\" (UID: \"792583b6-449b-4b55-b138-cf51971e2a17\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8cr9" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.391843 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/dc9fd238-6c7a-46db-bb03-3782b9cfc481-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8jvmc\" (UID: \"dc9fd238-6c7a-46db-bb03-3782b9cfc481\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8jvmc" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.391860 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/4684cfd6-1836-4218-9c6e-c6ca75a2da96-available-featuregates\") pod \"openshift-config-operator-7777fb866f-g2f49\" (UID: \"4684cfd6-1836-4218-9c6e-c6ca75a2da96\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-g2f49" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.391878 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.391906 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/70601ef2-ab17-4550-890b-b29bf1a7b85e-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9l9mq\" (UID: \"70601ef2-ab17-4550-890b-b29bf1a7b85e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.391924 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qb65\" (UniqueName: \"kubernetes.io/projected/27a9fc13-1b12-40f1-9de4-fc04f3810df9-kube-api-access-7qb65\") pod \"openshift-apiserver-operator-796bbdcf4f-m7qjx\" (UID: \"27a9fc13-1b12-40f1-9de4-fc04f3810df9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m7qjx" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.391938 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ff73cd60-bdf9-477a-8ede-1e781aaffd22-node-pullsecrets\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.391952 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/cc0b2145-4578-4709-837d-6bae20399714-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-d5rvd\" (UID: \"cc0b2145-4578-4709-837d-6bae20399714\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d5rvd" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.391968 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msbwr\" (UniqueName: \"kubernetes.io/projected/70601ef2-ab17-4550-890b-b29bf1a7b85e-kube-api-access-msbwr\") pod \"controller-manager-879f6c89f-9l9mq\" (UID: \"70601ef2-ab17-4550-890b-b29bf1a7b85e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.391983 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff73cd60-bdf9-477a-8ede-1e781aaffd22-config\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.391997 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ff73cd60-bdf9-477a-8ede-1e781aaffd22-encryption-config\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.392012 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-audit-policies\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.392027 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-serving-cert\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.392041 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntjmk\" (UniqueName: \"kubernetes.io/projected/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-kube-api-access-ntjmk\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.392056 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27a9fc13-1b12-40f1-9de4-fc04f3810df9-config\") pod \"openshift-apiserver-operator-796bbdcf4f-m7qjx\" (UID: \"27a9fc13-1b12-40f1-9de4-fc04f3810df9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m7qjx" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.392068 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfd29e9b-fbda-41d0-8300-a0d9337e0e6d-config\") pod \"machine-api-operator-5694c8668f-kjb6z\" (UID: \"bfd29e9b-fbda-41d0-8300-a0d9337e0e6d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-kjb6z" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.392083 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/792583b6-449b-4b55-b138-cf51971e2a17-config\") pod \"authentication-operator-69f744f599-s8cr9\" (UID: \"792583b6-449b-4b55-b138-cf51971e2a17\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8cr9" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.392098 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.392115 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.392131 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.392146 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.392159 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ff73cd60-bdf9-477a-8ede-1e781aaffd22-audit\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.392181 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.392425 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/bfd29e9b-fbda-41d0-8300-a0d9337e0e6d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-kjb6z\" (UID: \"bfd29e9b-fbda-41d0-8300-a0d9337e0e6d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-kjb6z" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.392560 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/70601ef2-ab17-4550-890b-b29bf1a7b85e-client-ca\") pod \"controller-manager-879f6c89f-9l9mq\" (UID: \"70601ef2-ab17-4550-890b-b29bf1a7b85e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.392635 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.392838 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.392897 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.392919 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg789\" (UniqueName: \"kubernetes.io/projected/bfd29e9b-fbda-41d0-8300-a0d9337e0e6d-kube-api-access-lg789\") pod \"machine-api-operator-5694c8668f-kjb6z\" (UID: \"bfd29e9b-fbda-41d0-8300-a0d9337e0e6d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-kjb6z" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.393078 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/cfd25023-4dd6-4527-bcde-836258333c37-machine-approver-tls\") pod \"machine-approver-56656f9798-p9pk8\" (UID: \"cfd25023-4dd6-4527-bcde-836258333c37\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p9pk8" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.393105 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-audit-dir\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.393119 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ff73cd60-bdf9-477a-8ede-1e781aaffd22-audit-dir\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.393161 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/792583b6-449b-4b55-b138-cf51971e2a17-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-s8cr9\" (UID: \"792583b6-449b-4b55-b138-cf51971e2a17\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8cr9" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.393188 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94495808-8600-45f6-a7fe-74ded4bd3ec4-config\") pod \"route-controller-manager-6576b87f9c-wcd54\" (UID: \"94495808-8600-45f6-a7fe-74ded4bd3ec4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.393205 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a0e597f4-b43a-47db-9f22-a38272060004-audit-dir\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.393244 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ff73cd60-bdf9-477a-8ede-1e781aaffd22-trusted-ca-bundle\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.393269 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dc9fd238-6c7a-46db-bb03-3782b9cfc481-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8jvmc\" (UID: \"dc9fd238-6c7a-46db-bb03-3782b9cfc481\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8jvmc" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.393339 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6dxl\" (UniqueName: \"kubernetes.io/projected/94495808-8600-45f6-a7fe-74ded4bd3ec4-kube-api-access-f6dxl\") pod \"route-controller-manager-6576b87f9c-wcd54\" (UID: \"94495808-8600-45f6-a7fe-74ded4bd3ec4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.393375 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70601ef2-ab17-4550-890b-b29bf1a7b85e-config\") pod \"controller-manager-879f6c89f-9l9mq\" (UID: \"70601ef2-ab17-4550-890b-b29bf1a7b85e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.393426 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.393468 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ff73cd60-bdf9-477a-8ede-1e781aaffd22-etcd-client\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.393610 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-pnffr"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.393607 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dc9fd238-6c7a-46db-bb03-3782b9cfc481-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8jvmc\" (UID: \"dc9fd238-6c7a-46db-bb03-3782b9cfc481\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8jvmc" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.394005 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cfd25023-4dd6-4527-bcde-836258333c37-auth-proxy-config\") pod \"machine-approver-56656f9798-p9pk8\" (UID: \"cfd25023-4dd6-4527-bcde-836258333c37\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p9pk8" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.394038 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70601ef2-ab17-4550-890b-b29bf1a7b85e-serving-cert\") pod \"controller-manager-879f6c89f-9l9mq\" (UID: \"70601ef2-ab17-4550-890b-b29bf1a7b85e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.394055 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ff73cd60-bdf9-477a-8ede-1e781aaffd22-etcd-serving-ca\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.394070 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff73cd60-bdf9-477a-8ede-1e781aaffd22-serving-cert\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.394091 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.394120 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4684cfd6-1836-4218-9c6e-c6ca75a2da96-serving-cert\") pod \"openshift-config-operator-7777fb866f-g2f49\" (UID: \"4684cfd6-1836-4218-9c6e-c6ca75a2da96\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-g2f49" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.394139 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.394154 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cv58x\" (UniqueName: \"kubernetes.io/projected/cc0b2145-4578-4709-837d-6bae20399714-kube-api-access-cv58x\") pod \"cluster-samples-operator-665b6dd947-d5rvd\" (UID: \"cc0b2145-4578-4709-837d-6bae20399714\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d5rvd" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.394173 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6fwj\" (UniqueName: \"kubernetes.io/projected/4684cfd6-1836-4218-9c6e-c6ca75a2da96-kube-api-access-t6fwj\") pod \"openshift-config-operator-7777fb866f-g2f49\" (UID: \"4684cfd6-1836-4218-9c6e-c6ca75a2da96\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-g2f49" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.394191 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.394206 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94495808-8600-45f6-a7fe-74ded4bd3ec4-serving-cert\") pod \"route-controller-manager-6576b87f9c-wcd54\" (UID: \"94495808-8600-45f6-a7fe-74ded4bd3ec4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.394221 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-audit-policies\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.394238 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws8tn\" (UniqueName: \"kubernetes.io/projected/cfd25023-4dd6-4527-bcde-836258333c37-kube-api-access-ws8tn\") pod \"machine-approver-56656f9798-p9pk8\" (UID: \"cfd25023-4dd6-4527-bcde-836258333c37\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p9pk8" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.394340 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-pnffr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.394594 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.394755 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.394916 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.395285 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.395366 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.395402 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.395491 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.395534 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.395598 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.395853 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.404519 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.404839 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.404988 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.405394 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.405579 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.406968 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.407208 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.407262 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.408592 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.412488 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-rf8k5"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.425314 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.430179 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.430605 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-rf8k5" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.431385 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.431996 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.433068 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.435386 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.436448 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.436625 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.436689 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qj7bk"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.437137 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-79fhl"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.437478 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-79fhl" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.437746 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qj7bk" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.448273 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.450906 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-twwmd"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.451568 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-twwmd" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.451871 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-gw45j"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.452250 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gw45j" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.462071 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-544q6"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.462147 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.462692 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-544q6" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.463218 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-pdnzp"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.463911 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pdnzp" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.465293 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-859lm"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.465720 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-859lm" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.466900 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.470475 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-gsj4z"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.471149 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-gsj4z" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.473524 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sxcm9"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.473997 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sxcm9" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.481077 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8nhwt"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.481704 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.495597 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-5wffk"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.496552 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-c7t4g"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.497151 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tfrtt"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.499862 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vnw2w"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.500396 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564408-2658d"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.500670 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c7t4g" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.500815 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qqfjh"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.500934 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tfrtt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.500950 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-5wffk" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.511767 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vnw2w" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.513125 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qqfjh" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.513863 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515046 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ff73cd60-bdf9-477a-8ede-1e781aaffd22-encryption-config\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515081 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ce6591e6-fb40-40a0-91a7-9aa2b3aa0437-etcd-client\") pod \"etcd-operator-b45778765-qj7bk\" (UID: \"ce6591e6-fb40-40a0-91a7-9aa2b3aa0437\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qj7bk" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515106 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-audit-policies\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515124 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-serving-cert\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515142 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1cec2292-f508-42b2-916f-9a5808045626-console-oauth-config\") pod \"console-f9d7485db-46hjn\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515168 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27a9fc13-1b12-40f1-9de4-fc04f3810df9-config\") pod \"openshift-apiserver-operator-796bbdcf4f-m7qjx\" (UID: \"27a9fc13-1b12-40f1-9de4-fc04f3810df9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m7qjx" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515186 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfd29e9b-fbda-41d0-8300-a0d9337e0e6d-config\") pod \"machine-api-operator-5694c8668f-kjb6z\" (UID: \"bfd29e9b-fbda-41d0-8300-a0d9337e0e6d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-kjb6z" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515204 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntjmk\" (UniqueName: \"kubernetes.io/projected/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-kube-api-access-ntjmk\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515219 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce6591e6-fb40-40a0-91a7-9aa2b3aa0437-config\") pod \"etcd-operator-b45778765-qj7bk\" (UID: \"ce6591e6-fb40-40a0-91a7-9aa2b3aa0437\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qj7bk" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515239 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/792583b6-449b-4b55-b138-cf51971e2a17-config\") pod \"authentication-operator-69f744f599-s8cr9\" (UID: \"792583b6-449b-4b55-b138-cf51971e2a17\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8cr9" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515258 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515280 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515301 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-console-config\") pod \"console-f9d7485db-46hjn\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515316 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515334 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515352 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ff73cd60-bdf9-477a-8ede-1e781aaffd22-audit\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515378 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515401 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/bfd29e9b-fbda-41d0-8300-a0d9337e0e6d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-kjb6z\" (UID: \"bfd29e9b-fbda-41d0-8300-a0d9337e0e6d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-kjb6z" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515436 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/70601ef2-ab17-4550-890b-b29bf1a7b85e-client-ca\") pod \"controller-manager-879f6c89f-9l9mq\" (UID: \"70601ef2-ab17-4550-890b-b29bf1a7b85e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515465 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515489 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515510 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515540 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg789\" (UniqueName: \"kubernetes.io/projected/bfd29e9b-fbda-41d0-8300-a0d9337e0e6d-kube-api-access-lg789\") pod \"machine-api-operator-5694c8668f-kjb6z\" (UID: \"bfd29e9b-fbda-41d0-8300-a0d9337e0e6d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-kjb6z" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515568 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/cfd25023-4dd6-4527-bcde-836258333c37-machine-approver-tls\") pod \"machine-approver-56656f9798-p9pk8\" (UID: \"cfd25023-4dd6-4527-bcde-836258333c37\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p9pk8" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515590 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-audit-dir\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515608 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ff73cd60-bdf9-477a-8ede-1e781aaffd22-audit-dir\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515626 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/792583b6-449b-4b55-b138-cf51971e2a17-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-s8cr9\" (UID: \"792583b6-449b-4b55-b138-cf51971e2a17\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8cr9" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515769 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94495808-8600-45f6-a7fe-74ded4bd3ec4-config\") pod \"route-controller-manager-6576b87f9c-wcd54\" (UID: \"94495808-8600-45f6-a7fe-74ded4bd3ec4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515819 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1cec2292-f508-42b2-916f-9a5808045626-console-serving-cert\") pod \"console-f9d7485db-46hjn\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515844 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a0e597f4-b43a-47db-9f22-a38272060004-audit-dir\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515863 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ff73cd60-bdf9-477a-8ede-1e781aaffd22-trusted-ca-bundle\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515886 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dc9fd238-6c7a-46db-bb03-3782b9cfc481-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8jvmc\" (UID: \"dc9fd238-6c7a-46db-bb03-3782b9cfc481\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8jvmc" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515904 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6dxl\" (UniqueName: \"kubernetes.io/projected/94495808-8600-45f6-a7fe-74ded4bd3ec4-kube-api-access-f6dxl\") pod \"route-controller-manager-6576b87f9c-wcd54\" (UID: \"94495808-8600-45f6-a7fe-74ded4bd3ec4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515927 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70601ef2-ab17-4550-890b-b29bf1a7b85e-config\") pod \"controller-manager-879f6c89f-9l9mq\" (UID: \"70601ef2-ab17-4550-890b-b29bf1a7b85e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515946 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515970 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ff73cd60-bdf9-477a-8ede-1e781aaffd22-etcd-client\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.515989 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dc9fd238-6c7a-46db-bb03-3782b9cfc481-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8jvmc\" (UID: \"dc9fd238-6c7a-46db-bb03-3782b9cfc481\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8jvmc" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516010 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-service-ca\") pod \"console-f9d7485db-46hjn\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516029 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cfd25023-4dd6-4527-bcde-836258333c37-auth-proxy-config\") pod \"machine-approver-56656f9798-p9pk8\" (UID: \"cfd25023-4dd6-4527-bcde-836258333c37\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p9pk8" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516048 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70601ef2-ab17-4550-890b-b29bf1a7b85e-serving-cert\") pod \"controller-manager-879f6c89f-9l9mq\" (UID: \"70601ef2-ab17-4550-890b-b29bf1a7b85e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516065 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ce6591e6-fb40-40a0-91a7-9aa2b3aa0437-etcd-service-ca\") pod \"etcd-operator-b45778765-qj7bk\" (UID: \"ce6591e6-fb40-40a0-91a7-9aa2b3aa0437\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qj7bk" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516087 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516110 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4684cfd6-1836-4218-9c6e-c6ca75a2da96-serving-cert\") pod \"openshift-config-operator-7777fb866f-g2f49\" (UID: \"4684cfd6-1836-4218-9c6e-c6ca75a2da96\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-g2f49" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516131 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ff73cd60-bdf9-477a-8ede-1e781aaffd22-etcd-serving-ca\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516151 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff73cd60-bdf9-477a-8ede-1e781aaffd22-serving-cert\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516170 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516192 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cv58x\" (UniqueName: \"kubernetes.io/projected/cc0b2145-4578-4709-837d-6bae20399714-kube-api-access-cv58x\") pod \"cluster-samples-operator-665b6dd947-d5rvd\" (UID: \"cc0b2145-4578-4709-837d-6bae20399714\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d5rvd" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516217 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516239 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6fwj\" (UniqueName: \"kubernetes.io/projected/4684cfd6-1836-4218-9c6e-c6ca75a2da96-kube-api-access-t6fwj\") pod \"openshift-config-operator-7777fb866f-g2f49\" (UID: \"4684cfd6-1836-4218-9c6e-c6ca75a2da96\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-g2f49" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516265 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws8tn\" (UniqueName: \"kubernetes.io/projected/cfd25023-4dd6-4527-bcde-836258333c37-kube-api-access-ws8tn\") pod \"machine-approver-56656f9798-p9pk8\" (UID: \"cfd25023-4dd6-4527-bcde-836258333c37\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p9pk8" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516288 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94495808-8600-45f6-a7fe-74ded4bd3ec4-serving-cert\") pod \"route-controller-manager-6576b87f9c-wcd54\" (UID: \"94495808-8600-45f6-a7fe-74ded4bd3ec4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516310 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-audit-policies\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516333 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ce6591e6-fb40-40a0-91a7-9aa2b3aa0437-etcd-ca\") pod \"etcd-operator-b45778765-qj7bk\" (UID: \"ce6591e6-fb40-40a0-91a7-9aa2b3aa0437\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qj7bk" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516355 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k2z5\" (UniqueName: \"kubernetes.io/projected/1cec2292-f508-42b2-916f-9a5808045626-kube-api-access-2k2z5\") pod \"console-f9d7485db-46hjn\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516374 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/792583b6-449b-4b55-b138-cf51971e2a17-serving-cert\") pod \"authentication-operator-69f744f599-s8cr9\" (UID: \"792583b6-449b-4b55-b138-cf51971e2a17\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8cr9" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516395 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-encryption-config\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516435 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7b8c\" (UniqueName: \"kubernetes.io/projected/ce6591e6-fb40-40a0-91a7-9aa2b3aa0437-kube-api-access-q7b8c\") pod \"etcd-operator-b45778765-qj7bk\" (UID: \"ce6591e6-fb40-40a0-91a7-9aa2b3aa0437\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qj7bk" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516471 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/94495808-8600-45f6-a7fe-74ded4bd3ec4-client-ca\") pod \"route-controller-manager-6576b87f9c-wcd54\" (UID: \"94495808-8600-45f6-a7fe-74ded4bd3ec4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516492 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5qm4\" (UniqueName: \"kubernetes.io/projected/a0e597f4-b43a-47db-9f22-a38272060004-kube-api-access-s5qm4\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516513 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-etcd-client\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516534 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ff73cd60-bdf9-477a-8ede-1e781aaffd22-image-import-ca\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516554 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxb2j\" (UniqueName: \"kubernetes.io/projected/792583b6-449b-4b55-b138-cf51971e2a17-kube-api-access-qxb2j\") pod \"authentication-operator-69f744f599-s8cr9\" (UID: \"792583b6-449b-4b55-b138-cf51971e2a17\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8cr9" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516576 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27a9fc13-1b12-40f1-9de4-fc04f3810df9-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-m7qjx\" (UID: \"27a9fc13-1b12-40f1-9de4-fc04f3810df9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m7qjx" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516598 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-trusted-ca-bundle\") pod \"console-f9d7485db-46hjn\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516622 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs286\" (UniqueName: \"kubernetes.io/projected/dc9fd238-6c7a-46db-bb03-3782b9cfc481-kube-api-access-rs286\") pod \"cluster-image-registry-operator-dc59b4c8b-8jvmc\" (UID: \"dc9fd238-6c7a-46db-bb03-3782b9cfc481\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8jvmc" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516641 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bfd29e9b-fbda-41d0-8300-a0d9337e0e6d-images\") pod \"machine-api-operator-5694c8668f-kjb6z\" (UID: \"bfd29e9b-fbda-41d0-8300-a0d9337e0e6d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-kjb6z" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516663 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfd25023-4dd6-4527-bcde-836258333c37-config\") pod \"machine-approver-56656f9798-p9pk8\" (UID: \"cfd25023-4dd6-4527-bcde-836258333c37\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p9pk8" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516686 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zggbj\" (UniqueName: \"kubernetes.io/projected/ff73cd60-bdf9-477a-8ede-1e781aaffd22-kube-api-access-zggbj\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516709 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd7wq\" (UniqueName: \"kubernetes.io/projected/290d1b89-d5c9-4928-8973-6c33f476b091-kube-api-access-pd7wq\") pod \"downloads-7954f5f757-rf8k5\" (UID: \"290d1b89-d5c9-4928-8973-6c33f476b091\") " pod="openshift-console/downloads-7954f5f757-rf8k5" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516730 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/792583b6-449b-4b55-b138-cf51971e2a17-service-ca-bundle\") pod \"authentication-operator-69f744f599-s8cr9\" (UID: \"792583b6-449b-4b55-b138-cf51971e2a17\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8cr9" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516752 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/dc9fd238-6c7a-46db-bb03-3782b9cfc481-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8jvmc\" (UID: \"dc9fd238-6c7a-46db-bb03-3782b9cfc481\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8jvmc" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516775 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/4684cfd6-1836-4218-9c6e-c6ca75a2da96-available-featuregates\") pod \"openshift-config-operator-7777fb866f-g2f49\" (UID: \"4684cfd6-1836-4218-9c6e-c6ca75a2da96\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-g2f49" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516798 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516818 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/70601ef2-ab17-4550-890b-b29bf1a7b85e-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9l9mq\" (UID: \"70601ef2-ab17-4550-890b-b29bf1a7b85e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516834 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce6591e6-fb40-40a0-91a7-9aa2b3aa0437-serving-cert\") pod \"etcd-operator-b45778765-qj7bk\" (UID: \"ce6591e6-fb40-40a0-91a7-9aa2b3aa0437\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qj7bk" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516856 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qb65\" (UniqueName: \"kubernetes.io/projected/27a9fc13-1b12-40f1-9de4-fc04f3810df9-kube-api-access-7qb65\") pod \"openshift-apiserver-operator-796bbdcf4f-m7qjx\" (UID: \"27a9fc13-1b12-40f1-9de4-fc04f3810df9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m7qjx" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516878 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/cc0b2145-4578-4709-837d-6bae20399714-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-d5rvd\" (UID: \"cc0b2145-4578-4709-837d-6bae20399714\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d5rvd" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516898 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msbwr\" (UniqueName: \"kubernetes.io/projected/70601ef2-ab17-4550-890b-b29bf1a7b85e-kube-api-access-msbwr\") pod \"controller-manager-879f6c89f-9l9mq\" (UID: \"70601ef2-ab17-4550-890b-b29bf1a7b85e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516916 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ff73cd60-bdf9-477a-8ede-1e781aaffd22-node-pullsecrets\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516935 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-oauth-serving-cert\") pod \"console-f9d7485db-46hjn\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.516956 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff73cd60-bdf9-477a-8ede-1e781aaffd22-config\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.517698 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff73cd60-bdf9-477a-8ede-1e781aaffd22-config\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.517738 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ff73cd60-bdf9-477a-8ede-1e781aaffd22-trusted-ca-bundle\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.518638 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564408-2658d" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.518886 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/70601ef2-ab17-4550-890b-b29bf1a7b85e-client-ca\") pod \"controller-manager-879f6c89f-9l9mq\" (UID: \"70601ef2-ab17-4550-890b-b29bf1a7b85e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.520424 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-audit-policies\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.521119 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/94495808-8600-45f6-a7fe-74ded4bd3ec4-client-ca\") pod \"route-controller-manager-6576b87f9c-wcd54\" (UID: \"94495808-8600-45f6-a7fe-74ded4bd3ec4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.536661 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dc9fd238-6c7a-46db-bb03-3782b9cfc481-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8jvmc\" (UID: \"dc9fd238-6c7a-46db-bb03-3782b9cfc481\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8jvmc" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.537712 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cfd25023-4dd6-4527-bcde-836258333c37-auth-proxy-config\") pod \"machine-approver-56656f9798-p9pk8\" (UID: \"cfd25023-4dd6-4527-bcde-836258333c37\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p9pk8" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.539207 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ff73cd60-bdf9-477a-8ede-1e781aaffd22-etcd-serving-ca\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.539631 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70601ef2-ab17-4550-890b-b29bf1a7b85e-config\") pod \"controller-manager-879f6c89f-9l9mq\" (UID: \"70601ef2-ab17-4550-890b-b29bf1a7b85e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.540963 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfd29e9b-fbda-41d0-8300-a0d9337e0e6d-config\") pod \"machine-api-operator-5694c8668f-kjb6z\" (UID: \"bfd29e9b-fbda-41d0-8300-a0d9337e0e6d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-kjb6z" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.541234 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.542456 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.542817 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-serving-cert\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.543321 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27a9fc13-1b12-40f1-9de4-fc04f3810df9-config\") pod \"openshift-apiserver-operator-796bbdcf4f-m7qjx\" (UID: \"27a9fc13-1b12-40f1-9de4-fc04f3810df9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m7qjx" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.544825 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-audit-dir\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.544899 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ff73cd60-bdf9-477a-8ede-1e781aaffd22-audit-dir\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.545855 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94495808-8600-45f6-a7fe-74ded4bd3ec4-config\") pod \"route-controller-manager-6576b87f9c-wcd54\" (UID: \"94495808-8600-45f6-a7fe-74ded4bd3ec4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.545922 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a0e597f4-b43a-47db-9f22-a38272060004-audit-dir\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.548300 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xj696"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.548554 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.549329 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.549749 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff73cd60-bdf9-477a-8ede-1e781aaffd22-serving-cert\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.551665 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.553942 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ff73cd60-bdf9-477a-8ede-1e781aaffd22-encryption-config\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.554226 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/4684cfd6-1836-4218-9c6e-c6ca75a2da96-available-featuregates\") pod \"openshift-config-operator-7777fb866f-g2f49\" (UID: \"4684cfd6-1836-4218-9c6e-c6ca75a2da96\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-g2f49" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.554759 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bfd29e9b-fbda-41d0-8300-a0d9337e0e6d-images\") pod \"machine-api-operator-5694c8668f-kjb6z\" (UID: \"bfd29e9b-fbda-41d0-8300-a0d9337e0e6d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-kjb6z" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.554966 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.555220 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ff73cd60-bdf9-477a-8ede-1e781aaffd22-image-import-ca\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.555251 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xj696" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.555977 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/792583b6-449b-4b55-b138-cf51971e2a17-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-s8cr9\" (UID: \"792583b6-449b-4b55-b138-cf51971e2a17\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8cr9" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.556045 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.556179 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/792583b6-449b-4b55-b138-cf51971e2a17-service-ca-bundle\") pod \"authentication-operator-69f744f599-s8cr9\" (UID: \"792583b6-449b-4b55-b138-cf51971e2a17\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8cr9" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.556254 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4684cfd6-1836-4218-9c6e-c6ca75a2da96-serving-cert\") pod \"openshift-config-operator-7777fb866f-g2f49\" (UID: \"4684cfd6-1836-4218-9c6e-c6ca75a2da96\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-g2f49" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.556257 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfd25023-4dd6-4527-bcde-836258333c37-config\") pod \"machine-approver-56656f9798-p9pk8\" (UID: \"cfd25023-4dd6-4527-bcde-836258333c37\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p9pk8" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.556480 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/792583b6-449b-4b55-b138-cf51971e2a17-config\") pod \"authentication-operator-69f744f599-s8cr9\" (UID: \"792583b6-449b-4b55-b138-cf51971e2a17\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8cr9" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.556530 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-audit-policies\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.557024 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-etcd-client\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.557267 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/70601ef2-ab17-4550-890b-b29bf1a7b85e-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9l9mq\" (UID: \"70601ef2-ab17-4550-890b-b29bf1a7b85e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.557711 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ff73cd60-bdf9-477a-8ede-1e781aaffd22-node-pullsecrets\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.557764 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.558079 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564410-9hqrd"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.558663 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.559046 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.560698 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70601ef2-ab17-4550-890b-b29bf1a7b85e-serving-cert\") pod \"controller-manager-879f6c89f-9l9mq\" (UID: \"70601ef2-ab17-4550-890b-b29bf1a7b85e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.561073 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.561100 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/bfd29e9b-fbda-41d0-8300-a0d9337e0e6d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-kjb6z\" (UID: \"bfd29e9b-fbda-41d0-8300-a0d9337e0e6d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-kjb6z" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.561124 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564410-9hqrd" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.561627 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94495808-8600-45f6-a7fe-74ded4bd3ec4-serving-cert\") pod \"route-controller-manager-6576b87f9c-wcd54\" (UID: \"94495808-8600-45f6-a7fe-74ded4bd3ec4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.561807 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ff73cd60-bdf9-477a-8ede-1e781aaffd22-audit\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.561867 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.565870 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-encryption-config\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.567461 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.567481 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.570727 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27a9fc13-1b12-40f1-9de4-fc04f3810df9-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-m7qjx\" (UID: \"27a9fc13-1b12-40f1-9de4-fc04f3810df9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m7qjx" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.571479 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.573636 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564400-d9vqv"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.574036 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/cc0b2145-4578-4709-837d-6bae20399714-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-d5rvd\" (UID: \"cc0b2145-4578-4709-837d-6bae20399714\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d5rvd" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.574224 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-ktwzg"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.574615 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-ktwzg" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.574716 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564400-d9vqv" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.574747 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/cfd25023-4dd6-4527-bcde-836258333c37-machine-approver-tls\") pod \"machine-approver-56656f9798-p9pk8\" (UID: \"cfd25023-4dd6-4527-bcde-836258333c37\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p9pk8" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.574609 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ff73cd60-bdf9-477a-8ede-1e781aaffd22-etcd-client\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.574788 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-kbjxd"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.575609 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-kbjxd" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.575769 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.576733 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.576879 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-s8cr9"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.578771 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/dc9fd238-6c7a-46db-bb03-3782b9cfc481-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8jvmc\" (UID: \"dc9fd238-6c7a-46db-bb03-3782b9cfc481\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8jvmc" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.580348 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.583134 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.587896 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/792583b6-449b-4b55-b138-cf51971e2a17-serving-cert\") pod \"authentication-operator-69f744f599-s8cr9\" (UID: \"792583b6-449b-4b55-b138-cf51971e2a17\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8cr9" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.590186 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.590250 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vnw2w"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.590402 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9pgcg"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.590438 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-544q6"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.590451 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-pdnzp"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.590462 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-47kcd"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.591703 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-pnffr"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.592593 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-46hjn"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.594585 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wv7kr"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.594631 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d5rvd"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.597228 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-859lm"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.597254 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-c7t4g"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.598629 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8jvmc"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.598988 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-g2f49"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.600101 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9l9mq"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.601109 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-rwl8z"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.603345 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-rwl8z" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.604891 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.612873 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-qqrpb"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.614517 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.617450 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-service-ca\") pod \"console-f9d7485db-46hjn\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.617477 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ce6591e6-fb40-40a0-91a7-9aa2b3aa0437-etcd-service-ca\") pod \"etcd-operator-b45778765-qj7bk\" (UID: \"ce6591e6-fb40-40a0-91a7-9aa2b3aa0437\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qj7bk" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.617508 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k2z5\" (UniqueName: \"kubernetes.io/projected/1cec2292-f508-42b2-916f-9a5808045626-kube-api-access-2k2z5\") pod \"console-f9d7485db-46hjn\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.617530 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ce6591e6-fb40-40a0-91a7-9aa2b3aa0437-etcd-ca\") pod \"etcd-operator-b45778765-qj7bk\" (UID: \"ce6591e6-fb40-40a0-91a7-9aa2b3aa0437\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qj7bk" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.617546 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7b8c\" (UniqueName: \"kubernetes.io/projected/ce6591e6-fb40-40a0-91a7-9aa2b3aa0437-kube-api-access-q7b8c\") pod \"etcd-operator-b45778765-qj7bk\" (UID: \"ce6591e6-fb40-40a0-91a7-9aa2b3aa0437\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qj7bk" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.617579 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-trusted-ca-bundle\") pod \"console-f9d7485db-46hjn\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.617605 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd7wq\" (UniqueName: \"kubernetes.io/projected/290d1b89-d5c9-4928-8973-6c33f476b091-kube-api-access-pd7wq\") pod \"downloads-7954f5f757-rf8k5\" (UID: \"290d1b89-d5c9-4928-8973-6c33f476b091\") " pod="openshift-console/downloads-7954f5f757-rf8k5" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.617624 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce6591e6-fb40-40a0-91a7-9aa2b3aa0437-serving-cert\") pod \"etcd-operator-b45778765-qj7bk\" (UID: \"ce6591e6-fb40-40a0-91a7-9aa2b3aa0437\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qj7bk" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.617651 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-oauth-serving-cert\") pod \"console-f9d7485db-46hjn\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.617668 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ce6591e6-fb40-40a0-91a7-9aa2b3aa0437-etcd-client\") pod \"etcd-operator-b45778765-qj7bk\" (UID: \"ce6591e6-fb40-40a0-91a7-9aa2b3aa0437\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qj7bk" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.617684 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1cec2292-f508-42b2-916f-9a5808045626-console-oauth-config\") pod \"console-f9d7485db-46hjn\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.617706 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce6591e6-fb40-40a0-91a7-9aa2b3aa0437-config\") pod \"etcd-operator-b45778765-qj7bk\" (UID: \"ce6591e6-fb40-40a0-91a7-9aa2b3aa0437\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qj7bk" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.617723 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-console-config\") pod \"console-f9d7485db-46hjn\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.617773 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1cec2292-f508-42b2-916f-9a5808045626-console-serving-cert\") pod \"console-f9d7485db-46hjn\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.618454 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564408-2658d"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.618490 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-79fhl"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.618502 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-twwmd"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.618512 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tfrtt"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.618521 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-rf8k5"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.618532 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.618542 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8nhwt"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.618553 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-5wffk"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.618564 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2mqb"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.618577 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-gw45j"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.618586 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sxcm9"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.618812 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-trusted-ca-bundle\") pod \"console-f9d7485db-46hjn\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.618962 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2kz74"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.619368 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-oauth-serving-cert\") pod \"console-f9d7485db-46hjn\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.619893 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-console-config\") pod \"console-f9d7485db-46hjn\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.620086 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-service-ca\") pod \"console-f9d7485db-46hjn\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.621096 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-qqrpb"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.622136 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qj7bk"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.622521 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1cec2292-f508-42b2-916f-9a5808045626-console-serving-cert\") pod \"console-f9d7485db-46hjn\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.623474 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-kbjxd"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.623732 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1cec2292-f508-42b2-916f-9a5808045626-console-oauth-config\") pod \"console-f9d7485db-46hjn\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.624062 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.624932 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-ktwzg"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.626774 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xj696"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.628635 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564410-9hqrd"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.629804 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.630812 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564400-d9vqv"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.631795 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-cdpgz"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.633113 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-cdpgz" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.633231 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qqfjh"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.634775 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-cdpgz"] Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.643138 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.662993 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.682989 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.705199 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.723693 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.743472 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.752650 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce6591e6-fb40-40a0-91a7-9aa2b3aa0437-serving-cert\") pod \"etcd-operator-b45778765-qj7bk\" (UID: \"ce6591e6-fb40-40a0-91a7-9aa2b3aa0437\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qj7bk" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.763755 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.768917 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce6591e6-fb40-40a0-91a7-9aa2b3aa0437-config\") pod \"etcd-operator-b45778765-qj7bk\" (UID: \"ce6591e6-fb40-40a0-91a7-9aa2b3aa0437\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qj7bk" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.783612 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.788425 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ce6591e6-fb40-40a0-91a7-9aa2b3aa0437-etcd-ca\") pod \"etcd-operator-b45778765-qj7bk\" (UID: \"ce6591e6-fb40-40a0-91a7-9aa2b3aa0437\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qj7bk" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.803485 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.809641 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ce6591e6-fb40-40a0-91a7-9aa2b3aa0437-etcd-service-ca\") pod \"etcd-operator-b45778765-qj7bk\" (UID: \"ce6591e6-fb40-40a0-91a7-9aa2b3aa0437\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qj7bk" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.823933 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.844397 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.851053 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ce6591e6-fb40-40a0-91a7-9aa2b3aa0437-etcd-client\") pod \"etcd-operator-b45778765-qj7bk\" (UID: \"ce6591e6-fb40-40a0-91a7-9aa2b3aa0437\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qj7bk" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.885262 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.904491 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.923181 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.954139 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.966013 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 18 20:10:02 crc kubenswrapper[4950]: I0318 20:10:02.983993 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.004481 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.024143 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.044770 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.064334 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.087522 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.105745 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.124302 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.144351 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.164789 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.184155 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.204376 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.224010 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.244724 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.283727 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.304209 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.323723 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.343454 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.362927 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.384530 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.405111 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.424702 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.443746 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.463633 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.483689 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.502377 4950 request.go:700] Waited for 1.020405359s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/secrets?fieldSelector=metadata.name%3Dmarketplace-operator-metrics&limit=500&resourceVersion=0 Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.504334 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.535622 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.544357 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.564952 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.588151 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.604870 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.624574 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.643070 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.663777 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.684084 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.704890 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.724929 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.744396 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.763489 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.784737 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.822182 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dc9fd238-6c7a-46db-bb03-3782b9cfc481-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8jvmc\" (UID: \"dc9fd238-6c7a-46db-bb03-3782b9cfc481\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8jvmc" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.853934 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6dxl\" (UniqueName: \"kubernetes.io/projected/94495808-8600-45f6-a7fe-74ded4bd3ec4-kube-api-access-f6dxl\") pod \"route-controller-manager-6576b87f9c-wcd54\" (UID: \"94495808-8600-45f6-a7fe-74ded4bd3ec4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.863631 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.872370 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntjmk\" (UniqueName: \"kubernetes.io/projected/be93cdca-9fe6-46a2-ae6a-94516d8eaa50-kube-api-access-ntjmk\") pod \"apiserver-7bbb656c7d-tnbzn\" (UID: \"be93cdca-9fe6-46a2-ae6a-94516d8eaa50\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.876048 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.907948 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5qm4\" (UniqueName: \"kubernetes.io/projected/a0e597f4-b43a-47db-9f22-a38272060004-kube-api-access-s5qm4\") pod \"oauth-openshift-558db77b4-wv7kr\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.925555 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.926821 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cv58x\" (UniqueName: \"kubernetes.io/projected/cc0b2145-4578-4709-837d-6bae20399714-kube-api-access-cv58x\") pod \"cluster-samples-operator-665b6dd947-d5rvd\" (UID: \"cc0b2145-4578-4709-837d-6bae20399714\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d5rvd" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.951989 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d5rvd" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.953222 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.972917 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg789\" (UniqueName: \"kubernetes.io/projected/bfd29e9b-fbda-41d0-8300-a0d9337e0e6d-kube-api-access-lg789\") pod \"machine-api-operator-5694c8668f-kjb6z\" (UID: \"bfd29e9b-fbda-41d0-8300-a0d9337e0e6d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-kjb6z" Mar 18 20:10:03 crc kubenswrapper[4950]: I0318 20:10:03.992294 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6fwj\" (UniqueName: \"kubernetes.io/projected/4684cfd6-1836-4218-9c6e-c6ca75a2da96-kube-api-access-t6fwj\") pod \"openshift-config-operator-7777fb866f-g2f49\" (UID: \"4684cfd6-1836-4218-9c6e-c6ca75a2da96\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-g2f49" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.003367 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws8tn\" (UniqueName: \"kubernetes.io/projected/cfd25023-4dd6-4527-bcde-836258333c37-kube-api-access-ws8tn\") pod \"machine-approver-56656f9798-p9pk8\" (UID: \"cfd25023-4dd6-4527-bcde-836258333c37\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p9pk8" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.027153 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxb2j\" (UniqueName: \"kubernetes.io/projected/792583b6-449b-4b55-b138-cf51971e2a17-kube-api-access-qxb2j\") pod \"authentication-operator-69f744f599-s8cr9\" (UID: \"792583b6-449b-4b55-b138-cf51971e2a17\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s8cr9" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.040819 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs286\" (UniqueName: \"kubernetes.io/projected/dc9fd238-6c7a-46db-bb03-3782b9cfc481-kube-api-access-rs286\") pod \"cluster-image-registry-operator-dc59b4c8b-8jvmc\" (UID: \"dc9fd238-6c7a-46db-bb03-3782b9cfc481\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8jvmc" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.060802 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zggbj\" (UniqueName: \"kubernetes.io/projected/ff73cd60-bdf9-477a-8ede-1e781aaffd22-kube-api-access-zggbj\") pod \"apiserver-76f77b778f-27jdq\" (UID: \"ff73cd60-bdf9-477a-8ede-1e781aaffd22\") " pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.076393 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54"] Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.078278 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qb65\" (UniqueName: \"kubernetes.io/projected/27a9fc13-1b12-40f1-9de4-fc04f3810df9-kube-api-access-7qb65\") pod \"openshift-apiserver-operator-796bbdcf4f-m7qjx\" (UID: \"27a9fc13-1b12-40f1-9de4-fc04f3810df9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m7qjx" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.084942 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.093915 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.103685 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.124548 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.124669 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m7qjx" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.137951 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-kjb6z" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.143122 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.143630 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.155180 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p9pk8" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.163512 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.163815 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-s8cr9" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.182997 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d5rvd"] Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.191817 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.198274 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msbwr\" (UniqueName: \"kubernetes.io/projected/70601ef2-ab17-4550-890b-b29bf1a7b85e-kube-api-access-msbwr\") pod \"controller-manager-879f6c89f-9l9mq\" (UID: \"70601ef2-ab17-4550-890b-b29bf1a7b85e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.204038 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.205106 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.219170 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8jvmc" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.223933 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.245707 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-g2f49" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.246603 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.270015 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.284050 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.295185 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-27jdq"] Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.304284 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.324747 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.345626 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.366816 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.385339 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.406782 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.429595 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.451697 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.463697 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.484521 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.488749 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn"] Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.505205 4950 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.522620 4950 request.go:700] Waited for 1.907848799s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.524551 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.544176 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.591146 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd7wq\" (UniqueName: \"kubernetes.io/projected/290d1b89-d5c9-4928-8973-6c33f476b091-kube-api-access-pd7wq\") pod \"downloads-7954f5f757-rf8k5\" (UID: \"290d1b89-d5c9-4928-8973-6c33f476b091\") " pod="openshift-console/downloads-7954f5f757-rf8k5" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.597172 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7b8c\" (UniqueName: \"kubernetes.io/projected/ce6591e6-fb40-40a0-91a7-9aa2b3aa0437-kube-api-access-q7b8c\") pod \"etcd-operator-b45778765-qj7bk\" (UID: \"ce6591e6-fb40-40a0-91a7-9aa2b3aa0437\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qj7bk" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.617319 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k2z5\" (UniqueName: \"kubernetes.io/projected/1cec2292-f508-42b2-916f-9a5808045626-kube-api-access-2k2z5\") pod \"console-f9d7485db-46hjn\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.623024 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.643829 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.663351 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.722315 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-rf8k5" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.736520 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m7qjx"] Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.739008 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qj7bk" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.753207 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5ptd\" (UniqueName: \"kubernetes.io/projected/d2c83122-863b-41e6-9118-54fa716c8485-kube-api-access-r5ptd\") pod \"dns-operator-744455d44c-pnffr\" (UID: \"d2c83122-863b-41e6-9118-54fa716c8485\") " pod="openshift-dns-operator/dns-operator-744455d44c-pnffr" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.753283 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fedd09d-87e7-4bfc-ab58-589189c984d8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-9pgcg\" (UID: \"1fedd09d-87e7-4bfc-ab58-589189c984d8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9pgcg" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.753343 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fcd663f-3b4c-42fe-a585-c1e1142875ca-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-79fhl\" (UID: \"7fcd663f-3b4c-42fe-a585-c1e1142875ca\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-79fhl" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.753366 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d2c83122-863b-41e6-9118-54fa716c8485-metrics-tls\") pod \"dns-operator-744455d44c-pnffr\" (UID: \"d2c83122-863b-41e6-9118-54fa716c8485\") " pod="openshift-dns-operator/dns-operator-744455d44c-pnffr" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.753435 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc3e8cef-c459-47ef-9531-da9acd77054e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n2mqb\" (UID: \"fc3e8cef-c459-47ef-9531-da9acd77054e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2mqb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.753459 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fcd663f-3b4c-42fe-a585-c1e1142875ca-config\") pod \"kube-apiserver-operator-766d6c64bb-79fhl\" (UID: \"7fcd663f-3b4c-42fe-a585-c1e1142875ca\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-79fhl" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.753542 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.753567 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/28e264f3-49f6-406a-b700-213e196574c1-registry-tls\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.753592 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/28e264f3-49f6-406a-b700-213e196574c1-registry-certificates\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.753624 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34f027ae-fb71-460c-b978-e254fac5d50f-config\") pod \"console-operator-58897d9998-47kcd\" (UID: \"34f027ae-fb71-460c-b978-e254fac5d50f\") " pod="openshift-console-operator/console-operator-58897d9998-47kcd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.753644 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fc3e8cef-c459-47ef-9531-da9acd77054e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n2mqb\" (UID: \"fc3e8cef-c459-47ef-9531-da9acd77054e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2mqb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.753743 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg8mb\" (UniqueName: \"kubernetes.io/projected/28e264f3-49f6-406a-b700-213e196574c1-kube-api-access-pg8mb\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.753809 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lshrt\" (UniqueName: \"kubernetes.io/projected/34f027ae-fb71-460c-b978-e254fac5d50f-kube-api-access-lshrt\") pod \"console-operator-58897d9998-47kcd\" (UID: \"34f027ae-fb71-460c-b978-e254fac5d50f\") " pod="openshift-console-operator/console-operator-58897d9998-47kcd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.753846 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7fcd663f-3b4c-42fe-a585-c1e1142875ca-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-79fhl\" (UID: \"7fcd663f-3b4c-42fe-a585-c1e1142875ca\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-79fhl" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.753872 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwghk\" (UniqueName: \"kubernetes.io/projected/1fedd09d-87e7-4bfc-ab58-589189c984d8-kube-api-access-kwghk\") pod \"openshift-controller-manager-operator-756b6f6bc6-9pgcg\" (UID: \"1fedd09d-87e7-4bfc-ab58-589189c984d8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9pgcg" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.753898 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/28e264f3-49f6-406a-b700-213e196574c1-trusted-ca\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.753955 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/28e264f3-49f6-406a-b700-213e196574c1-installation-pull-secrets\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.753995 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/28e264f3-49f6-406a-b700-213e196574c1-ca-trust-extracted\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.754025 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/28e264f3-49f6-406a-b700-213e196574c1-bound-sa-token\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.754049 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/34f027ae-fb71-460c-b978-e254fac5d50f-trusted-ca\") pod \"console-operator-58897d9998-47kcd\" (UID: \"34f027ae-fb71-460c-b978-e254fac5d50f\") " pod="openshift-console-operator/console-operator-58897d9998-47kcd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.754160 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fedd09d-87e7-4bfc-ab58-589189c984d8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-9pgcg\" (UID: \"1fedd09d-87e7-4bfc-ab58-589189c984d8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9pgcg" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.754233 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc3e8cef-c459-47ef-9531-da9acd77054e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n2mqb\" (UID: \"fc3e8cef-c459-47ef-9531-da9acd77054e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2mqb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.754255 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34f027ae-fb71-460c-b978-e254fac5d50f-serving-cert\") pod \"console-operator-58897d9998-47kcd\" (UID: \"34f027ae-fb71-460c-b978-e254fac5d50f\") " pod="openshift-console-operator/console-operator-58897d9998-47kcd" Mar 18 20:10:04 crc kubenswrapper[4950]: E0318 20:10:04.776232 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:05.276199192 +0000 UTC m=+218.517041070 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.781923 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-s8cr9"] Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.815888 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-kjb6z"] Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.819042 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8jvmc"] Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.824897 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m7qjx" event={"ID":"27a9fc13-1b12-40f1-9de4-fc04f3810df9","Type":"ContainerStarted","Data":"09b2356c2308100d823a2cd804f2d499d2ce2468441ef96ade2672a4d28b3570"} Mar 18 20:10:04 crc kubenswrapper[4950]: W0318 20:10:04.835093 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod792583b6_449b_4b55_b138_cf51971e2a17.slice/crio-96201c2f5830d3706fd4ca0ea88de47dbb8af44b177bec23fefdbe72132c6c2b WatchSource:0}: Error finding container 96201c2f5830d3706fd4ca0ea88de47dbb8af44b177bec23fefdbe72132c6c2b: Status 404 returned error can't find the container with id 96201c2f5830d3706fd4ca0ea88de47dbb8af44b177bec23fefdbe72132c6c2b Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.839615 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" event={"ID":"94495808-8600-45f6-a7fe-74ded4bd3ec4","Type":"ContainerStarted","Data":"6554a4338c0d8074ba1cafa349754420b2a43f0b1a801ee6abad4f7293d58bb5"} Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.839653 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" event={"ID":"94495808-8600-45f6-a7fe-74ded4bd3ec4","Type":"ContainerStarted","Data":"01c005459aac84dc7f80dbdbc25307698b3d7c74eed4d546d239541914d5cbf7"} Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.840622 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.840674 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9l9mq"] Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.844618 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p9pk8" event={"ID":"cfd25023-4dd6-4527-bcde-836258333c37","Type":"ContainerStarted","Data":"8edebb04adce3015d9d2c2f31f5551b76411fc56d5f9e6f0e392002ae0ff3ac8"} Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.844662 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p9pk8" event={"ID":"cfd25023-4dd6-4527-bcde-836258333c37","Type":"ContainerStarted","Data":"0028e3731008252540f6b996f6e3fb9ced55306d4c2113cd54a9ceccb26f101a"} Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.844749 4950 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-wcd54 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.844773 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" podUID="94495808-8600-45f6-a7fe-74ded4bd3ec4" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.854298 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-g2f49"] Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.854770 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.854971 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/30bb01b5-2b19-403e-9930-a1991532e787-bound-sa-token\") pod \"ingress-operator-5b745b69d9-twwmd\" (UID: \"30bb01b5-2b19-403e-9930-a1991532e787\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-twwmd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855005 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc3e8cef-c459-47ef-9531-da9acd77054e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n2mqb\" (UID: \"fc3e8cef-c459-47ef-9531-da9acd77054e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2mqb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855021 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/164efb7c-0b83-42ed-8d30-acc8a1ff1794-secret-volume\") pod \"collect-profiles-29564400-d9vqv\" (UID: \"164efb7c-0b83-42ed-8d30-acc8a1ff1794\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564400-d9vqv" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855045 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fcd663f-3b4c-42fe-a585-c1e1142875ca-config\") pod \"kube-apiserver-operator-766d6c64bb-79fhl\" (UID: \"7fcd663f-3b4c-42fe-a585-c1e1142875ca\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-79fhl" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855060 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjf7l\" (UniqueName: \"kubernetes.io/projected/627e6f3d-cde5-4849-bd02-6437622f4978-kube-api-access-cjf7l\") pod \"auto-csr-approver-29564410-9hqrd\" (UID: \"627e6f3d-cde5-4849-bd02-6437622f4978\") " pod="openshift-infra/auto-csr-approver-29564410-9hqrd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855078 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncvzb\" (UniqueName: \"kubernetes.io/projected/5d345627-baa3-4354-a422-1b3646e5cc48-kube-api-access-ncvzb\") pod \"ingress-canary-kbjxd\" (UID: \"5d345627-baa3-4354-a422-1b3646e5cc48\") " pod="openshift-ingress-canary/ingress-canary-kbjxd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855095 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/28e264f3-49f6-406a-b700-213e196574c1-registry-tls\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855109 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkgdb\" (UniqueName: \"kubernetes.io/projected/28014e20-2d95-42bc-881b-026137a7a016-kube-api-access-zkgdb\") pod \"package-server-manager-789f6589d5-qqfjh\" (UID: \"28014e20-2d95-42bc-881b-026137a7a016\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qqfjh" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855131 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/94700472-065a-4371-8025-5c4e28b1e722-srv-cert\") pod \"catalog-operator-68c6474976-vnw2w\" (UID: \"94700472-065a-4371-8025-5c4e28b1e722\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vnw2w" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855152 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcbnx\" (UniqueName: \"kubernetes.io/projected/f962a9b5-16d9-41e3-93a1-d7347e8cc27f-kube-api-access-zcbnx\") pod \"migrator-59844c95c7-c7t4g\" (UID: \"f962a9b5-16d9-41e3-93a1-d7347e8cc27f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c7t4g" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855169 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdb55\" (UniqueName: \"kubernetes.io/projected/8edc3dbe-4856-4b87-b5d8-bbf2c55c0127-kube-api-access-fdb55\") pod \"multus-admission-controller-857f4d67dd-5wffk\" (UID: \"8edc3dbe-4856-4b87-b5d8-bbf2c55c0127\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5wffk" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855183 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/94700472-065a-4371-8025-5c4e28b1e722-profile-collector-cert\") pod \"catalog-operator-68c6474976-vnw2w\" (UID: \"94700472-065a-4371-8025-5c4e28b1e722\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vnw2w" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855215 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg8mb\" (UniqueName: \"kubernetes.io/projected/28e264f3-49f6-406a-b700-213e196574c1-kube-api-access-pg8mb\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855231 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8596ab2e-852c-4d94-a8ce-e02c823a76b1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sxcm9\" (UID: \"8596ab2e-852c-4d94-a8ce-e02c823a76b1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sxcm9" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855259 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fd0504b1-d681-4d1d-a9bb-8be90167f30f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-tfrtt\" (UID: \"fd0504b1-d681-4d1d-a9bb-8be90167f30f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tfrtt" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855273 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf43d927-8bd0-46df-9115-a95e6f747190-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-859lm\" (UID: \"cf43d927-8bd0-46df-9115-a95e6f747190\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-859lm" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855290 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/90ab43dd-38e6-4264-a086-5cef08691214-stats-auth\") pod \"router-default-5444994796-gsj4z\" (UID: \"90ab43dd-38e6-4264-a086-5cef08691214\") " pod="openshift-ingress/router-default-5444994796-gsj4z" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855305 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a2456057-5ff6-458f-af94-53edca249b7f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8nhwt\" (UID: \"a2456057-5ff6-458f-af94-53edca249b7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855321 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/28e264f3-49f6-406a-b700-213e196574c1-installation-pull-secrets\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855338 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/34f027ae-fb71-460c-b978-e254fac5d50f-trusted-ca\") pod \"console-operator-58897d9998-47kcd\" (UID: \"34f027ae-fb71-460c-b978-e254fac5d50f\") " pod="openshift-console-operator/console-operator-58897d9998-47kcd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855355 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/28014e20-2d95-42bc-881b-026137a7a016-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-qqfjh\" (UID: \"28014e20-2d95-42bc-881b-026137a7a016\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qqfjh" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855372 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/0560ae4e-4142-48c1-9829-2e770a3b56b4-csi-data-dir\") pod \"csi-hostpathplugin-qqrpb\" (UID: \"0560ae4e-4142-48c1-9829-2e770a3b56b4\") " pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855388 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8edc3dbe-4856-4b87-b5d8-bbf2c55c0127-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-5wffk\" (UID: \"8edc3dbe-4856-4b87-b5d8-bbf2c55c0127\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5wffk" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855404 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/30bb01b5-2b19-403e-9930-a1991532e787-trusted-ca\") pod \"ingress-operator-5b745b69d9-twwmd\" (UID: \"30bb01b5-2b19-403e-9930-a1991532e787\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-twwmd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855442 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn5w6\" (UniqueName: \"kubernetes.io/projected/fd0504b1-d681-4d1d-a9bb-8be90167f30f-kube-api-access-zn5w6\") pod \"olm-operator-6b444d44fb-tfrtt\" (UID: \"fd0504b1-d681-4d1d-a9bb-8be90167f30f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tfrtt" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855467 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb5hp\" (UniqueName: \"kubernetes.io/projected/94700472-065a-4371-8025-5c4e28b1e722-kube-api-access-gb5hp\") pod \"catalog-operator-68c6474976-vnw2w\" (UID: \"94700472-065a-4371-8025-5c4e28b1e722\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vnw2w" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855485 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd2f8\" (UniqueName: \"kubernetes.io/projected/30bb01b5-2b19-403e-9930-a1991532e787-kube-api-access-wd2f8\") pod \"ingress-operator-5b745b69d9-twwmd\" (UID: \"30bb01b5-2b19-403e-9930-a1991532e787\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-twwmd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855505 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f68ebf05-2b93-49f5-874a-351cf9c9d974-node-bootstrap-token\") pod \"machine-config-server-rwl8z\" (UID: \"f68ebf05-2b93-49f5-874a-351cf9c9d974\") " pod="openshift-machine-config-operator/machine-config-server-rwl8z" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855529 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fedd09d-87e7-4bfc-ab58-589189c984d8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-9pgcg\" (UID: \"1fedd09d-87e7-4bfc-ab58-589189c984d8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9pgcg" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855549 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qscws\" (UniqueName: \"kubernetes.io/projected/0560ae4e-4142-48c1-9829-2e770a3b56b4-kube-api-access-qscws\") pod \"csi-hostpathplugin-qqrpb\" (UID: \"0560ae4e-4142-48c1-9829-2e770a3b56b4\") " pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855569 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/167a9973-d839-4aaf-973b-2e00be41f59f-images\") pod \"machine-config-operator-74547568cd-pdnzp\" (UID: \"167a9973-d839-4aaf-973b-2e00be41f59f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pdnzp" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855588 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90ab43dd-38e6-4264-a086-5cef08691214-service-ca-bundle\") pod \"router-default-5444994796-gsj4z\" (UID: \"90ab43dd-38e6-4264-a086-5cef08691214\") " pod="openshift-ingress/router-default-5444994796-gsj4z" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855606 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/167a9973-d839-4aaf-973b-2e00be41f59f-proxy-tls\") pod \"machine-config-operator-74547568cd-pdnzp\" (UID: \"167a9973-d839-4aaf-973b-2e00be41f59f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pdnzp" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855637 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc3e8cef-c459-47ef-9531-da9acd77054e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n2mqb\" (UID: \"fc3e8cef-c459-47ef-9531-da9acd77054e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2mqb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855676 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c56j5\" (UniqueName: \"kubernetes.io/projected/cf43d927-8bd0-46df-9115-a95e6f747190-kube-api-access-c56j5\") pod \"kube-storage-version-migrator-operator-b67b599dd-859lm\" (UID: \"cf43d927-8bd0-46df-9115-a95e6f747190\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-859lm" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855704 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb10c1c4-bc22-46af-9f3f-c948d34e0017-config\") pod \"service-ca-operator-777779d784-xj696\" (UID: \"bb10c1c4-bc22-46af-9f3f-c948d34e0017\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xj696" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855734 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/0560ae4e-4142-48c1-9829-2e770a3b56b4-socket-dir\") pod \"csi-hostpathplugin-qqrpb\" (UID: \"0560ae4e-4142-48c1-9829-2e770a3b56b4\") " pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855760 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5ptd\" (UniqueName: \"kubernetes.io/projected/d2c83122-863b-41e6-9118-54fa716c8485-kube-api-access-r5ptd\") pod \"dns-operator-744455d44c-pnffr\" (UID: \"d2c83122-863b-41e6-9118-54fa716c8485\") " pod="openshift-dns-operator/dns-operator-744455d44c-pnffr" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855797 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fedd09d-87e7-4bfc-ab58-589189c984d8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-9pgcg\" (UID: \"1fedd09d-87e7-4bfc-ab58-589189c984d8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9pgcg" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855820 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/035f024b-dad2-4ad3-96bb-ebc3c52ff9d4-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-544q6\" (UID: \"035f024b-dad2-4ad3-96bb-ebc3c52ff9d4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-544q6" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855842 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0274f224-f29d-4c71-88e7-2111719f598b-tmpfs\") pod \"packageserver-d55dfcdfc-jq2zs\" (UID: \"0274f224-f29d-4c71-88e7-2111719f598b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855861 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fd0504b1-d681-4d1d-a9bb-8be90167f30f-srv-cert\") pod \"olm-operator-6b444d44fb-tfrtt\" (UID: \"fd0504b1-d681-4d1d-a9bb-8be90167f30f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tfrtt" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855882 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5d345627-baa3-4354-a422-1b3646e5cc48-cert\") pod \"ingress-canary-kbjxd\" (UID: \"5d345627-baa3-4354-a422-1b3646e5cc48\") " pod="openshift-ingress-canary/ingress-canary-kbjxd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855903 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0274f224-f29d-4c71-88e7-2111719f598b-apiservice-cert\") pod \"packageserver-d55dfcdfc-jq2zs\" (UID: \"0274f224-f29d-4c71-88e7-2111719f598b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855922 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0274f224-f29d-4c71-88e7-2111719f598b-webhook-cert\") pod \"packageserver-d55dfcdfc-jq2zs\" (UID: \"0274f224-f29d-4c71-88e7-2111719f598b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855944 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fcd663f-3b4c-42fe-a585-c1e1142875ca-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-79fhl\" (UID: \"7fcd663f-3b4c-42fe-a585-c1e1142875ca\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-79fhl" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.855980 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d2c83122-863b-41e6-9118-54fa716c8485-metrics-tls\") pod \"dns-operator-744455d44c-pnffr\" (UID: \"d2c83122-863b-41e6-9118-54fa716c8485\") " pod="openshift-dns-operator/dns-operator-744455d44c-pnffr" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856003 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/90ab43dd-38e6-4264-a086-5cef08691214-metrics-certs\") pod \"router-default-5444994796-gsj4z\" (UID: \"90ab43dd-38e6-4264-a086-5cef08691214\") " pod="openshift-ingress/router-default-5444994796-gsj4z" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856023 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a2456057-5ff6-458f-af94-53edca249b7f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8nhwt\" (UID: \"a2456057-5ff6-458f-af94-53edca249b7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856044 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f68ebf05-2b93-49f5-874a-351cf9c9d974-certs\") pod \"machine-config-server-rwl8z\" (UID: \"f68ebf05-2b93-49f5-874a-351cf9c9d974\") " pod="openshift-machine-config-operator/machine-config-server-rwl8z" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856079 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rprg\" (UniqueName: \"kubernetes.io/projected/164efb7c-0b83-42ed-8d30-acc8a1ff1794-kube-api-access-2rprg\") pod \"collect-profiles-29564400-d9vqv\" (UID: \"164efb7c-0b83-42ed-8d30-acc8a1ff1794\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564400-d9vqv" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856107 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/28e264f3-49f6-406a-b700-213e196574c1-registry-certificates\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856128 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34f027ae-fb71-460c-b978-e254fac5d50f-config\") pod \"console-operator-58897d9998-47kcd\" (UID: \"34f027ae-fb71-460c-b978-e254fac5d50f\") " pod="openshift-console-operator/console-operator-58897d9998-47kcd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856152 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fc3e8cef-c459-47ef-9531-da9acd77054e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n2mqb\" (UID: \"fc3e8cef-c459-47ef-9531-da9acd77054e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2mqb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856178 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/0560ae4e-4142-48c1-9829-2e770a3b56b4-mountpoint-dir\") pod \"csi-hostpathplugin-qqrpb\" (UID: \"0560ae4e-4142-48c1-9829-2e770a3b56b4\") " pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856204 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64khz\" (UniqueName: \"kubernetes.io/projected/36384cbe-f2f7-48aa-8b54-ea07d8cc3678-kube-api-access-64khz\") pod \"machine-config-controller-84d6567774-gw45j\" (UID: \"36384cbe-f2f7-48aa-8b54-ea07d8cc3678\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gw45j" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856224 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vz67\" (UniqueName: \"kubernetes.io/projected/f68ebf05-2b93-49f5-874a-351cf9c9d974-kube-api-access-5vz67\") pod \"machine-config-server-rwl8z\" (UID: \"f68ebf05-2b93-49f5-874a-351cf9c9d974\") " pod="openshift-machine-config-operator/machine-config-server-rwl8z" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856245 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/90ab43dd-38e6-4264-a086-5cef08691214-default-certificate\") pod \"router-default-5444994796-gsj4z\" (UID: \"90ab43dd-38e6-4264-a086-5cef08691214\") " pod="openshift-ingress/router-default-5444994796-gsj4z" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856317 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b3f2101e-4192-42db-9ff8-3dd152bc24bb-signing-key\") pod \"service-ca-9c57cc56f-ktwzg\" (UID: \"b3f2101e-4192-42db-9ff8-3dd152bc24bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-ktwzg" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856350 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g6m2\" (UniqueName: \"kubernetes.io/projected/bb10c1c4-bc22-46af-9f3f-c948d34e0017-kube-api-access-5g6m2\") pod \"service-ca-operator-777779d784-xj696\" (UID: \"bb10c1c4-bc22-46af-9f3f-c948d34e0017\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xj696" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856390 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/0560ae4e-4142-48c1-9829-2e770a3b56b4-plugins-dir\") pod \"csi-hostpathplugin-qqrpb\" (UID: \"0560ae4e-4142-48c1-9829-2e770a3b56b4\") " pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856434 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lshrt\" (UniqueName: \"kubernetes.io/projected/34f027ae-fb71-460c-b978-e254fac5d50f-kube-api-access-lshrt\") pod \"console-operator-58897d9998-47kcd\" (UID: \"34f027ae-fb71-460c-b978-e254fac5d50f\") " pod="openshift-console-operator/console-operator-58897d9998-47kcd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856460 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/167a9973-d839-4aaf-973b-2e00be41f59f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-pdnzp\" (UID: \"167a9973-d839-4aaf-973b-2e00be41f59f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pdnzp" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856478 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf43d927-8bd0-46df-9115-a95e6f747190-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-859lm\" (UID: \"cf43d927-8bd0-46df-9115-a95e6f747190\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-859lm" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856495 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a8ef5a1d-9e3e-4057-89f5-672b0b1ca73c-metrics-tls\") pod \"dns-default-cdpgz\" (UID: \"a8ef5a1d-9e3e-4057-89f5-672b0b1ca73c\") " pod="openshift-dns/dns-default-cdpgz" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856525 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/035f024b-dad2-4ad3-96bb-ebc3c52ff9d4-config\") pod \"kube-controller-manager-operator-78b949d7b-544q6\" (UID: \"035f024b-dad2-4ad3-96bb-ebc3c52ff9d4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-544q6" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856562 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/28e264f3-49f6-406a-b700-213e196574c1-trusted-ca\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856581 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7fcd663f-3b4c-42fe-a585-c1e1142875ca-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-79fhl\" (UID: \"7fcd663f-3b4c-42fe-a585-c1e1142875ca\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-79fhl" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856603 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwghk\" (UniqueName: \"kubernetes.io/projected/1fedd09d-87e7-4bfc-ab58-589189c984d8-kube-api-access-kwghk\") pod \"openshift-controller-manager-operator-756b6f6bc6-9pgcg\" (UID: \"1fedd09d-87e7-4bfc-ab58-589189c984d8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9pgcg" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856622 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/36384cbe-f2f7-48aa-8b54-ea07d8cc3678-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-gw45j\" (UID: \"36384cbe-f2f7-48aa-8b54-ea07d8cc3678\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gw45j" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.856640 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/035f024b-dad2-4ad3-96bb-ebc3c52ff9d4-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-544q6\" (UID: \"035f024b-dad2-4ad3-96bb-ebc3c52ff9d4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-544q6" Mar 18 20:10:04 crc kubenswrapper[4950]: E0318 20:10:04.857977 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:05.357949306 +0000 UTC m=+218.598791174 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.860379 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/34f027ae-fb71-460c-b978-e254fac5d50f-trusted-ca\") pod \"console-operator-58897d9998-47kcd\" (UID: \"34f027ae-fb71-460c-b978-e254fac5d50f\") " pod="openshift-console-operator/console-operator-58897d9998-47kcd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.860891 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fedd09d-87e7-4bfc-ab58-589189c984d8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-9pgcg\" (UID: \"1fedd09d-87e7-4bfc-ab58-589189c984d8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9pgcg" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.862208 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wv7kr"] Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.858817 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34f027ae-fb71-460c-b978-e254fac5d50f-config\") pod \"console-operator-58897d9998-47kcd\" (UID: \"34f027ae-fb71-460c-b978-e254fac5d50f\") " pod="openshift-console-operator/console-operator-58897d9998-47kcd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.862925 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/28e264f3-49f6-406a-b700-213e196574c1-installation-pull-secrets\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.863121 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fcd663f-3b4c-42fe-a585-c1e1142875ca-config\") pod \"kube-apiserver-operator-766d6c64bb-79fhl\" (UID: \"7fcd663f-3b4c-42fe-a585-c1e1142875ca\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-79fhl" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.864065 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/28e264f3-49f6-406a-b700-213e196574c1-trusted-ca\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.864229 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b3f2101e-4192-42db-9ff8-3dd152bc24bb-signing-cabundle\") pod \"service-ca-9c57cc56f-ktwzg\" (UID: \"b3f2101e-4192-42db-9ff8-3dd152bc24bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-ktwzg" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.864265 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/36384cbe-f2f7-48aa-8b54-ea07d8cc3678-proxy-tls\") pod \"machine-config-controller-84d6567774-gw45j\" (UID: \"36384cbe-f2f7-48aa-8b54-ea07d8cc3678\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gw45j" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.864343 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/28e264f3-49f6-406a-b700-213e196574c1-ca-trust-extracted\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.864365 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/28e264f3-49f6-406a-b700-213e196574c1-bound-sa-token\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.864438 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/30bb01b5-2b19-403e-9930-a1991532e787-metrics-tls\") pod \"ingress-operator-5b745b69d9-twwmd\" (UID: \"30bb01b5-2b19-403e-9930-a1991532e787\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-twwmd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.865010 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/28e264f3-49f6-406a-b700-213e196574c1-ca-trust-extracted\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.865107 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/28e264f3-49f6-406a-b700-213e196574c1-registry-certificates\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.865171 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwwjp\" (UniqueName: \"kubernetes.io/projected/0274f224-f29d-4c71-88e7-2111719f598b-kube-api-access-lwwjp\") pod \"packageserver-d55dfcdfc-jq2zs\" (UID: \"0274f224-f29d-4c71-88e7-2111719f598b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.865196 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a8ef5a1d-9e3e-4057-89f5-672b0b1ca73c-config-volume\") pod \"dns-default-cdpgz\" (UID: \"a8ef5a1d-9e3e-4057-89f5-672b0b1ca73c\") " pod="openshift-dns/dns-default-cdpgz" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.865314 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34f027ae-fb71-460c-b978-e254fac5d50f-serving-cert\") pod \"console-operator-58897d9998-47kcd\" (UID: \"34f027ae-fb71-460c-b978-e254fac5d50f\") " pod="openshift-console-operator/console-operator-58897d9998-47kcd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.865482 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t4jq\" (UniqueName: \"kubernetes.io/projected/167a9973-d839-4aaf-973b-2e00be41f59f-kube-api-access-4t4jq\") pod \"machine-config-operator-74547568cd-pdnzp\" (UID: \"167a9973-d839-4aaf-973b-2e00be41f59f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pdnzp" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.865533 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb10c1c4-bc22-46af-9f3f-c948d34e0017-serving-cert\") pod \"service-ca-operator-777779d784-xj696\" (UID: \"bb10c1c4-bc22-46af-9f3f-c948d34e0017\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xj696" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.865911 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/0560ae4e-4142-48c1-9829-2e770a3b56b4-registration-dir\") pod \"csi-hostpathplugin-qqrpb\" (UID: \"0560ae4e-4142-48c1-9829-2e770a3b56b4\") " pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.865947 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2lpl\" (UniqueName: \"kubernetes.io/projected/8596ab2e-852c-4d94-a8ce-e02c823a76b1-kube-api-access-v2lpl\") pod \"control-plane-machine-set-operator-78cbb6b69f-sxcm9\" (UID: \"8596ab2e-852c-4d94-a8ce-e02c823a76b1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sxcm9" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.865994 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjg8b\" (UniqueName: \"kubernetes.io/projected/97cb884c-5f55-4ca0-8639-193c2f1df139-kube-api-access-sjg8b\") pod \"auto-csr-approver-29564408-2658d\" (UID: \"97cb884c-5f55-4ca0-8639-193c2f1df139\") " pod="openshift-infra/auto-csr-approver-29564408-2658d" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.866050 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/164efb7c-0b83-42ed-8d30-acc8a1ff1794-config-volume\") pod \"collect-profiles-29564400-d9vqv\" (UID: \"164efb7c-0b83-42ed-8d30-acc8a1ff1794\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564400-d9vqv" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.866068 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h96q2\" (UniqueName: \"kubernetes.io/projected/a8ef5a1d-9e3e-4057-89f5-672b0b1ca73c-kube-api-access-h96q2\") pod \"dns-default-cdpgz\" (UID: \"a8ef5a1d-9e3e-4057-89f5-672b0b1ca73c\") " pod="openshift-dns/dns-default-cdpgz" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.866286 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c4x9\" (UniqueName: \"kubernetes.io/projected/b3f2101e-4192-42db-9ff8-3dd152bc24bb-kube-api-access-5c4x9\") pod \"service-ca-9c57cc56f-ktwzg\" (UID: \"b3f2101e-4192-42db-9ff8-3dd152bc24bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-ktwzg" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.866318 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f75b5\" (UniqueName: \"kubernetes.io/projected/90ab43dd-38e6-4264-a086-5cef08691214-kube-api-access-f75b5\") pod \"router-default-5444994796-gsj4z\" (UID: \"90ab43dd-38e6-4264-a086-5cef08691214\") " pod="openshift-ingress/router-default-5444994796-gsj4z" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.866337 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sgrs\" (UniqueName: \"kubernetes.io/projected/a2456057-5ff6-458f-af94-53edca249b7f-kube-api-access-2sgrs\") pod \"marketplace-operator-79b997595-8nhwt\" (UID: \"a2456057-5ff6-458f-af94-53edca249b7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.866554 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/28e264f3-49f6-406a-b700-213e196574c1-registry-tls\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.871094 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fcd663f-3b4c-42fe-a585-c1e1142875ca-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-79fhl\" (UID: \"7fcd663f-3b4c-42fe-a585-c1e1142875ca\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-79fhl" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.871705 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc3e8cef-c459-47ef-9531-da9acd77054e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n2mqb\" (UID: \"fc3e8cef-c459-47ef-9531-da9acd77054e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2mqb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.871971 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fedd09d-87e7-4bfc-ab58-589189c984d8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-9pgcg\" (UID: \"1fedd09d-87e7-4bfc-ab58-589189c984d8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9pgcg" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.872936 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34f027ae-fb71-460c-b978-e254fac5d50f-serving-cert\") pod \"console-operator-58897d9998-47kcd\" (UID: \"34f027ae-fb71-460c-b978-e254fac5d50f\") " pod="openshift-console-operator/console-operator-58897d9998-47kcd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.873707 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc3e8cef-c459-47ef-9531-da9acd77054e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n2mqb\" (UID: \"fc3e8cef-c459-47ef-9531-da9acd77054e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2mqb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.878112 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d2c83122-863b-41e6-9118-54fa716c8485-metrics-tls\") pod \"dns-operator-744455d44c-pnffr\" (UID: \"d2c83122-863b-41e6-9118-54fa716c8485\") " pod="openshift-dns-operator/dns-operator-744455d44c-pnffr" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.881493 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-27jdq" event={"ID":"ff73cd60-bdf9-477a-8ede-1e781aaffd22","Type":"ContainerStarted","Data":"b4a77e1ac9e7fc5693ac19d7b6c212c1118788563e86ffb2063da7efab8648e3"} Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.893518 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" event={"ID":"be93cdca-9fe6-46a2-ae6a-94516d8eaa50","Type":"ContainerStarted","Data":"99dff108e6a741c0cb557e3ce67dda5adc198d264a444dc8a76f5fd631a22be6"} Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.896219 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d5rvd" event={"ID":"cc0b2145-4578-4709-837d-6bae20399714","Type":"ContainerStarted","Data":"396ca88f0795c53bf7784c3e2fb1b53b38100cfa83e62626c5d72e45fe6e56b9"} Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.896249 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d5rvd" event={"ID":"cc0b2145-4578-4709-837d-6bae20399714","Type":"ContainerStarted","Data":"2eee547d81a03697521145fa814cbafca6afbdd6045e4c64f72c28e5dcc820b6"} Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.900058 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg8mb\" (UniqueName: \"kubernetes.io/projected/28e264f3-49f6-406a-b700-213e196574c1-kube-api-access-pg8mb\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.900907 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.934120 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lshrt\" (UniqueName: \"kubernetes.io/projected/34f027ae-fb71-460c-b978-e254fac5d50f-kube-api-access-lshrt\") pod \"console-operator-58897d9998-47kcd\" (UID: \"34f027ae-fb71-460c-b978-e254fac5d50f\") " pod="openshift-console-operator/console-operator-58897d9998-47kcd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.936344 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5ptd\" (UniqueName: \"kubernetes.io/projected/d2c83122-863b-41e6-9118-54fa716c8485-kube-api-access-r5ptd\") pod \"dns-operator-744455d44c-pnffr\" (UID: \"d2c83122-863b-41e6-9118-54fa716c8485\") " pod="openshift-dns-operator/dns-operator-744455d44c-pnffr" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.960985 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fc3e8cef-c459-47ef-9531-da9acd77054e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n2mqb\" (UID: \"fc3e8cef-c459-47ef-9531-da9acd77054e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2mqb" Mar 18 20:10:04 crc kubenswrapper[4950]: W0318 20:10:04.967502 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4684cfd6_1836_4218_9c6e_c6ca75a2da96.slice/crio-6c421fd7ff8d53d0c82d315c646b14b6ad14b33c1d5ef10c5d566647522b6b91 WatchSource:0}: Error finding container 6c421fd7ff8d53d0c82d315c646b14b6ad14b33c1d5ef10c5d566647522b6b91: Status 404 returned error can't find the container with id 6c421fd7ff8d53d0c82d315c646b14b6ad14b33c1d5ef10c5d566647522b6b91 Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969235 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8596ab2e-852c-4d94-a8ce-e02c823a76b1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sxcm9\" (UID: \"8596ab2e-852c-4d94-a8ce-e02c823a76b1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sxcm9" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969297 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fd0504b1-d681-4d1d-a9bb-8be90167f30f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-tfrtt\" (UID: \"fd0504b1-d681-4d1d-a9bb-8be90167f30f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tfrtt" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969315 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a2456057-5ff6-458f-af94-53edca249b7f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8nhwt\" (UID: \"a2456057-5ff6-458f-af94-53edca249b7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969337 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf43d927-8bd0-46df-9115-a95e6f747190-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-859lm\" (UID: \"cf43d927-8bd0-46df-9115-a95e6f747190\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-859lm" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969365 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/90ab43dd-38e6-4264-a086-5cef08691214-stats-auth\") pod \"router-default-5444994796-gsj4z\" (UID: \"90ab43dd-38e6-4264-a086-5cef08691214\") " pod="openshift-ingress/router-default-5444994796-gsj4z" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969382 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/28014e20-2d95-42bc-881b-026137a7a016-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-qqfjh\" (UID: \"28014e20-2d95-42bc-881b-026137a7a016\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qqfjh" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969401 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/0560ae4e-4142-48c1-9829-2e770a3b56b4-csi-data-dir\") pod \"csi-hostpathplugin-qqrpb\" (UID: \"0560ae4e-4142-48c1-9829-2e770a3b56b4\") " pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969435 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8edc3dbe-4856-4b87-b5d8-bbf2c55c0127-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-5wffk\" (UID: \"8edc3dbe-4856-4b87-b5d8-bbf2c55c0127\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5wffk" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969451 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn5w6\" (UniqueName: \"kubernetes.io/projected/fd0504b1-d681-4d1d-a9bb-8be90167f30f-kube-api-access-zn5w6\") pod \"olm-operator-6b444d44fb-tfrtt\" (UID: \"fd0504b1-d681-4d1d-a9bb-8be90167f30f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tfrtt" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969466 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb5hp\" (UniqueName: \"kubernetes.io/projected/94700472-065a-4371-8025-5c4e28b1e722-kube-api-access-gb5hp\") pod \"catalog-operator-68c6474976-vnw2w\" (UID: \"94700472-065a-4371-8025-5c4e28b1e722\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vnw2w" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969483 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/30bb01b5-2b19-403e-9930-a1991532e787-trusted-ca\") pod \"ingress-operator-5b745b69d9-twwmd\" (UID: \"30bb01b5-2b19-403e-9930-a1991532e787\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-twwmd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969514 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qscws\" (UniqueName: \"kubernetes.io/projected/0560ae4e-4142-48c1-9829-2e770a3b56b4-kube-api-access-qscws\") pod \"csi-hostpathplugin-qqrpb\" (UID: \"0560ae4e-4142-48c1-9829-2e770a3b56b4\") " pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969529 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/167a9973-d839-4aaf-973b-2e00be41f59f-images\") pod \"machine-config-operator-74547568cd-pdnzp\" (UID: \"167a9973-d839-4aaf-973b-2e00be41f59f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pdnzp" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969544 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd2f8\" (UniqueName: \"kubernetes.io/projected/30bb01b5-2b19-403e-9930-a1991532e787-kube-api-access-wd2f8\") pod \"ingress-operator-5b745b69d9-twwmd\" (UID: \"30bb01b5-2b19-403e-9930-a1991532e787\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-twwmd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969566 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f68ebf05-2b93-49f5-874a-351cf9c9d974-node-bootstrap-token\") pod \"machine-config-server-rwl8z\" (UID: \"f68ebf05-2b93-49f5-874a-351cf9c9d974\") " pod="openshift-machine-config-operator/machine-config-server-rwl8z" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969599 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90ab43dd-38e6-4264-a086-5cef08691214-service-ca-bundle\") pod \"router-default-5444994796-gsj4z\" (UID: \"90ab43dd-38e6-4264-a086-5cef08691214\") " pod="openshift-ingress/router-default-5444994796-gsj4z" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969619 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/167a9973-d839-4aaf-973b-2e00be41f59f-proxy-tls\") pod \"machine-config-operator-74547568cd-pdnzp\" (UID: \"167a9973-d839-4aaf-973b-2e00be41f59f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pdnzp" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969667 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c56j5\" (UniqueName: \"kubernetes.io/projected/cf43d927-8bd0-46df-9115-a95e6f747190-kube-api-access-c56j5\") pod \"kube-storage-version-migrator-operator-b67b599dd-859lm\" (UID: \"cf43d927-8bd0-46df-9115-a95e6f747190\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-859lm" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969685 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb10c1c4-bc22-46af-9f3f-c948d34e0017-config\") pod \"service-ca-operator-777779d784-xj696\" (UID: \"bb10c1c4-bc22-46af-9f3f-c948d34e0017\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xj696" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969708 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/0560ae4e-4142-48c1-9829-2e770a3b56b4-socket-dir\") pod \"csi-hostpathplugin-qqrpb\" (UID: \"0560ae4e-4142-48c1-9829-2e770a3b56b4\") " pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969744 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/035f024b-dad2-4ad3-96bb-ebc3c52ff9d4-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-544q6\" (UID: \"035f024b-dad2-4ad3-96bb-ebc3c52ff9d4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-544q6" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969760 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0274f224-f29d-4c71-88e7-2111719f598b-tmpfs\") pod \"packageserver-d55dfcdfc-jq2zs\" (UID: \"0274f224-f29d-4c71-88e7-2111719f598b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969776 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0274f224-f29d-4c71-88e7-2111719f598b-apiservice-cert\") pod \"packageserver-d55dfcdfc-jq2zs\" (UID: \"0274f224-f29d-4c71-88e7-2111719f598b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969791 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0274f224-f29d-4c71-88e7-2111719f598b-webhook-cert\") pod \"packageserver-d55dfcdfc-jq2zs\" (UID: \"0274f224-f29d-4c71-88e7-2111719f598b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969822 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fd0504b1-d681-4d1d-a9bb-8be90167f30f-srv-cert\") pod \"olm-operator-6b444d44fb-tfrtt\" (UID: \"fd0504b1-d681-4d1d-a9bb-8be90167f30f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tfrtt" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969838 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5d345627-baa3-4354-a422-1b3646e5cc48-cert\") pod \"ingress-canary-kbjxd\" (UID: \"5d345627-baa3-4354-a422-1b3646e5cc48\") " pod="openshift-ingress-canary/ingress-canary-kbjxd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969856 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/90ab43dd-38e6-4264-a086-5cef08691214-metrics-certs\") pod \"router-default-5444994796-gsj4z\" (UID: \"90ab43dd-38e6-4264-a086-5cef08691214\") " pod="openshift-ingress/router-default-5444994796-gsj4z" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969870 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f68ebf05-2b93-49f5-874a-351cf9c9d974-certs\") pod \"machine-config-server-rwl8z\" (UID: \"f68ebf05-2b93-49f5-874a-351cf9c9d974\") " pod="openshift-machine-config-operator/machine-config-server-rwl8z" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969904 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a2456057-5ff6-458f-af94-53edca249b7f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8nhwt\" (UID: \"a2456057-5ff6-458f-af94-53edca249b7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969924 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rprg\" (UniqueName: \"kubernetes.io/projected/164efb7c-0b83-42ed-8d30-acc8a1ff1794-kube-api-access-2rprg\") pod \"collect-profiles-29564400-d9vqv\" (UID: \"164efb7c-0b83-42ed-8d30-acc8a1ff1794\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564400-d9vqv" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969944 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969979 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/0560ae4e-4142-48c1-9829-2e770a3b56b4-mountpoint-dir\") pod \"csi-hostpathplugin-qqrpb\" (UID: \"0560ae4e-4142-48c1-9829-2e770a3b56b4\") " pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.969996 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64khz\" (UniqueName: \"kubernetes.io/projected/36384cbe-f2f7-48aa-8b54-ea07d8cc3678-kube-api-access-64khz\") pod \"machine-config-controller-84d6567774-gw45j\" (UID: \"36384cbe-f2f7-48aa-8b54-ea07d8cc3678\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gw45j" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970012 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vz67\" (UniqueName: \"kubernetes.io/projected/f68ebf05-2b93-49f5-874a-351cf9c9d974-kube-api-access-5vz67\") pod \"machine-config-server-rwl8z\" (UID: \"f68ebf05-2b93-49f5-874a-351cf9c9d974\") " pod="openshift-machine-config-operator/machine-config-server-rwl8z" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970027 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/90ab43dd-38e6-4264-a086-5cef08691214-default-certificate\") pod \"router-default-5444994796-gsj4z\" (UID: \"90ab43dd-38e6-4264-a086-5cef08691214\") " pod="openshift-ingress/router-default-5444994796-gsj4z" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970063 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b3f2101e-4192-42db-9ff8-3dd152bc24bb-signing-key\") pod \"service-ca-9c57cc56f-ktwzg\" (UID: \"b3f2101e-4192-42db-9ff8-3dd152bc24bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-ktwzg" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970081 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g6m2\" (UniqueName: \"kubernetes.io/projected/bb10c1c4-bc22-46af-9f3f-c948d34e0017-kube-api-access-5g6m2\") pod \"service-ca-operator-777779d784-xj696\" (UID: \"bb10c1c4-bc22-46af-9f3f-c948d34e0017\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xj696" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970103 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/0560ae4e-4142-48c1-9829-2e770a3b56b4-plugins-dir\") pod \"csi-hostpathplugin-qqrpb\" (UID: \"0560ae4e-4142-48c1-9829-2e770a3b56b4\") " pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970145 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/167a9973-d839-4aaf-973b-2e00be41f59f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-pdnzp\" (UID: \"167a9973-d839-4aaf-973b-2e00be41f59f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pdnzp" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970169 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf43d927-8bd0-46df-9115-a95e6f747190-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-859lm\" (UID: \"cf43d927-8bd0-46df-9115-a95e6f747190\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-859lm" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970189 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a8ef5a1d-9e3e-4057-89f5-672b0b1ca73c-metrics-tls\") pod \"dns-default-cdpgz\" (UID: \"a8ef5a1d-9e3e-4057-89f5-672b0b1ca73c\") " pod="openshift-dns/dns-default-cdpgz" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970237 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/035f024b-dad2-4ad3-96bb-ebc3c52ff9d4-config\") pod \"kube-controller-manager-operator-78b949d7b-544q6\" (UID: \"035f024b-dad2-4ad3-96bb-ebc3c52ff9d4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-544q6" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970265 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/36384cbe-f2f7-48aa-8b54-ea07d8cc3678-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-gw45j\" (UID: \"36384cbe-f2f7-48aa-8b54-ea07d8cc3678\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gw45j" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970312 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/035f024b-dad2-4ad3-96bb-ebc3c52ff9d4-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-544q6\" (UID: \"035f024b-dad2-4ad3-96bb-ebc3c52ff9d4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-544q6" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970328 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b3f2101e-4192-42db-9ff8-3dd152bc24bb-signing-cabundle\") pod \"service-ca-9c57cc56f-ktwzg\" (UID: \"b3f2101e-4192-42db-9ff8-3dd152bc24bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-ktwzg" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970347 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/36384cbe-f2f7-48aa-8b54-ea07d8cc3678-proxy-tls\") pod \"machine-config-controller-84d6567774-gw45j\" (UID: \"36384cbe-f2f7-48aa-8b54-ea07d8cc3678\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gw45j" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970382 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/30bb01b5-2b19-403e-9930-a1991532e787-metrics-tls\") pod \"ingress-operator-5b745b69d9-twwmd\" (UID: \"30bb01b5-2b19-403e-9930-a1991532e787\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-twwmd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970399 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwwjp\" (UniqueName: \"kubernetes.io/projected/0274f224-f29d-4c71-88e7-2111719f598b-kube-api-access-lwwjp\") pod \"packageserver-d55dfcdfc-jq2zs\" (UID: \"0274f224-f29d-4c71-88e7-2111719f598b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970599 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a8ef5a1d-9e3e-4057-89f5-672b0b1ca73c-config-volume\") pod \"dns-default-cdpgz\" (UID: \"a8ef5a1d-9e3e-4057-89f5-672b0b1ca73c\") " pod="openshift-dns/dns-default-cdpgz" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970618 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t4jq\" (UniqueName: \"kubernetes.io/projected/167a9973-d839-4aaf-973b-2e00be41f59f-kube-api-access-4t4jq\") pod \"machine-config-operator-74547568cd-pdnzp\" (UID: \"167a9973-d839-4aaf-973b-2e00be41f59f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pdnzp" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970633 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb10c1c4-bc22-46af-9f3f-c948d34e0017-serving-cert\") pod \"service-ca-operator-777779d784-xj696\" (UID: \"bb10c1c4-bc22-46af-9f3f-c948d34e0017\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xj696" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970698 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/0560ae4e-4142-48c1-9829-2e770a3b56b4-registration-dir\") pod \"csi-hostpathplugin-qqrpb\" (UID: \"0560ae4e-4142-48c1-9829-2e770a3b56b4\") " pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970721 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2lpl\" (UniqueName: \"kubernetes.io/projected/8596ab2e-852c-4d94-a8ce-e02c823a76b1-kube-api-access-v2lpl\") pod \"control-plane-machine-set-operator-78cbb6b69f-sxcm9\" (UID: \"8596ab2e-852c-4d94-a8ce-e02c823a76b1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sxcm9" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970760 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjg8b\" (UniqueName: \"kubernetes.io/projected/97cb884c-5f55-4ca0-8639-193c2f1df139-kube-api-access-sjg8b\") pod \"auto-csr-approver-29564408-2658d\" (UID: \"97cb884c-5f55-4ca0-8639-193c2f1df139\") " pod="openshift-infra/auto-csr-approver-29564408-2658d" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970776 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/164efb7c-0b83-42ed-8d30-acc8a1ff1794-config-volume\") pod \"collect-profiles-29564400-d9vqv\" (UID: \"164efb7c-0b83-42ed-8d30-acc8a1ff1794\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564400-d9vqv" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970791 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h96q2\" (UniqueName: \"kubernetes.io/projected/a8ef5a1d-9e3e-4057-89f5-672b0b1ca73c-kube-api-access-h96q2\") pod \"dns-default-cdpgz\" (UID: \"a8ef5a1d-9e3e-4057-89f5-672b0b1ca73c\") " pod="openshift-dns/dns-default-cdpgz" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970824 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f75b5\" (UniqueName: \"kubernetes.io/projected/90ab43dd-38e6-4264-a086-5cef08691214-kube-api-access-f75b5\") pod \"router-default-5444994796-gsj4z\" (UID: \"90ab43dd-38e6-4264-a086-5cef08691214\") " pod="openshift-ingress/router-default-5444994796-gsj4z" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970840 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sgrs\" (UniqueName: \"kubernetes.io/projected/a2456057-5ff6-458f-af94-53edca249b7f-kube-api-access-2sgrs\") pod \"marketplace-operator-79b997595-8nhwt\" (UID: \"a2456057-5ff6-458f-af94-53edca249b7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970856 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c4x9\" (UniqueName: \"kubernetes.io/projected/b3f2101e-4192-42db-9ff8-3dd152bc24bb-kube-api-access-5c4x9\") pod \"service-ca-9c57cc56f-ktwzg\" (UID: \"b3f2101e-4192-42db-9ff8-3dd152bc24bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-ktwzg" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970875 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/164efb7c-0b83-42ed-8d30-acc8a1ff1794-secret-volume\") pod \"collect-profiles-29564400-d9vqv\" (UID: \"164efb7c-0b83-42ed-8d30-acc8a1ff1794\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564400-d9vqv" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970910 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/30bb01b5-2b19-403e-9930-a1991532e787-bound-sa-token\") pod \"ingress-operator-5b745b69d9-twwmd\" (UID: \"30bb01b5-2b19-403e-9930-a1991532e787\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-twwmd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970928 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjf7l\" (UniqueName: \"kubernetes.io/projected/627e6f3d-cde5-4849-bd02-6437622f4978-kube-api-access-cjf7l\") pod \"auto-csr-approver-29564410-9hqrd\" (UID: \"627e6f3d-cde5-4849-bd02-6437622f4978\") " pod="openshift-infra/auto-csr-approver-29564410-9hqrd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970946 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncvzb\" (UniqueName: \"kubernetes.io/projected/5d345627-baa3-4354-a422-1b3646e5cc48-kube-api-access-ncvzb\") pod \"ingress-canary-kbjxd\" (UID: \"5d345627-baa3-4354-a422-1b3646e5cc48\") " pod="openshift-ingress-canary/ingress-canary-kbjxd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970962 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkgdb\" (UniqueName: \"kubernetes.io/projected/28014e20-2d95-42bc-881b-026137a7a016-kube-api-access-zkgdb\") pod \"package-server-manager-789f6589d5-qqfjh\" (UID: \"28014e20-2d95-42bc-881b-026137a7a016\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qqfjh" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.970997 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/94700472-065a-4371-8025-5c4e28b1e722-srv-cert\") pod \"catalog-operator-68c6474976-vnw2w\" (UID: \"94700472-065a-4371-8025-5c4e28b1e722\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vnw2w" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.971015 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcbnx\" (UniqueName: \"kubernetes.io/projected/f962a9b5-16d9-41e3-93a1-d7347e8cc27f-kube-api-access-zcbnx\") pod \"migrator-59844c95c7-c7t4g\" (UID: \"f962a9b5-16d9-41e3-93a1-d7347e8cc27f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c7t4g" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.971032 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdb55\" (UniqueName: \"kubernetes.io/projected/8edc3dbe-4856-4b87-b5d8-bbf2c55c0127-kube-api-access-fdb55\") pod \"multus-admission-controller-857f4d67dd-5wffk\" (UID: \"8edc3dbe-4856-4b87-b5d8-bbf2c55c0127\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5wffk" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.971066 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/94700472-065a-4371-8025-5c4e28b1e722-profile-collector-cert\") pod \"catalog-operator-68c6474976-vnw2w\" (UID: \"94700472-065a-4371-8025-5c4e28b1e722\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vnw2w" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.971128 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb10c1c4-bc22-46af-9f3f-c948d34e0017-config\") pod \"service-ca-operator-777779d784-xj696\" (UID: \"bb10c1c4-bc22-46af-9f3f-c948d34e0017\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xj696" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.972191 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90ab43dd-38e6-4264-a086-5cef08691214-service-ca-bundle\") pod \"router-default-5444994796-gsj4z\" (UID: \"90ab43dd-38e6-4264-a086-5cef08691214\") " pod="openshift-ingress/router-default-5444994796-gsj4z" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.972792 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/0560ae4e-4142-48c1-9829-2e770a3b56b4-socket-dir\") pod \"csi-hostpathplugin-qqrpb\" (UID: \"0560ae4e-4142-48c1-9829-2e770a3b56b4\") " pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.973096 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/0560ae4e-4142-48c1-9829-2e770a3b56b4-csi-data-dir\") pod \"csi-hostpathplugin-qqrpb\" (UID: \"0560ae4e-4142-48c1-9829-2e770a3b56b4\") " pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.975864 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0274f224-f29d-4c71-88e7-2111719f598b-tmpfs\") pod \"packageserver-d55dfcdfc-jq2zs\" (UID: \"0274f224-f29d-4c71-88e7-2111719f598b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.979336 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf43d927-8bd0-46df-9115-a95e6f747190-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-859lm\" (UID: \"cf43d927-8bd0-46df-9115-a95e6f747190\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-859lm" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.983766 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a8ef5a1d-9e3e-4057-89f5-672b0b1ca73c-config-volume\") pod \"dns-default-cdpgz\" (UID: \"a8ef5a1d-9e3e-4057-89f5-672b0b1ca73c\") " pod="openshift-dns/dns-default-cdpgz" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.985636 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8edc3dbe-4856-4b87-b5d8-bbf2c55c0127-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-5wffk\" (UID: \"8edc3dbe-4856-4b87-b5d8-bbf2c55c0127\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5wffk" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.985767 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/30bb01b5-2b19-403e-9930-a1991532e787-trusted-ca\") pod \"ingress-operator-5b745b69d9-twwmd\" (UID: \"30bb01b5-2b19-403e-9930-a1991532e787\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-twwmd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.989286 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8596ab2e-852c-4d94-a8ce-e02c823a76b1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sxcm9\" (UID: \"8596ab2e-852c-4d94-a8ce-e02c823a76b1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sxcm9" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.989603 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/167a9973-d839-4aaf-973b-2e00be41f59f-images\") pod \"machine-config-operator-74547568cd-pdnzp\" (UID: \"167a9973-d839-4aaf-973b-2e00be41f59f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pdnzp" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.989744 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/035f024b-dad2-4ad3-96bb-ebc3c52ff9d4-config\") pod \"kube-controller-manager-operator-78b949d7b-544q6\" (UID: \"035f024b-dad2-4ad3-96bb-ebc3c52ff9d4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-544q6" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.991205 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/30bb01b5-2b19-403e-9930-a1991532e787-metrics-tls\") pod \"ingress-operator-5b745b69d9-twwmd\" (UID: \"30bb01b5-2b19-403e-9930-a1991532e787\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-twwmd" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.992477 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/94700472-065a-4371-8025-5c4e28b1e722-profile-collector-cert\") pod \"catalog-operator-68c6474976-vnw2w\" (UID: \"94700472-065a-4371-8025-5c4e28b1e722\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vnw2w" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.992542 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/28014e20-2d95-42bc-881b-026137a7a016-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-qqfjh\" (UID: \"28014e20-2d95-42bc-881b-026137a7a016\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qqfjh" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.992780 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/035f024b-dad2-4ad3-96bb-ebc3c52ff9d4-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-544q6\" (UID: \"035f024b-dad2-4ad3-96bb-ebc3c52ff9d4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-544q6" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.992931 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf43d927-8bd0-46df-9115-a95e6f747190-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-859lm\" (UID: \"cf43d927-8bd0-46df-9115-a95e6f747190\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-859lm" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.995513 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/0560ae4e-4142-48c1-9829-2e770a3b56b4-registration-dir\") pod \"csi-hostpathplugin-qqrpb\" (UID: \"0560ae4e-4142-48c1-9829-2e770a3b56b4\") " pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.996798 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/167a9973-d839-4aaf-973b-2e00be41f59f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-pdnzp\" (UID: \"167a9973-d839-4aaf-973b-2e00be41f59f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pdnzp" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.998190 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fd0504b1-d681-4d1d-a9bb-8be90167f30f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-tfrtt\" (UID: \"fd0504b1-d681-4d1d-a9bb-8be90167f30f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tfrtt" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.998239 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/0560ae4e-4142-48c1-9829-2e770a3b56b4-plugins-dir\") pod \"csi-hostpathplugin-qqrpb\" (UID: \"0560ae4e-4142-48c1-9829-2e770a3b56b4\") " pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.998481 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/36384cbe-f2f7-48aa-8b54-ea07d8cc3678-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-gw45j\" (UID: \"36384cbe-f2f7-48aa-8b54-ea07d8cc3678\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gw45j" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.998566 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/167a9973-d839-4aaf-973b-2e00be41f59f-proxy-tls\") pod \"machine-config-operator-74547568cd-pdnzp\" (UID: \"167a9973-d839-4aaf-973b-2e00be41f59f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pdnzp" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.998823 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/90ab43dd-38e6-4264-a086-5cef08691214-stats-auth\") pod \"router-default-5444994796-gsj4z\" (UID: \"90ab43dd-38e6-4264-a086-5cef08691214\") " pod="openshift-ingress/router-default-5444994796-gsj4z" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.999212 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0274f224-f29d-4c71-88e7-2111719f598b-webhook-cert\") pod \"packageserver-d55dfcdfc-jq2zs\" (UID: \"0274f224-f29d-4c71-88e7-2111719f598b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs" Mar 18 20:10:04 crc kubenswrapper[4950]: I0318 20:10:04.999495 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/164efb7c-0b83-42ed-8d30-acc8a1ff1794-secret-volume\") pod \"collect-profiles-29564400-d9vqv\" (UID: \"164efb7c-0b83-42ed-8d30-acc8a1ff1794\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564400-d9vqv" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.003148 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/90ab43dd-38e6-4264-a086-5cef08691214-default-certificate\") pod \"router-default-5444994796-gsj4z\" (UID: \"90ab43dd-38e6-4264-a086-5cef08691214\") " pod="openshift-ingress/router-default-5444994796-gsj4z" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.008123 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb10c1c4-bc22-46af-9f3f-c948d34e0017-serving-cert\") pod \"service-ca-operator-777779d784-xj696\" (UID: \"bb10c1c4-bc22-46af-9f3f-c948d34e0017\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xj696" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.008516 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a8ef5a1d-9e3e-4057-89f5-672b0b1ca73c-metrics-tls\") pod \"dns-default-cdpgz\" (UID: \"a8ef5a1d-9e3e-4057-89f5-672b0b1ca73c\") " pod="openshift-dns/dns-default-cdpgz" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.009968 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f68ebf05-2b93-49f5-874a-351cf9c9d974-node-bootstrap-token\") pod \"machine-config-server-rwl8z\" (UID: \"f68ebf05-2b93-49f5-874a-351cf9c9d974\") " pod="openshift-machine-config-operator/machine-config-server-rwl8z" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.010987 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b3f2101e-4192-42db-9ff8-3dd152bc24bb-signing-key\") pod \"service-ca-9c57cc56f-ktwzg\" (UID: \"b3f2101e-4192-42db-9ff8-3dd152bc24bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-ktwzg" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.011866 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-pnffr" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.012282 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/90ab43dd-38e6-4264-a086-5cef08691214-metrics-certs\") pod \"router-default-5444994796-gsj4z\" (UID: \"90ab43dd-38e6-4264-a086-5cef08691214\") " pod="openshift-ingress/router-default-5444994796-gsj4z" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.012888 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/164efb7c-0b83-42ed-8d30-acc8a1ff1794-config-volume\") pod \"collect-profiles-29564400-d9vqv\" (UID: \"164efb7c-0b83-42ed-8d30-acc8a1ff1794\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564400-d9vqv" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.013753 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-47kcd" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.014093 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b3f2101e-4192-42db-9ff8-3dd152bc24bb-signing-cabundle\") pod \"service-ca-9c57cc56f-ktwzg\" (UID: \"b3f2101e-4192-42db-9ff8-3dd152bc24bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-ktwzg" Mar 18 20:10:05 crc kubenswrapper[4950]: E0318 20:10:05.014447 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:05.514402211 +0000 UTC m=+218.755244079 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.014491 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/0560ae4e-4142-48c1-9829-2e770a3b56b4-mountpoint-dir\") pod \"csi-hostpathplugin-qqrpb\" (UID: \"0560ae4e-4142-48c1-9829-2e770a3b56b4\") " pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.016250 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0274f224-f29d-4c71-88e7-2111719f598b-apiservice-cert\") pod \"packageserver-d55dfcdfc-jq2zs\" (UID: \"0274f224-f29d-4c71-88e7-2111719f598b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.017636 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fd0504b1-d681-4d1d-a9bb-8be90167f30f-srv-cert\") pod \"olm-operator-6b444d44fb-tfrtt\" (UID: \"fd0504b1-d681-4d1d-a9bb-8be90167f30f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tfrtt" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.018833 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a2456057-5ff6-458f-af94-53edca249b7f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8nhwt\" (UID: \"a2456057-5ff6-458f-af94-53edca249b7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.019483 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a2456057-5ff6-458f-af94-53edca249b7f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8nhwt\" (UID: \"a2456057-5ff6-458f-af94-53edca249b7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.020518 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/94700472-065a-4371-8025-5c4e28b1e722-srv-cert\") pod \"catalog-operator-68c6474976-vnw2w\" (UID: \"94700472-065a-4371-8025-5c4e28b1e722\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vnw2w" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.026265 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7fcd663f-3b4c-42fe-a585-c1e1142875ca-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-79fhl\" (UID: \"7fcd663f-3b4c-42fe-a585-c1e1142875ca\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-79fhl" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.029850 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwghk\" (UniqueName: \"kubernetes.io/projected/1fedd09d-87e7-4bfc-ab58-589189c984d8-kube-api-access-kwghk\") pod \"openshift-controller-manager-operator-756b6f6bc6-9pgcg\" (UID: \"1fedd09d-87e7-4bfc-ab58-589189c984d8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9pgcg" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.030604 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f68ebf05-2b93-49f5-874a-351cf9c9d974-certs\") pod \"machine-config-server-rwl8z\" (UID: \"f68ebf05-2b93-49f5-874a-351cf9c9d974\") " pod="openshift-machine-config-operator/machine-config-server-rwl8z" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.030730 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/36384cbe-f2f7-48aa-8b54-ea07d8cc3678-proxy-tls\") pod \"machine-config-controller-84d6567774-gw45j\" (UID: \"36384cbe-f2f7-48aa-8b54-ea07d8cc3678\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gw45j" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.031135 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-79fhl" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.032309 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5d345627-baa3-4354-a422-1b3646e5cc48-cert\") pod \"ingress-canary-kbjxd\" (UID: \"5d345627-baa3-4354-a422-1b3646e5cc48\") " pod="openshift-ingress-canary/ingress-canary-kbjxd" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.037867 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/28e264f3-49f6-406a-b700-213e196574c1-bound-sa-token\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.046005 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qj7bk"] Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.063253 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qscws\" (UniqueName: \"kubernetes.io/projected/0560ae4e-4142-48c1-9829-2e770a3b56b4-kube-api-access-qscws\") pod \"csi-hostpathplugin-qqrpb\" (UID: \"0560ae4e-4142-48c1-9829-2e770a3b56b4\") " pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.073108 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:05 crc kubenswrapper[4950]: E0318 20:10:05.073670 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:05.573652048 +0000 UTC m=+218.814493916 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.079687 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c56j5\" (UniqueName: \"kubernetes.io/projected/cf43d927-8bd0-46df-9115-a95e6f747190-kube-api-access-c56j5\") pod \"kube-storage-version-migrator-operator-b67b599dd-859lm\" (UID: \"cf43d927-8bd0-46df-9115-a95e6f747190\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-859lm" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.100253 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-rf8k5"] Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.105115 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64khz\" (UniqueName: \"kubernetes.io/projected/36384cbe-f2f7-48aa-8b54-ea07d8cc3678-kube-api-access-64khz\") pod \"machine-config-controller-84d6567774-gw45j\" (UID: \"36384cbe-f2f7-48aa-8b54-ea07d8cc3678\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gw45j" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.122281 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd2f8\" (UniqueName: \"kubernetes.io/projected/30bb01b5-2b19-403e-9930-a1991532e787-kube-api-access-wd2f8\") pod \"ingress-operator-5b745b69d9-twwmd\" (UID: \"30bb01b5-2b19-403e-9930-a1991532e787\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-twwmd" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.143808 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn5w6\" (UniqueName: \"kubernetes.io/projected/fd0504b1-d681-4d1d-a9bb-8be90167f30f-kube-api-access-zn5w6\") pod \"olm-operator-6b444d44fb-tfrtt\" (UID: \"fd0504b1-d681-4d1d-a9bb-8be90167f30f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tfrtt" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.161736 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9pgcg" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.162062 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb5hp\" (UniqueName: \"kubernetes.io/projected/94700472-065a-4371-8025-5c4e28b1e722-kube-api-access-gb5hp\") pod \"catalog-operator-68c6474976-vnw2w\" (UID: \"94700472-065a-4371-8025-5c4e28b1e722\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vnw2w" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.175712 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:05 crc kubenswrapper[4950]: E0318 20:10:05.176158 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:05.676141043 +0000 UTC m=+218.916982911 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.179325 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2mqb" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.183958 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vz67\" (UniqueName: \"kubernetes.io/projected/f68ebf05-2b93-49f5-874a-351cf9c9d974-kube-api-access-5vz67\") pod \"machine-config-server-rwl8z\" (UID: \"f68ebf05-2b93-49f5-874a-351cf9c9d974\") " pod="openshift-machine-config-operator/machine-config-server-rwl8z" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.217594 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-rwl8z" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.231387 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.240907 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjf7l\" (UniqueName: \"kubernetes.io/projected/627e6f3d-cde5-4849-bd02-6437622f4978-kube-api-access-cjf7l\") pod \"auto-csr-approver-29564410-9hqrd\" (UID: \"627e6f3d-cde5-4849-bd02-6437622f4978\") " pod="openshift-infra/auto-csr-approver-29564410-9hqrd" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.251998 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkgdb\" (UniqueName: \"kubernetes.io/projected/28014e20-2d95-42bc-881b-026137a7a016-kube-api-access-zkgdb\") pod \"package-server-manager-789f6589d5-qqfjh\" (UID: \"28014e20-2d95-42bc-881b-026137a7a016\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qqfjh" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.254267 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncvzb\" (UniqueName: \"kubernetes.io/projected/5d345627-baa3-4354-a422-1b3646e5cc48-kube-api-access-ncvzb\") pod \"ingress-canary-kbjxd\" (UID: \"5d345627-baa3-4354-a422-1b3646e5cc48\") " pod="openshift-ingress-canary/ingress-canary-kbjxd" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.256889 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcbnx\" (UniqueName: \"kubernetes.io/projected/f962a9b5-16d9-41e3-93a1-d7347e8cc27f-kube-api-access-zcbnx\") pod \"migrator-59844c95c7-c7t4g\" (UID: \"f962a9b5-16d9-41e3-93a1-d7347e8cc27f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c7t4g" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.276589 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:05 crc kubenswrapper[4950]: E0318 20:10:05.277576 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:05.777552861 +0000 UTC m=+219.018394729 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.280638 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdb55\" (UniqueName: \"kubernetes.io/projected/8edc3dbe-4856-4b87-b5d8-bbf2c55c0127-kube-api-access-fdb55\") pod \"multus-admission-controller-857f4d67dd-5wffk\" (UID: \"8edc3dbe-4856-4b87-b5d8-bbf2c55c0127\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5wffk" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.298076 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwwjp\" (UniqueName: \"kubernetes.io/projected/0274f224-f29d-4c71-88e7-2111719f598b-kube-api-access-lwwjp\") pod \"packageserver-d55dfcdfc-jq2zs\" (UID: \"0274f224-f29d-4c71-88e7-2111719f598b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.330550 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t4jq\" (UniqueName: \"kubernetes.io/projected/167a9973-d839-4aaf-973b-2e00be41f59f-kube-api-access-4t4jq\") pod \"machine-config-operator-74547568cd-pdnzp\" (UID: \"167a9973-d839-4aaf-973b-2e00be41f59f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pdnzp" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.342893 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/30bb01b5-2b19-403e-9930-a1991532e787-bound-sa-token\") pod \"ingress-operator-5b745b69d9-twwmd\" (UID: \"30bb01b5-2b19-403e-9930-a1991532e787\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-twwmd" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.347724 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-twwmd" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.355581 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gw45j" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.362362 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h96q2\" (UniqueName: \"kubernetes.io/projected/a8ef5a1d-9e3e-4057-89f5-672b0b1ca73c-kube-api-access-h96q2\") pod \"dns-default-cdpgz\" (UID: \"a8ef5a1d-9e3e-4057-89f5-672b0b1ca73c\") " pod="openshift-dns/dns-default-cdpgz" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.370825 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pdnzp" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.376539 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-859lm" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.378913 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:05 crc kubenswrapper[4950]: E0318 20:10:05.379630 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:05.879613856 +0000 UTC m=+219.120455724 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.390083 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f75b5\" (UniqueName: \"kubernetes.io/projected/90ab43dd-38e6-4264-a086-5cef08691214-kube-api-access-f75b5\") pod \"router-default-5444994796-gsj4z\" (UID: \"90ab43dd-38e6-4264-a086-5cef08691214\") " pod="openshift-ingress/router-default-5444994796-gsj4z" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.397282 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c7t4g" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.409157 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-46hjn"] Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.415138 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tfrtt" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.419477 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qqfjh" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.419977 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-5wffk" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.426888 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vnw2w" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.439195 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sgrs\" (UniqueName: \"kubernetes.io/projected/a2456057-5ff6-458f-af94-53edca249b7f-kube-api-access-2sgrs\") pod \"marketplace-operator-79b997595-8nhwt\" (UID: \"a2456057-5ff6-458f-af94-53edca249b7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.439582 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c4x9\" (UniqueName: \"kubernetes.io/projected/b3f2101e-4192-42db-9ff8-3dd152bc24bb-kube-api-access-5c4x9\") pod \"service-ca-9c57cc56f-ktwzg\" (UID: \"b3f2101e-4192-42db-9ff8-3dd152bc24bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-ktwzg" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.442712 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2lpl\" (UniqueName: \"kubernetes.io/projected/8596ab2e-852c-4d94-a8ce-e02c823a76b1-kube-api-access-v2lpl\") pod \"control-plane-machine-set-operator-78cbb6b69f-sxcm9\" (UID: \"8596ab2e-852c-4d94-a8ce-e02c823a76b1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sxcm9" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.458901 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564410-9hqrd" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.465566 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjg8b\" (UniqueName: \"kubernetes.io/projected/97cb884c-5f55-4ca0-8639-193c2f1df139-kube-api-access-sjg8b\") pod \"auto-csr-approver-29564408-2658d\" (UID: \"97cb884c-5f55-4ca0-8639-193c2f1df139\") " pod="openshift-infra/auto-csr-approver-29564408-2658d" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.479720 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:05 crc kubenswrapper[4950]: E0318 20:10:05.480075 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:05.980060398 +0000 UTC m=+219.220902256 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.480471 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-ktwzg" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.481818 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g6m2\" (UniqueName: \"kubernetes.io/projected/bb10c1c4-bc22-46af-9f3f-c948d34e0017-kube-api-access-5g6m2\") pod \"service-ca-operator-777779d784-xj696\" (UID: \"bb10c1c4-bc22-46af-9f3f-c948d34e0017\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xj696" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.497653 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-kbjxd" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.500923 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.511401 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/035f024b-dad2-4ad3-96bb-ebc3c52ff9d4-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-544q6\" (UID: \"035f024b-dad2-4ad3-96bb-ebc3c52ff9d4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-544q6" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.535231 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rprg\" (UniqueName: \"kubernetes.io/projected/164efb7c-0b83-42ed-8d30-acc8a1ff1794-kube-api-access-2rprg\") pod \"collect-profiles-29564400-d9vqv\" (UID: \"164efb7c-0b83-42ed-8d30-acc8a1ff1794\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564400-d9vqv" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.550969 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-cdpgz" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.585486 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:05 crc kubenswrapper[4950]: E0318 20:10:05.585916 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:06.085871361 +0000 UTC m=+219.326713229 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.659958 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-544q6" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.679697 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-gsj4z" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.685876 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sxcm9" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.686530 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:05 crc kubenswrapper[4950]: E0318 20:10:05.686776 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:06.186754325 +0000 UTC m=+219.427596193 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:05 crc kubenswrapper[4950]: E0318 20:10:05.688242 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:06.188225893 +0000 UTC m=+219.429067761 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.687782 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.698576 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.747199 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564408-2658d" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.752754 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xj696" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.786354 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564400-d9vqv" Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.789143 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-47kcd"] Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.791152 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:05 crc kubenswrapper[4950]: E0318 20:10:05.791302 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:06.291280964 +0000 UTC m=+219.532122832 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.791619 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:05 crc kubenswrapper[4950]: E0318 20:10:05.792036 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:06.292017303 +0000 UTC m=+219.532859171 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.852156 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-79fhl"] Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.852228 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-pnffr"] Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.894468 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:05 crc kubenswrapper[4950]: E0318 20:10:05.895210 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:06.395160296 +0000 UTC m=+219.636002164 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.932474 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-s8cr9" event={"ID":"792583b6-449b-4b55-b138-cf51971e2a17","Type":"ContainerStarted","Data":"96201c2f5830d3706fd4ca0ea88de47dbb8af44b177bec23fefdbe72132c6c2b"} Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.962568 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d5rvd" event={"ID":"cc0b2145-4578-4709-837d-6bae20399714","Type":"ContainerStarted","Data":"6192bf5e88d905d54d960df9c7cfc0d79204545732cfab343e432defb6a98b14"} Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.968242 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m7qjx" event={"ID":"27a9fc13-1b12-40f1-9de4-fc04f3810df9","Type":"ContainerStarted","Data":"7e19a198abca3f529ae320607e8b1b7c3a5db04e3185b4ca01b44fb13cc2b820"} Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.990054 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p9pk8" event={"ID":"cfd25023-4dd6-4527-bcde-836258333c37","Type":"ContainerStarted","Data":"b2847f54c8f5343ce415fc1750aa03283593a71ec1528f7ad786a4bdb3b16edd"} Mar 18 20:10:05 crc kubenswrapper[4950]: I0318 20:10:05.998825 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:05 crc kubenswrapper[4950]: E0318 20:10:05.999201 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:06.499188982 +0000 UTC m=+219.740030850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.006259 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qj7bk" event={"ID":"ce6591e6-fb40-40a0-91a7-9aa2b3aa0437","Type":"ContainerStarted","Data":"aab7bf7eaf8d2621b849feb2067e92bed38a73bcd42bb8126d91d5b6a5e40de2"} Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.023880 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" event={"ID":"70601ef2-ab17-4550-890b-b29bf1a7b85e","Type":"ContainerStarted","Data":"6a6ac9f276afaa5327c38326539da8d4956e49fcf92a25f8ca48d1dfced47c93"} Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.023938 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" event={"ID":"70601ef2-ab17-4550-890b-b29bf1a7b85e","Type":"ContainerStarted","Data":"c40fe5f787730ada42f53a058b4a24f96d9949f2e312bb23c66a12e3063b094d"} Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.024387 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.046293 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-g2f49" event={"ID":"4684cfd6-1836-4218-9c6e-c6ca75a2da96","Type":"ContainerStarted","Data":"63cdf6e75758234b82cd4ef7d563c503d87b4b769d03d5beff019e06d6346e55"} Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.046337 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-g2f49" event={"ID":"4684cfd6-1836-4218-9c6e-c6ca75a2da96","Type":"ContainerStarted","Data":"6c421fd7ff8d53d0c82d315c646b14b6ad14b33c1d5ef10c5d566647522b6b91"} Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.053829 4950 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-9l9mq container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.053896 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" podUID="70601ef2-ab17-4550-890b-b29bf1a7b85e" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.054440 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-rf8k5" event={"ID":"290d1b89-d5c9-4928-8973-6c33f476b091","Type":"ContainerStarted","Data":"505bc9a431cca4d5991d7a0956f9f3643f20a4e0daa2659a7e5476456c857ee9"} Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.056981 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-kjb6z" event={"ID":"bfd29e9b-fbda-41d0-8300-a0d9337e0e6d","Type":"ContainerStarted","Data":"b788557af672cb409b0f4eccd0adc2dd2c087fa122fc2dc0753264693081fd25"} Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.057008 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-kjb6z" event={"ID":"bfd29e9b-fbda-41d0-8300-a0d9337e0e6d","Type":"ContainerStarted","Data":"08ae8368669f1faa7fa40cef5edeafe66b8d3802037131effb147ec40e24f32d"} Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.060334 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-46hjn" event={"ID":"1cec2292-f508-42b2-916f-9a5808045626","Type":"ContainerStarted","Data":"b44d263b6e57afdd082bb8aef49f2502668371c453f03cf85053218b0197ca42"} Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.061724 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8jvmc" event={"ID":"dc9fd238-6c7a-46db-bb03-3782b9cfc481","Type":"ContainerStarted","Data":"1357e71648f68d3206ddb9ab27598fb1dec3e1a7cc72b898b7332f6625339179"} Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.061767 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8jvmc" event={"ID":"dc9fd238-6c7a-46db-bb03-3782b9cfc481","Type":"ContainerStarted","Data":"c5874587ec6fe25c5e8220cb8f38cf8e4d818b81ad68b7ccb6a0dbd550417e58"} Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.065066 4950 generic.go:334] "Generic (PLEG): container finished" podID="be93cdca-9fe6-46a2-ae6a-94516d8eaa50" containerID="831930f72df5aec61b2e5c0b5cb2ffb1d29112a1935f72b8ed008482ce8f790c" exitCode=0 Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.065103 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" event={"ID":"be93cdca-9fe6-46a2-ae6a-94516d8eaa50","Type":"ContainerDied","Data":"831930f72df5aec61b2e5c0b5cb2ffb1d29112a1935f72b8ed008482ce8f790c"} Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.071444 4950 generic.go:334] "Generic (PLEG): container finished" podID="ff73cd60-bdf9-477a-8ede-1e781aaffd22" containerID="e20f8a81cd845a4c1970b4b70d1294864892cd74777f73900a2435daaaa8b28e" exitCode=0 Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.071517 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-27jdq" event={"ID":"ff73cd60-bdf9-477a-8ede-1e781aaffd22","Type":"ContainerDied","Data":"e20f8a81cd845a4c1970b4b70d1294864892cd74777f73900a2435daaaa8b28e"} Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.077975 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" event={"ID":"a0e597f4-b43a-47db-9f22-a38272060004","Type":"ContainerStarted","Data":"9319cf28b6a29f82fc9c9f97d9bd49eb6cea689d474de0971b094d11aab3c153"} Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.078004 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" event={"ID":"a0e597f4-b43a-47db-9f22-a38272060004","Type":"ContainerStarted","Data":"0cf19fb082ef62fb43e346aa04417a8b6176e8a669c3039920e8a2f2d58d6852"} Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.078571 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.094210 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" Mar 18 20:10:06 crc kubenswrapper[4950]: W0318 20:10:06.095185 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34f027ae_fb71_460c_b978_e254fac5d50f.slice/crio-9c4e6fe2ff0dd06200fa0f74779a4895d8d020cfe7df29dd34a4559498d8c064 WatchSource:0}: Error finding container 9c4e6fe2ff0dd06200fa0f74779a4895d8d020cfe7df29dd34a4559498d8c064: Status 404 returned error can't find the container with id 9c4e6fe2ff0dd06200fa0f74779a4895d8d020cfe7df29dd34a4559498d8c064 Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.101747 4950 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-wv7kr container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.101806 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" podUID="a0e597f4-b43a-47db-9f22-a38272060004" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.102598 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:06 crc kubenswrapper[4950]: E0318 20:10:06.105107 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:06.605078916 +0000 UTC m=+219.845920784 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:06 crc kubenswrapper[4950]: W0318 20:10:06.110875 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2c83122_863b_41e6_9118_54fa716c8485.slice/crio-841ba1b9bdf4a7651ee1508f39afb8aa0a46adebb0faccdb0b25016c713e519b WatchSource:0}: Error finding container 841ba1b9bdf4a7651ee1508f39afb8aa0a46adebb0faccdb0b25016c713e519b: Status 404 returned error can't find the container with id 841ba1b9bdf4a7651ee1508f39afb8aa0a46adebb0faccdb0b25016c713e519b Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.150236 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-pdnzp"] Mar 18 20:10:06 crc kubenswrapper[4950]: W0318 20:10:06.188190 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fcd663f_3b4c_42fe_a585_c1e1142875ca.slice/crio-4aa84cfddf020b1c415e978590ee4d48908154d0100151c278812f6a3fc2b02a WatchSource:0}: Error finding container 4aa84cfddf020b1c415e978590ee4d48908154d0100151c278812f6a3fc2b02a: Status 404 returned error can't find the container with id 4aa84cfddf020b1c415e978590ee4d48908154d0100151c278812f6a3fc2b02a Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.209621 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:06 crc kubenswrapper[4950]: E0318 20:10:06.223472 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:06.723457217 +0000 UTC m=+219.964299085 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.310526 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:06 crc kubenswrapper[4950]: E0318 20:10:06.311116 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:06.811098835 +0000 UTC m=+220.051940703 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.341642 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-qqrpb"] Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.411855 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:06 crc kubenswrapper[4950]: E0318 20:10:06.412219 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:06.912203365 +0000 UTC m=+220.153045233 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.419505 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2mqb"] Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.498180 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" podStartSLOduration=171.498157569 podStartE2EDuration="2m51.498157569s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:06.477839018 +0000 UTC m=+219.718680886" watchObservedRunningTime="2026-03-18 20:10:06.498157569 +0000 UTC m=+219.738999437" Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.513341 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:06 crc kubenswrapper[4950]: E0318 20:10:06.513808 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:07.013791616 +0000 UTC m=+220.254633484 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.626064 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:06 crc kubenswrapper[4950]: E0318 20:10:06.626350 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:07.126339734 +0000 UTC m=+220.367181602 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.682026 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs"] Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.709994 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" podStartSLOduration=171.709975058 podStartE2EDuration="2m51.709975058s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:06.708825118 +0000 UTC m=+219.949666986" watchObservedRunningTime="2026-03-18 20:10:06.709975058 +0000 UTC m=+219.950816926" Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.726734 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:06 crc kubenswrapper[4950]: E0318 20:10:06.726990 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:07.226963391 +0000 UTC m=+220.467805259 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.727165 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:06 crc kubenswrapper[4950]: E0318 20:10:06.727634 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:07.227625049 +0000 UTC m=+220.468466927 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.754159 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" podStartSLOduration=171.754147131 podStartE2EDuration="2m51.754147131s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:06.752883408 +0000 UTC m=+219.993725276" watchObservedRunningTime="2026-03-18 20:10:06.754147131 +0000 UTC m=+219.994988999" Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.799087 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564410-9hqrd"] Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.820761 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9pgcg"] Mar 18 20:10:06 crc kubenswrapper[4950]: W0318 20:10:06.826686 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0274f224_f29d_4c71_88e7_2111719f598b.slice/crio-2a7b54aebf6ae1e1abc0e91bd168596ca20401979df7d76ca077631c90b4d982 WatchSource:0}: Error finding container 2a7b54aebf6ae1e1abc0e91bd168596ca20401979df7d76ca077631c90b4d982: Status 404 returned error can't find the container with id 2a7b54aebf6ae1e1abc0e91bd168596ca20401979df7d76ca077631c90b4d982 Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.832148 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:06 crc kubenswrapper[4950]: E0318 20:10:06.832502 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:07.332482976 +0000 UTC m=+220.573324844 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.849158 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-p9pk8" podStartSLOduration=171.849142981 podStartE2EDuration="2m51.849142981s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:06.847725714 +0000 UTC m=+220.088567582" watchObservedRunningTime="2026-03-18 20:10:06.849142981 +0000 UTC m=+220.089984849" Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.892400 4950 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.933994 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:06 crc kubenswrapper[4950]: E0318 20:10:06.934720 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:07.434708495 +0000 UTC m=+220.675550363 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:06 crc kubenswrapper[4950]: I0318 20:10:06.935304 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-d5rvd" podStartSLOduration=171.93529426 podStartE2EDuration="2m51.93529426s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:06.918018199 +0000 UTC m=+220.158860067" watchObservedRunningTime="2026-03-18 20:10:06.93529426 +0000 UTC m=+220.176136128" Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.041972 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:07 crc kubenswrapper[4950]: E0318 20:10:07.042537 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:07.54252097 +0000 UTC m=+220.783362838 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.091890 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m7qjx" podStartSLOduration=172.091871829 podStartE2EDuration="2m52.091871829s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:07.087660299 +0000 UTC m=+220.328502167" watchObservedRunningTime="2026-03-18 20:10:07.091871829 +0000 UTC m=+220.332713697" Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.115491 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-twwmd"] Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.115746 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qqfjh"] Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.129951 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-859lm"] Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.155827 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:07 crc kubenswrapper[4950]: E0318 20:10:07.156191 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:07.656175697 +0000 UTC m=+220.897017565 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.242948 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8jvmc" podStartSLOduration=172.242935422 podStartE2EDuration="2m52.242935422s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:07.241129235 +0000 UTC m=+220.481971103" watchObservedRunningTime="2026-03-18 20:10:07.242935422 +0000 UTC m=+220.483777290" Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.246740 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-ktwzg"] Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.250831 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-c7t4g"] Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.256538 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:07 crc kubenswrapper[4950]: E0318 20:10:07.256861 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:07.756844446 +0000 UTC m=+220.997686314 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.265228 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-5wffk"] Mar 18 20:10:07 crc kubenswrapper[4950]: W0318 20:10:07.280308 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30bb01b5_2b19_403e_9930_a1991532e787.slice/crio-42e40d4983a5b38d5a17a66850830ad18cef5f02e922ea898e74e2d649b997a5 WatchSource:0}: Error finding container 42e40d4983a5b38d5a17a66850830ad18cef5f02e922ea898e74e2d649b997a5: Status 404 returned error can't find the container with id 42e40d4983a5b38d5a17a66850830ad18cef5f02e922ea898e74e2d649b997a5 Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.293530 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-46hjn" event={"ID":"1cec2292-f508-42b2-916f-9a5808045626","Type":"ContainerStarted","Data":"60eb1f53d56060d604da66326a97602acb70000d70709e4e1a6cc167a6f57a65"} Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.311872 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tfrtt"] Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.361757 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.362842 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-kjb6z" event={"ID":"bfd29e9b-fbda-41d0-8300-a0d9337e0e6d","Type":"ContainerStarted","Data":"0662fe8f6d98da19ebca47690fc52fb266f29d14f1ed9e5a405d0051f59ce559"} Mar 18 20:10:07 crc kubenswrapper[4950]: E0318 20:10:07.363065 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:07.863051628 +0000 UTC m=+221.103893496 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.377648 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vnw2w"] Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.419872 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-47kcd" event={"ID":"34f027ae-fb71-460c-b978-e254fac5d50f","Type":"ContainerStarted","Data":"9c4e6fe2ff0dd06200fa0f74779a4895d8d020cfe7df29dd34a4559498d8c064"} Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.463558 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-rf8k5" event={"ID":"290d1b89-d5c9-4928-8973-6c33f476b091","Type":"ContainerStarted","Data":"4fb386f1a5dd049fe98912e8fed43881826275f68539fc48343e9c5a03c402e2"} Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.466033 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-rf8k5" Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.466098 4950 patch_prober.go:28] interesting pod/downloads-7954f5f757-rf8k5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.466138 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rf8k5" podUID="290d1b89-d5c9-4928-8973-6c33f476b091" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.469513 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:07 crc kubenswrapper[4950]: E0318 20:10:07.469815 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:07.969800285 +0000 UTC m=+221.210642153 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.476076 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564408-2658d"] Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.492682 4950 ???:1] "http: TLS handshake error from 192.168.126.11:39722: no serving certificate available for the kubelet" Mar 18 20:10:07 crc kubenswrapper[4950]: W0318 20:10:07.528935 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf43d927_8bd0_46df_9115_a95e6f747190.slice/crio-a1ea86aa2c97dfe2082ce045d35d4ce535d53134e86f11f74a23243f85424e32 WatchSource:0}: Error finding container a1ea86aa2c97dfe2082ce045d35d4ce535d53134e86f11f74a23243f85424e32: Status 404 returned error can't find the container with id a1ea86aa2c97dfe2082ce045d35d4ce535d53134e86f11f74a23243f85424e32 Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.534591 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-79fhl" event={"ID":"7fcd663f-3b4c-42fe-a585-c1e1142875ca","Type":"ContainerStarted","Data":"4aa84cfddf020b1c415e978590ee4d48908154d0100151c278812f6a3fc2b02a"} Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.567210 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-s8cr9" event={"ID":"792583b6-449b-4b55-b138-cf51971e2a17","Type":"ContainerStarted","Data":"a9a49056e58527619a0971e1c8d2b39a1338d0e832dcce187ecee719ac02d13e"} Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.572551 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:07 crc kubenswrapper[4950]: E0318 20:10:07.574445 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:08.074432877 +0000 UTC m=+221.315274745 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.576560 4950 ???:1] "http: TLS handshake error from 192.168.126.11:39730: no serving certificate available for the kubelet" Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.582494 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sxcm9"] Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.597768 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-544q6"] Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.644773 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-rf8k5" podStartSLOduration=172.644757753 podStartE2EDuration="2m52.644757753s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:07.643162842 +0000 UTC m=+220.884004710" watchObservedRunningTime="2026-03-18 20:10:07.644757753 +0000 UTC m=+220.885599621" Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.672324 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-s8cr9" podStartSLOduration=172.672293632 podStartE2EDuration="2m52.672293632s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:07.671390249 +0000 UTC m=+220.912232117" watchObservedRunningTime="2026-03-18 20:10:07.672293632 +0000 UTC m=+220.913135490" Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.681659 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:07 crc kubenswrapper[4950]: E0318 20:10:07.682086 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:08.182071127 +0000 UTC m=+221.422912995 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.688835 4950 ???:1] "http: TLS handshake error from 192.168.126.11:39732: no serving certificate available for the kubelet" Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.702056 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8nhwt"] Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.749015 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-46hjn" podStartSLOduration=172.748996035 podStartE2EDuration="2m52.748996035s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:07.721262641 +0000 UTC m=+220.962104509" watchObservedRunningTime="2026-03-18 20:10:07.748996035 +0000 UTC m=+220.989837903" Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.762814 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-kjb6z" podStartSLOduration=172.761258005 podStartE2EDuration="2m52.761258005s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:07.756324876 +0000 UTC m=+220.997166744" watchObservedRunningTime="2026-03-18 20:10:07.761258005 +0000 UTC m=+221.002099873" Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.763341 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-kbjxd"] Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.794461 4950 ???:1] "http: TLS handshake error from 192.168.126.11:39734: no serving certificate available for the kubelet" Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.794705 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xj696"] Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.795001 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:07 crc kubenswrapper[4950]: E0318 20:10:07.795439 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:08.295407826 +0000 UTC m=+221.536249694 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.808734 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564410-9hqrd" event={"ID":"627e6f3d-cde5-4849-bd02-6437622f4978","Type":"ContainerStarted","Data":"586f0b3af3e8f89f832874751a5f050777edfe372bea7a93c8e4a7e2a5768640"} Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.830267 4950 generic.go:334] "Generic (PLEG): container finished" podID="4684cfd6-1836-4218-9c6e-c6ca75a2da96" containerID="63cdf6e75758234b82cd4ef7d563c503d87b4b769d03d5beff019e06d6346e55" exitCode=0 Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.830332 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-g2f49" event={"ID":"4684cfd6-1836-4218-9c6e-c6ca75a2da96","Type":"ContainerDied","Data":"63cdf6e75758234b82cd4ef7d563c503d87b4b769d03d5beff019e06d6346e55"} Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.873608 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9pgcg" event={"ID":"1fedd09d-87e7-4bfc-ab58-589189c984d8","Type":"ContainerStarted","Data":"a8a20d84f03545a64870ee7dd75ecfe17b14f5af00c2b0943be5188ebe7600cb"} Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.900738 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:07 crc kubenswrapper[4950]: E0318 20:10:07.901202 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:08.401183978 +0000 UTC m=+221.642025836 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.901262 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:07 crc kubenswrapper[4950]: E0318 20:10:07.901959 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:08.401940748 +0000 UTC m=+221.642782616 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.918173 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs" event={"ID":"0274f224-f29d-4c71-88e7-2111719f598b","Type":"ContainerStarted","Data":"2a7b54aebf6ae1e1abc0e91bd168596ca20401979df7d76ca077631c90b4d982"} Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.923832 4950 ???:1] "http: TLS handshake error from 192.168.126.11:39738: no serving certificate available for the kubelet" Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.925144 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564400-d9vqv"] Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.939049 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-gw45j"] Mar 18 20:10:07 crc kubenswrapper[4950]: I0318 20:10:07.989398 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pdnzp" event={"ID":"167a9973-d839-4aaf-973b-2e00be41f59f","Type":"ContainerStarted","Data":"28e7e16601be75e55a5c9d47a90a40042285c996678e303acf982a0ee3911485"} Mar 18 20:10:08 crc kubenswrapper[4950]: I0318 20:10:08.007943 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:08 crc kubenswrapper[4950]: E0318 20:10:08.008373 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:08.508343026 +0000 UTC m=+221.749184894 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:08 crc kubenswrapper[4950]: I0318 20:10:08.016701 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-cdpgz"] Mar 18 20:10:08 crc kubenswrapper[4950]: I0318 20:10:08.034824 4950 ???:1] "http: TLS handshake error from 192.168.126.11:39750: no serving certificate available for the kubelet" Mar 18 20:10:08 crc kubenswrapper[4950]: I0318 20:10:08.088684 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-rwl8z" event={"ID":"f68ebf05-2b93-49f5-874a-351cf9c9d974","Type":"ContainerStarted","Data":"b9c387e1d87ad1a56b052c663fa5e0b030d3bedc0a105108db2b69e8be96c954"} Mar 18 20:10:08 crc kubenswrapper[4950]: I0318 20:10:08.088726 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-rwl8z" event={"ID":"f68ebf05-2b93-49f5-874a-351cf9c9d974","Type":"ContainerStarted","Data":"0a2842123271cfab5c33c068e7788f873aa2717b86f430de810d19695528a9b1"} Mar 18 20:10:08 crc kubenswrapper[4950]: I0318 20:10:08.109268 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:08 crc kubenswrapper[4950]: E0318 20:10:08.109619 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:08.60960914 +0000 UTC m=+221.850451008 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:08 crc kubenswrapper[4950]: I0318 20:10:08.117561 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-pnffr" event={"ID":"d2c83122-863b-41e6-9118-54fa716c8485","Type":"ContainerStarted","Data":"841ba1b9bdf4a7651ee1508f39afb8aa0a46adebb0faccdb0b25016c713e519b"} Mar 18 20:10:08 crc kubenswrapper[4950]: I0318 20:10:08.143325 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2mqb" event={"ID":"fc3e8cef-c459-47ef-9531-da9acd77054e","Type":"ContainerStarted","Data":"e21258736e636986f125d41e821c28f33ffcfcef825c2f641eb85d22dceb3b30"} Mar 18 20:10:08 crc kubenswrapper[4950]: I0318 20:10:08.174793 4950 ???:1] "http: TLS handshake error from 192.168.126.11:39766: no serving certificate available for the kubelet" Mar 18 20:10:08 crc kubenswrapper[4950]: I0318 20:10:08.223198 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:08 crc kubenswrapper[4950]: E0318 20:10:08.225136 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:08.725046213 +0000 UTC m=+221.965888081 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:08 crc kubenswrapper[4950]: I0318 20:10:08.272330 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-gsj4z" event={"ID":"90ab43dd-38e6-4264-a086-5cef08691214","Type":"ContainerStarted","Data":"be899fcb162213c07be260191b4868a35f7910a2409283b49bf980891c840027"} Mar 18 20:10:08 crc kubenswrapper[4950]: I0318 20:10:08.284115 4950 ???:1] "http: TLS handshake error from 192.168.126.11:39770: no serving certificate available for the kubelet" Mar 18 20:10:08 crc kubenswrapper[4950]: I0318 20:10:08.324979 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:08 crc kubenswrapper[4950]: E0318 20:10:08.326211 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:08.826198694 +0000 UTC m=+222.067040562 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:08 crc kubenswrapper[4950]: I0318 20:10:08.358217 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" event={"ID":"0560ae4e-4142-48c1-9829-2e770a3b56b4","Type":"ContainerStarted","Data":"d1334d8421f00ad1013f9d8bc9e76fb01e3b8eaeb7bf5b0d4d32652f253bd67e"} Mar 18 20:10:08 crc kubenswrapper[4950]: I0318 20:10:08.386351 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" Mar 18 20:10:08 crc kubenswrapper[4950]: I0318 20:10:08.431706 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:08 crc kubenswrapper[4950]: E0318 20:10:08.433042 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:08.933026833 +0000 UTC m=+222.173868701 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:08 crc kubenswrapper[4950]: I0318 20:10:08.435192 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:10:08 crc kubenswrapper[4950]: I0318 20:10:08.533355 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:08 crc kubenswrapper[4950]: E0318 20:10:08.537929 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:09.037912892 +0000 UTC m=+222.278754750 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:08 crc kubenswrapper[4950]: I0318 20:10:08.606890 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-rwl8z" podStartSLOduration=6.606872862 podStartE2EDuration="6.606872862s" podCreationTimestamp="2026-03-18 20:10:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:08.605333712 +0000 UTC m=+221.846175580" watchObservedRunningTime="2026-03-18 20:10:08.606872862 +0000 UTC m=+221.847714730" Mar 18 20:10:08 crc kubenswrapper[4950]: I0318 20:10:08.634019 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:08 crc kubenswrapper[4950]: E0318 20:10:08.636121 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:09.136105605 +0000 UTC m=+222.376947473 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:08 crc kubenswrapper[4950]: I0318 20:10:08.737519 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:08 crc kubenswrapper[4950]: E0318 20:10:08.737898 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:09.237885333 +0000 UTC m=+222.478727201 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:08 crc kubenswrapper[4950]: I0318 20:10:08.838153 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:08 crc kubenswrapper[4950]: E0318 20:10:08.838536 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:09.33852057 +0000 UTC m=+222.579362438 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:08 crc kubenswrapper[4950]: I0318 20:10:08.951776 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:08 crc kubenswrapper[4950]: E0318 20:10:08.953298 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:09.453275396 +0000 UTC m=+222.694117264 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.067115 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:09 crc kubenswrapper[4950]: E0318 20:10:09.067470 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:09.567452457 +0000 UTC m=+222.808294315 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.073467 4950 ???:1] "http: TLS handshake error from 192.168.126.11:39782: no serving certificate available for the kubelet" Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.168198 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:09 crc kubenswrapper[4950]: E0318 20:10:09.168679 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:09.668664999 +0000 UTC m=+222.909506867 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.269538 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:09 crc kubenswrapper[4950]: E0318 20:10:09.270548 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:09.770525999 +0000 UTC m=+223.011367867 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.376910 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:09 crc kubenswrapper[4950]: E0318 20:10:09.377398 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:09.877382739 +0000 UTC m=+223.118224607 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.388933 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-79fhl" event={"ID":"7fcd663f-3b4c-42fe-a585-c1e1142875ca","Type":"ContainerStarted","Data":"f13779b0cf184ab4f06ec688fac98b14122f81bf827460b8e10cc40e170ccc72"} Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.415265 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2mqb" event={"ID":"fc3e8cef-c459-47ef-9531-da9acd77054e","Type":"ContainerStarted","Data":"291edd9e94776fc6801ae1921edf79f7b266a90fdc2292d7e3de6f600336f865"} Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.435446 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-qj7bk" podStartSLOduration=174.435423084 podStartE2EDuration="2m54.435423084s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:08.805873648 +0000 UTC m=+222.046715516" watchObservedRunningTime="2026-03-18 20:10:09.435423084 +0000 UTC m=+222.676264952" Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.435913 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-79fhl" podStartSLOduration=174.435908937 podStartE2EDuration="2m54.435908937s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:09.428929505 +0000 UTC m=+222.669771373" watchObservedRunningTime="2026-03-18 20:10:09.435908937 +0000 UTC m=+222.676750805" Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.478123 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n2mqb" podStartSLOduration=174.478095558 podStartE2EDuration="2m54.478095558s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:09.470820508 +0000 UTC m=+222.711662376" watchObservedRunningTime="2026-03-18 20:10:09.478095558 +0000 UTC m=+222.718937426" Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.479845 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:09 crc kubenswrapper[4950]: E0318 20:10:09.481882 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:09.981854746 +0000 UTC m=+223.222696614 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.522186 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9pgcg" event={"ID":"1fedd09d-87e7-4bfc-ab58-589189c984d8","Type":"ContainerStarted","Data":"6578b52831803063239f02eb0d49efa3170eacdab5db834605cac7360b708b2f"} Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.547544 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-859lm" event={"ID":"cf43d927-8bd0-46df-9115-a95e6f747190","Type":"ContainerStarted","Data":"a1ea86aa2c97dfe2082ce045d35d4ce535d53134e86f11f74a23243f85424e32"} Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.554850 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-twwmd" event={"ID":"30bb01b5-2b19-403e-9930-a1991532e787","Type":"ContainerStarted","Data":"42e40d4983a5b38d5a17a66850830ad18cef5f02e922ea898e74e2d649b997a5"} Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.574611 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-ktwzg" event={"ID":"b3f2101e-4192-42db-9ff8-3dd152bc24bb","Type":"ContainerStarted","Data":"998d0cac1d21e2d9e26097be7fae4f3fb4eaaa7fcedb4ed531426e7bb61a004f"} Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.582062 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9pgcg" podStartSLOduration=174.582026062 podStartE2EDuration="2m54.582026062s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:09.574959517 +0000 UTC m=+222.815801385" watchObservedRunningTime="2026-03-18 20:10:09.582026062 +0000 UTC m=+222.822867930" Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.582195 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:09 crc kubenswrapper[4950]: E0318 20:10:09.583700 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:10.083683135 +0000 UTC m=+223.324525003 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.589689 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-5wffk" event={"ID":"8edc3dbe-4856-4b87-b5d8-bbf2c55c0127","Type":"ContainerStarted","Data":"919290ef5b9519c85580292a843d90a5c6236cd8a3b861d8907b5cc9aa1210c7"} Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.673738 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-47kcd" event={"ID":"34f027ae-fb71-460c-b978-e254fac5d50f","Type":"ContainerStarted","Data":"8080748230b77dca9e38dfa25a5f743aeae53bc29993e9268f2e88c5a1dfca31"} Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.674835 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-47kcd" Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.675798 4950 patch_prober.go:28] interesting pod/console-operator-58897d9998-47kcd container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/readyz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.675841 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-47kcd" podUID="34f027ae-fb71-460c-b978-e254fac5d50f" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.17:8443/readyz\": dial tcp 10.217.0.17:8443: connect: connection refused" Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.680469 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" event={"ID":"be93cdca-9fe6-46a2-ae6a-94516d8eaa50","Type":"ContainerStarted","Data":"0c8e324d05475dc9821f77b72e85a47ae676c3bbdc33fbe01223a3c0c4c2ff1e"} Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.683229 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:09 crc kubenswrapper[4950]: E0318 20:10:09.684264 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:10.18424873 +0000 UTC m=+223.425090598 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.708288 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-47kcd" podStartSLOduration=174.708269897 podStartE2EDuration="2m54.708269897s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:09.707842506 +0000 UTC m=+222.948684374" watchObservedRunningTime="2026-03-18 20:10:09.708269897 +0000 UTC m=+222.949111765" Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.747518 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pdnzp" event={"ID":"167a9973-d839-4aaf-973b-2e00be41f59f","Type":"ContainerStarted","Data":"4bf2e1761fe9071c1a14de91f57e0c412f6496224ce72f8cac86115353cc7742"} Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.777731 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" podStartSLOduration=174.77771112 podStartE2EDuration="2m54.77771112s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:09.775644146 +0000 UTC m=+223.016486014" watchObservedRunningTime="2026-03-18 20:10:09.77771112 +0000 UTC m=+223.018552998" Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.784671 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:09 crc kubenswrapper[4950]: E0318 20:10:09.785909 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:10.285874464 +0000 UTC m=+223.526716332 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.885448 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-27jdq" event={"ID":"ff73cd60-bdf9-477a-8ede-1e781aaffd22","Type":"ContainerStarted","Data":"c160f18e3da03c927a67c49797fdbbb6ac01aa2d73193b3ca99e19c750dd31c5"} Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.887909 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:09 crc kubenswrapper[4950]: E0318 20:10:09.888353 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:10.388338409 +0000 UTC m=+223.629180277 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.911719 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tfrtt" event={"ID":"fd0504b1-d681-4d1d-a9bb-8be90167f30f","Type":"ContainerStarted","Data":"f429a7e03a8149691e3d0c6b9fca7fb35de14b20d24bd1d0b116a767f3dec152"} Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.959229 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564400-d9vqv" event={"ID":"164efb7c-0b83-42ed-8d30-acc8a1ff1794","Type":"ContainerStarted","Data":"bbefef7236c11b25684d8d2d54cfe0434ffc226bf5beaa23de40c535693387cc"} Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.977547 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gw45j" event={"ID":"36384cbe-f2f7-48aa-8b54-ea07d8cc3678","Type":"ContainerStarted","Data":"5e90464aba3fe9eba434b1f5a19e8fd1c19becb88227aa5257960c190cdaf9b5"} Mar 18 20:10:09 crc kubenswrapper[4950]: I0318 20:10:09.996098 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:09 crc kubenswrapper[4950]: E0318 20:10:09.996450 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:10.496437761 +0000 UTC m=+223.737279629 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:09.999931 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c7t4g" event={"ID":"f962a9b5-16d9-41e3-93a1-d7347e8cc27f","Type":"ContainerStarted","Data":"c485e48ca7adb73be8a97f0f6ec970ca027ea0d87e6c5908c6a71043f7a82812"} Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.021883 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" event={"ID":"a2456057-5ff6-458f-af94-53edca249b7f","Type":"ContainerStarted","Data":"f923a6528bbab2ec46d1fd4396ff4afd6429d0bf4de07c2b95d05fe8f0abf876"} Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.036013 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29564400-d9vqv" podStartSLOduration=175.035988154 podStartE2EDuration="2m55.035988154s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:10.024764111 +0000 UTC m=+223.265605979" watchObservedRunningTime="2026-03-18 20:10:10.035988154 +0000 UTC m=+223.276830022" Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.051065 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564408-2658d" event={"ID":"97cb884c-5f55-4ca0-8639-193c2f1df139","Type":"ContainerStarted","Data":"cb0d25f2e7d0a0dbdeebd645a69c14dd59c8f100deff662973fdcc0747e297ad"} Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.057980 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-cdpgz" event={"ID":"a8ef5a1d-9e3e-4057-89f5-672b0b1ca73c","Type":"ContainerStarted","Data":"21347fe074506c5bfdc115a10dfa8f422ec89d97b692f8531f1c53a62e0047e2"} Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.100108 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xj696" event={"ID":"bb10c1c4-bc22-46af-9f3f-c948d34e0017","Type":"ContainerStarted","Data":"b9b107a0e0260f92271ccf33992291c4432dc0aa64b36f963d9a1ea58991611e"} Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.100292 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:10 crc kubenswrapper[4950]: E0318 20:10:10.109627 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:10.609580384 +0000 UTC m=+223.850422242 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.160816 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs" event={"ID":"0274f224-f29d-4c71-88e7-2111719f598b","Type":"ContainerStarted","Data":"8299a43c33848d2a3e821c186c59778fb2e39c54e1c91aa7836f50201bea1a76"} Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.160856 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs" Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.164809 4950 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-jq2zs container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.165187 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs" podUID="0274f224-f29d-4c71-88e7-2111719f598b" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.194175 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sxcm9" event={"ID":"8596ab2e-852c-4d94-a8ce-e02c823a76b1","Type":"ContainerStarted","Data":"da6ad5ab6f160c918cac2d5263eb2373039dd41e4d283905ed13aa3fc3d76091"} Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.209036 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:10 crc kubenswrapper[4950]: E0318 20:10:10.210351 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:10.710339775 +0000 UTC m=+223.951181643 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.212666 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qj7bk" event={"ID":"ce6591e6-fb40-40a0-91a7-9aa2b3aa0437","Type":"ContainerStarted","Data":"cdd541ccba415bd326d3d4568c198efbb7c158a2aaa4ad4bdf4c5adf0b16b8b5"} Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.244741 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-gsj4z" event={"ID":"90ab43dd-38e6-4264-a086-5cef08691214","Type":"ContainerStarted","Data":"d887491fa0d8230090dbf6f726da57ce20c0d9d7db774471b433373334887dd7"} Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.253112 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qqfjh" event={"ID":"28014e20-2d95-42bc-881b-026137a7a016","Type":"ContainerStarted","Data":"075e72753ab3318fb5e9845b0c55fbfff68f2d82ba07f34227ba1efe5451011b"} Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.308269 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-g2f49" event={"ID":"4684cfd6-1836-4218-9c6e-c6ca75a2da96","Type":"ContainerStarted","Data":"d01b04c2dcad3172d11509dafbc5c410a7e85cc39adecc407a845e81750e0e62"} Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.308421 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-g2f49" Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.309600 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:10 crc kubenswrapper[4950]: E0318 20:10:10.309728 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:10.809704899 +0000 UTC m=+224.050546767 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.309979 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:10 crc kubenswrapper[4950]: E0318 20:10:10.311287 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:10.81127429 +0000 UTC m=+224.052116158 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.354603 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-kbjxd" event={"ID":"5d345627-baa3-4354-a422-1b3646e5cc48","Type":"ContainerStarted","Data":"1fd80d3ab9a98d12851ba9e35393b938c947beec7d55b2050baabb29ac1f9ee8"} Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.354926 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-kbjxd" event={"ID":"5d345627-baa3-4354-a422-1b3646e5cc48","Type":"ContainerStarted","Data":"8e23761f7031e5748ea6eedcb91a9944db14dd73f67e6bc2d144081342f508e4"} Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.373666 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xj696" podStartSLOduration=175.373650628 podStartE2EDuration="2m55.373650628s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:10.187642102 +0000 UTC m=+223.428483970" watchObservedRunningTime="2026-03-18 20:10:10.373650628 +0000 UTC m=+223.614492496" Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.391660 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-544q6" event={"ID":"035f024b-dad2-4ad3-96bb-ebc3c52ff9d4","Type":"ContainerStarted","Data":"36c8695f78f914b5b56cfba6a2919b71035b007f89b58bdfd113e3836bc9a79c"} Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.401076 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vnw2w" event={"ID":"94700472-065a-4371-8025-5c4e28b1e722","Type":"ContainerStarted","Data":"b6e7eca854467fd6531b9ee06396b87d8008e0df333bbf8cad534cbb0df742dc"} Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.401124 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vnw2w" event={"ID":"94700472-065a-4371-8025-5c4e28b1e722","Type":"ContainerStarted","Data":"c950ba28356ee1919e66e8d5874160833ca96d91a40726368728f90020dc1c96"} Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.402111 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vnw2w" Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.407553 4950 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-vnw2w container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.407599 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vnw2w" podUID="94700472-065a-4371-8025-5c4e28b1e722" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.410928 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:10 crc kubenswrapper[4950]: E0318 20:10:10.412207 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:10.912193265 +0000 UTC m=+224.153035133 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.426804 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-pnffr" event={"ID":"d2c83122-863b-41e6-9118-54fa716c8485","Type":"ContainerStarted","Data":"ee0a907c6e106be6a3e1d24eafc95241d276a528e12041057bda190adb26a5fd"} Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.428478 4950 patch_prober.go:28] interesting pod/downloads-7954f5f757-rf8k5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.428518 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rf8k5" podUID="290d1b89-d5c9-4928-8973-6c33f476b091" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.512252 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:10 crc kubenswrapper[4950]: E0318 20:10:10.519529 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:11.019513686 +0000 UTC m=+224.260355544 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.590934 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-g2f49" podStartSLOduration=175.59089793 podStartE2EDuration="2m55.59089793s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:10.590667294 +0000 UTC m=+223.831509172" watchObservedRunningTime="2026-03-18 20:10:10.59089793 +0000 UTC m=+223.831739798" Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.598673 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs" podStartSLOduration=175.598618482 podStartE2EDuration="2m55.598618482s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:10.391050803 +0000 UTC m=+223.631892671" watchObservedRunningTime="2026-03-18 20:10:10.598618482 +0000 UTC m=+223.839460350" Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.616701 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:10 crc kubenswrapper[4950]: E0318 20:10:10.617142 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:11.117128675 +0000 UTC m=+224.357970543 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.681205 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-gsj4z" Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.689732 4950 patch_prober.go:28] interesting pod/router-default-5444994796-gsj4z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 20:10:10 crc kubenswrapper[4950]: [-]has-synced failed: reason withheld Mar 18 20:10:10 crc kubenswrapper[4950]: [+]process-running ok Mar 18 20:10:10 crc kubenswrapper[4950]: healthz check failed Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.689796 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsj4z" podUID="90ab43dd-38e6-4264-a086-5cef08691214" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.707227 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-kbjxd" podStartSLOduration=8.707207437 podStartE2EDuration="8.707207437s" podCreationTimestamp="2026-03-18 20:10:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:10.702569686 +0000 UTC m=+223.943411554" watchObservedRunningTime="2026-03-18 20:10:10.707207437 +0000 UTC m=+223.948049315" Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.722244 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:10 crc kubenswrapper[4950]: E0318 20:10:10.722604 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:11.222592028 +0000 UTC m=+224.463433886 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.738639 4950 ???:1] "http: TLS handshake error from 192.168.126.11:39798: no serving certificate available for the kubelet" Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.832041 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:10 crc kubenswrapper[4950]: E0318 20:10:10.832574 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:11.332555829 +0000 UTC m=+224.573397697 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.890556 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-gsj4z" podStartSLOduration=175.890538063 podStartE2EDuration="2m55.890538063s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:10.889011353 +0000 UTC m=+224.129853221" watchObservedRunningTime="2026-03-18 20:10:10.890538063 +0000 UTC m=+224.131379931" Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.890877 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sxcm9" podStartSLOduration=175.890872962 podStartE2EDuration="2m55.890872962s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:10.777644036 +0000 UTC m=+224.018485904" watchObservedRunningTime="2026-03-18 20:10:10.890872962 +0000 UTC m=+224.131714820" Mar 18 20:10:10 crc kubenswrapper[4950]: I0318 20:10:10.933816 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:10 crc kubenswrapper[4950]: E0318 20:10:10.934153 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:11.434139521 +0000 UTC m=+224.674981389 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.014739 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-544q6" podStartSLOduration=176.014724975 podStartE2EDuration="2m56.014724975s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:11.012782155 +0000 UTC m=+224.253624023" watchObservedRunningTime="2026-03-18 20:10:11.014724975 +0000 UTC m=+224.255566843" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.034677 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:11 crc kubenswrapper[4950]: E0318 20:10:11.034904 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:11.534873452 +0000 UTC m=+224.775715320 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.035141 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:11 crc kubenswrapper[4950]: E0318 20:10:11.035482 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:11.535475637 +0000 UTC m=+224.776317505 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.136617 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:11 crc kubenswrapper[4950]: E0318 20:10:11.137291 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:11.637275935 +0000 UTC m=+224.878117803 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.154048 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vnw2w" podStartSLOduration=176.154031453 podStartE2EDuration="2m56.154031453s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:11.152146543 +0000 UTC m=+224.392988411" watchObservedRunningTime="2026-03-18 20:10:11.154031453 +0000 UTC m=+224.394873321" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.240222 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:11 crc kubenswrapper[4950]: E0318 20:10:11.240505 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:11.74049306 +0000 UTC m=+224.981334928 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.341178 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:11 crc kubenswrapper[4950]: E0318 20:10:11.344568 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:11.844551747 +0000 UTC m=+225.085393615 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.445198 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.445253 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.445276 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.445318 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.445344 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:10:11 crc kubenswrapper[4950]: E0318 20:10:11.446068 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:11.946056837 +0000 UTC m=+225.186898705 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.446166 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.452053 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.454614 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.457934 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.458913 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-544q6" event={"ID":"035f024b-dad2-4ad3-96bb-ebc3c52ff9d4","Type":"ContainerStarted","Data":"bc239b68c69b1ff75515bbd551e0529c9266a630f2f33bd42c70cf98bd4c26c3"} Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.461850 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xj696" event={"ID":"bb10c1c4-bc22-46af-9f3f-c948d34e0017","Type":"ContainerStarted","Data":"21e0dd9e141890e23daf618fc5ce12e9466febc7729479b7d29d1f336947f184"} Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.476197 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-ktwzg" event={"ID":"b3f2101e-4192-42db-9ff8-3dd152bc24bb","Type":"ContainerStarted","Data":"047e4cd55d49a399000cb76b477fc81920182a6f8ca07278e01916a34997e1b0"} Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.504368 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.526505 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.539473 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-27jdq" event={"ID":"ff73cd60-bdf9-477a-8ede-1e781aaffd22","Type":"ContainerStarted","Data":"f2f23a45d83d8fc714ff18321d47ee26861f081c9ff649a789d5ce8ffeca0955"} Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.544339 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-ktwzg" podStartSLOduration=176.544326462 podStartE2EDuration="2m56.544326462s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:11.542595357 +0000 UTC m=+224.783437235" watchObservedRunningTime="2026-03-18 20:10:11.544326462 +0000 UTC m=+224.785168330" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.551368 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:11 crc kubenswrapper[4950]: E0318 20:10:11.551911 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:12.05189466 +0000 UTC m=+225.292736518 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.556068 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sxcm9" event={"ID":"8596ab2e-852c-4d94-a8ce-e02c823a76b1","Type":"ContainerStarted","Data":"5939118945ad28467af4aae08b880ce353cbe6e1827bec19bad76aca39495812"} Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.590523 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.614806 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-cdpgz" event={"ID":"a8ef5a1d-9e3e-4057-89f5-672b0b1ca73c","Type":"ContainerStarted","Data":"6d19f9ab86533189847d4ef34393f32aad4ca228bfd2da1b5c2e2c5c13dba37d"} Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.614864 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-cdpgz" event={"ID":"a8ef5a1d-9e3e-4057-89f5-672b0b1ca73c","Type":"ContainerStarted","Data":"9adfecca9abdd5d09207305766d559aae87f6de40f5703acff39149e35af6653"} Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.615667 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-cdpgz" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.619894 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-twwmd" event={"ID":"30bb01b5-2b19-403e-9930-a1991532e787","Type":"ContainerStarted","Data":"56e58a3856c1c11e3f76c80ef813c39df4133fa0ebbf2821a4155e538d05e1f2"} Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.619927 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-twwmd" event={"ID":"30bb01b5-2b19-403e-9930-a1991532e787","Type":"ContainerStarted","Data":"d7ef9e01958fc0d131ea2a689df314ef6d41dff20b90c4c830953b6f9d96b5d4"} Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.620166 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-27jdq" podStartSLOduration=176.620157322 podStartE2EDuration="2m56.620157322s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:11.61856346 +0000 UTC m=+224.859405328" watchObservedRunningTime="2026-03-18 20:10:11.620157322 +0000 UTC m=+224.860999190" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.626596 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564400-d9vqv" event={"ID":"164efb7c-0b83-42ed-8d30-acc8a1ff1794","Type":"ContainerStarted","Data":"25e62c249dbad8da15635226eb6e8947d89177ad45c9ad3022cf2b7827bfcaff"} Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.656038 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.657766 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" event={"ID":"0560ae4e-4142-48c1-9829-2e770a3b56b4","Type":"ContainerStarted","Data":"bcce6b32aeeab337cb922eb856b84e61b0a80d2131aa169fcf978938533a71d7"} Mar 18 20:10:11 crc kubenswrapper[4950]: E0318 20:10:11.657915 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:12.157902438 +0000 UTC m=+225.398744296 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.676314 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qqfjh" event={"ID":"28014e20-2d95-42bc-881b-026137a7a016","Type":"ContainerStarted","Data":"7f07223436bcc989ae10bfb340b819c897ae1f4d6532c741a4944e79534c11e3"} Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.676359 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qqfjh" event={"ID":"28014e20-2d95-42bc-881b-026137a7a016","Type":"ContainerStarted","Data":"104a23f3110bc6d9ef26f381c5510be5d98a712251ec6d72510f40ecc6a12b36"} Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.677045 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qqfjh" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.698527 4950 patch_prober.go:28] interesting pod/router-default-5444994796-gsj4z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 20:10:11 crc kubenswrapper[4950]: [-]has-synced failed: reason withheld Mar 18 20:10:11 crc kubenswrapper[4950]: [+]process-running ok Mar 18 20:10:11 crc kubenswrapper[4950]: healthz check failed Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.698581 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsj4z" podUID="90ab43dd-38e6-4264-a086-5cef08691214" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.713167 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pdnzp" event={"ID":"167a9973-d839-4aaf-973b-2e00be41f59f","Type":"ContainerStarted","Data":"436856631d22d74312f9ce207b140b5d6fcb3a74f5306bc4f30857a6cabff9de"} Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.717574 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-cdpgz" podStartSLOduration=9.717556545 podStartE2EDuration="9.717556545s" podCreationTimestamp="2026-03-18 20:10:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:11.716846466 +0000 UTC m=+224.957688334" watchObservedRunningTime="2026-03-18 20:10:11.717556545 +0000 UTC m=+224.958398413" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.723955 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tfrtt" event={"ID":"fd0504b1-d681-4d1d-a9bb-8be90167f30f","Type":"ContainerStarted","Data":"d805a923f34e053ccbc9796107fb2cc0d3c4e56b90e40a4036331dacfb22fc66"} Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.724467 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tfrtt" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.741354 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-859lm" event={"ID":"cf43d927-8bd0-46df-9115-a95e6f747190","Type":"ContainerStarted","Data":"ec3eb6c498d41e44804e191e8c7733089f51f5aa1a400e0767e6ddb3233f7bab"} Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.762397 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-5wffk" event={"ID":"8edc3dbe-4856-4b87-b5d8-bbf2c55c0127","Type":"ContainerStarted","Data":"506e81c609f8228bfb48012e816e3303340427c0b917081ea9aa669aaa40fbb2"} Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.762451 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-5wffk" event={"ID":"8edc3dbe-4856-4b87-b5d8-bbf2c55c0127","Type":"ContainerStarted","Data":"faaab943c159690cdb58d14eaa03f854eca3cdede3a25b9091921e653d3b7fc6"} Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.766947 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:11 crc kubenswrapper[4950]: E0318 20:10:11.770356 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:12.270337443 +0000 UTC m=+225.511179311 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.780959 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c7t4g" event={"ID":"f962a9b5-16d9-41e3-93a1-d7347e8cc27f","Type":"ContainerStarted","Data":"05d85023f996b9614a690f0bcb05604139633e68a10de02471ec454a1091ac69"} Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.781002 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c7t4g" event={"ID":"f962a9b5-16d9-41e3-93a1-d7347e8cc27f","Type":"ContainerStarted","Data":"b5dede57c5226bc24415ccac6548074b2465f9b1c9255ec7b5fff9a01ea089fb"} Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.792440 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pdnzp" podStartSLOduration=176.792404699 podStartE2EDuration="2m56.792404699s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:11.782763397 +0000 UTC m=+225.023605265" watchObservedRunningTime="2026-03-18 20:10:11.792404699 +0000 UTC m=+225.033246567" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.810434 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tfrtt" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.811130 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" event={"ID":"a2456057-5ff6-458f-af94-53edca249b7f","Type":"ContainerStarted","Data":"bb43772a7f0be4bd5df4460e9a3438e9991f1cee7c7ca14fcaf9f7e315f4121e"} Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.811311 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.813096 4950 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-8nhwt container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.813140 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" podUID="a2456057-5ff6-458f-af94-53edca249b7f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.850638 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-pnffr" event={"ID":"d2c83122-863b-41e6-9118-54fa716c8485","Type":"ContainerStarted","Data":"d45e7ce4c4010018def2cf8f22d62efe7a9e931bc49e1b889abbe2feafd5f04b"} Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.857818 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qqfjh" podStartSLOduration=176.857803717 podStartE2EDuration="2m56.857803717s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:11.853922425 +0000 UTC m=+225.094764293" watchObservedRunningTime="2026-03-18 20:10:11.857803717 +0000 UTC m=+225.098645585" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.868362 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:11 crc kubenswrapper[4950]: E0318 20:10:11.869738 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:12.369722068 +0000 UTC m=+225.610563926 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.872499 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gw45j" event={"ID":"36384cbe-f2f7-48aa-8b54-ea07d8cc3678","Type":"ContainerStarted","Data":"a14cddf3cfe676aaf1d5b30c1d33e55fd346664f6f3dd3a8197fcdf3994b14ae"} Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.872534 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gw45j" event={"ID":"36384cbe-f2f7-48aa-8b54-ea07d8cc3678","Type":"ContainerStarted","Data":"59a5db7089d6728dc788c19df917c690199523801684abbf7dbbe83ab37b3d23"} Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.882054 4950 patch_prober.go:28] interesting pod/downloads-7954f5f757-rf8k5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.882097 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rf8k5" podUID="290d1b89-d5c9-4928-8973-6c33f476b091" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Mar 18 20:10:11 crc kubenswrapper[4950]: I0318 20:10:11.969702 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:11 crc kubenswrapper[4950]: E0318 20:10:11.971498 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:12.471471444 +0000 UTC m=+225.712313372 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.043885 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-859lm" podStartSLOduration=177.043866794 podStartE2EDuration="2m57.043866794s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:12.043628688 +0000 UTC m=+225.284470556" watchObservedRunningTime="2026-03-18 20:10:12.043866794 +0000 UTC m=+225.284708662" Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.044708 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-twwmd" podStartSLOduration=177.044702406 podStartE2EDuration="2m57.044702406s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:11.95370469 +0000 UTC m=+225.194546558" watchObservedRunningTime="2026-03-18 20:10:12.044702406 +0000 UTC m=+225.285544264" Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.071817 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:12 crc kubenswrapper[4950]: E0318 20:10:12.072203 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:12.572189744 +0000 UTC m=+225.813031612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.100675 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" podStartSLOduration=177.100661087 podStartE2EDuration="2m57.100661087s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:12.099912897 +0000 UTC m=+225.340754765" watchObservedRunningTime="2026-03-18 20:10:12.100661087 +0000 UTC m=+225.341502945" Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.147681 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vnw2w" Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.147928 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gw45j" podStartSLOduration=177.147909831 podStartE2EDuration="2m57.147909831s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:12.144936933 +0000 UTC m=+225.385778791" watchObservedRunningTime="2026-03-18 20:10:12.147909831 +0000 UTC m=+225.388751699" Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.173845 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:12 crc kubenswrapper[4950]: E0318 20:10:12.174249 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:12.674233938 +0000 UTC m=+225.915075806 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.275458 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:12 crc kubenswrapper[4950]: E0318 20:10:12.275789 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:12.775776319 +0000 UTC m=+226.016618187 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.285723 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tfrtt" podStartSLOduration=177.285710468 podStartE2EDuration="2m57.285710468s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:12.23060683 +0000 UTC m=+225.471448698" watchObservedRunningTime="2026-03-18 20:10:12.285710468 +0000 UTC m=+225.526552336" Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.286378 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-c7t4g" podStartSLOduration=177.286372026 podStartE2EDuration="2m57.286372026s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:12.281340194 +0000 UTC m=+225.522182062" watchObservedRunningTime="2026-03-18 20:10:12.286372026 +0000 UTC m=+225.527213894" Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.356948 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-pnffr" podStartSLOduration=177.356932708 podStartE2EDuration="2m57.356932708s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:12.354872574 +0000 UTC m=+225.595714442" watchObservedRunningTime="2026-03-18 20:10:12.356932708 +0000 UTC m=+225.597774576" Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.378896 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:12 crc kubenswrapper[4950]: E0318 20:10:12.379334 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:12.879316372 +0000 UTC m=+226.120158240 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.394615 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-5wffk" podStartSLOduration=177.394597801 podStartE2EDuration="2m57.394597801s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:12.394198651 +0000 UTC m=+225.635040519" watchObservedRunningTime="2026-03-18 20:10:12.394597801 +0000 UTC m=+225.635439669" Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.480875 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:12 crc kubenswrapper[4950]: E0318 20:10:12.481229 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:12.981214753 +0000 UTC m=+226.222056621 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.582054 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:12 crc kubenswrapper[4950]: E0318 20:10:12.582368 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:13.082353653 +0000 UTC m=+226.323195521 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.682900 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:12 crc kubenswrapper[4950]: E0318 20:10:12.683246 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:13.183235497 +0000 UTC m=+226.424077365 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.685738 4950 patch_prober.go:28] interesting pod/router-default-5444994796-gsj4z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 20:10:12 crc kubenswrapper[4950]: [-]has-synced failed: reason withheld Mar 18 20:10:12 crc kubenswrapper[4950]: [+]process-running ok Mar 18 20:10:12 crc kubenswrapper[4950]: healthz check failed Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.685800 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsj4z" podUID="90ab43dd-38e6-4264-a086-5cef08691214" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.788885 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:12 crc kubenswrapper[4950]: E0318 20:10:12.789253 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:13.289225304 +0000 UTC m=+226.530067172 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.851745 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-47kcd" Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.881728 4950 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-jq2zs container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.881788 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs" podUID="0274f224-f29d-4c71-88e7-2111719f598b" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.883321 4950 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-8nhwt container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.883355 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" podUID="a2456057-5ff6-458f-af94-53edca249b7f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.890247 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:12 crc kubenswrapper[4950]: E0318 20:10:12.890731 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:13.390718004 +0000 UTC m=+226.631559872 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:12 crc kubenswrapper[4950]: I0318 20:10:12.994717 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:12 crc kubenswrapper[4950]: E0318 20:10:12.996214 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:13.496199458 +0000 UTC m=+226.737041326 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.098122 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:13 crc kubenswrapper[4950]: E0318 20:10:13.098385 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:13.598373545 +0000 UTC m=+226.839215403 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.153947 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8kwk7"] Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.154947 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8kwk7" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.168558 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.201942 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:13 crc kubenswrapper[4950]: E0318 20:10:13.202338 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:13.702324099 +0000 UTC m=+226.943165967 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.228528 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8kwk7"] Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.303506 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgcwj\" (UniqueName: \"kubernetes.io/projected/4a386d81-3e2c-4df7-988d-5fb1758be616-kube-api-access-bgcwj\") pod \"certified-operators-8kwk7\" (UID: \"4a386d81-3e2c-4df7-988d-5fb1758be616\") " pod="openshift-marketplace/certified-operators-8kwk7" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.303559 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a386d81-3e2c-4df7-988d-5fb1758be616-utilities\") pod \"certified-operators-8kwk7\" (UID: \"4a386d81-3e2c-4df7-988d-5fb1758be616\") " pod="openshift-marketplace/certified-operators-8kwk7" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.303581 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a386d81-3e2c-4df7-988d-5fb1758be616-catalog-content\") pod \"certified-operators-8kwk7\" (UID: \"4a386d81-3e2c-4df7-988d-5fb1758be616\") " pod="openshift-marketplace/certified-operators-8kwk7" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.303610 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:13 crc kubenswrapper[4950]: E0318 20:10:13.303976 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:13.803958983 +0000 UTC m=+227.044800851 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.353680 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gzj5q"] Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.354677 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gzj5q" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.374857 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.405889 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.406099 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgcwj\" (UniqueName: \"kubernetes.io/projected/4a386d81-3e2c-4df7-988d-5fb1758be616-kube-api-access-bgcwj\") pod \"certified-operators-8kwk7\" (UID: \"4a386d81-3e2c-4df7-988d-5fb1758be616\") " pod="openshift-marketplace/certified-operators-8kwk7" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.406136 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a386d81-3e2c-4df7-988d-5fb1758be616-utilities\") pod \"certified-operators-8kwk7\" (UID: \"4a386d81-3e2c-4df7-988d-5fb1758be616\") " pod="openshift-marketplace/certified-operators-8kwk7" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.406154 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a386d81-3e2c-4df7-988d-5fb1758be616-catalog-content\") pod \"certified-operators-8kwk7\" (UID: \"4a386d81-3e2c-4df7-988d-5fb1758be616\") " pod="openshift-marketplace/certified-operators-8kwk7" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.406588 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a386d81-3e2c-4df7-988d-5fb1758be616-catalog-content\") pod \"certified-operators-8kwk7\" (UID: \"4a386d81-3e2c-4df7-988d-5fb1758be616\") " pod="openshift-marketplace/certified-operators-8kwk7" Mar 18 20:10:13 crc kubenswrapper[4950]: E0318 20:10:13.406662 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:13.906648254 +0000 UTC m=+227.147490122 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.407578 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a386d81-3e2c-4df7-988d-5fb1758be616-utilities\") pod \"certified-operators-8kwk7\" (UID: \"4a386d81-3e2c-4df7-988d-5fb1758be616\") " pod="openshift-marketplace/certified-operators-8kwk7" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.408268 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gzj5q"] Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.478355 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgcwj\" (UniqueName: \"kubernetes.io/projected/4a386d81-3e2c-4df7-988d-5fb1758be616-kube-api-access-bgcwj\") pod \"certified-operators-8kwk7\" (UID: \"4a386d81-3e2c-4df7-988d-5fb1758be616\") " pod="openshift-marketplace/certified-operators-8kwk7" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.509075 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slf9m\" (UniqueName: \"kubernetes.io/projected/044d9d2a-6b99-49ee-b00d-728c7832815c-kube-api-access-slf9m\") pod \"community-operators-gzj5q\" (UID: \"044d9d2a-6b99-49ee-b00d-728c7832815c\") " pod="openshift-marketplace/community-operators-gzj5q" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.509128 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.509174 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/044d9d2a-6b99-49ee-b00d-728c7832815c-utilities\") pod \"community-operators-gzj5q\" (UID: \"044d9d2a-6b99-49ee-b00d-728c7832815c\") " pod="openshift-marketplace/community-operators-gzj5q" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.509208 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/044d9d2a-6b99-49ee-b00d-728c7832815c-catalog-content\") pod \"community-operators-gzj5q\" (UID: \"044d9d2a-6b99-49ee-b00d-728c7832815c\") " pod="openshift-marketplace/community-operators-gzj5q" Mar 18 20:10:13 crc kubenswrapper[4950]: E0318 20:10:13.509495 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:14.009478619 +0000 UTC m=+227.250320487 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.525087 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8kwk7" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.608593 4950 ???:1] "http: TLS handshake error from 192.168.126.11:34510: no serving certificate available for the kubelet" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.610023 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.610194 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/044d9d2a-6b99-49ee-b00d-728c7832815c-utilities\") pod \"community-operators-gzj5q\" (UID: \"044d9d2a-6b99-49ee-b00d-728c7832815c\") " pod="openshift-marketplace/community-operators-gzj5q" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.610229 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/044d9d2a-6b99-49ee-b00d-728c7832815c-catalog-content\") pod \"community-operators-gzj5q\" (UID: \"044d9d2a-6b99-49ee-b00d-728c7832815c\") " pod="openshift-marketplace/community-operators-gzj5q" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.610287 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slf9m\" (UniqueName: \"kubernetes.io/projected/044d9d2a-6b99-49ee-b00d-728c7832815c-kube-api-access-slf9m\") pod \"community-operators-gzj5q\" (UID: \"044d9d2a-6b99-49ee-b00d-728c7832815c\") " pod="openshift-marketplace/community-operators-gzj5q" Mar 18 20:10:13 crc kubenswrapper[4950]: E0318 20:10:13.610613 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:14.110599839 +0000 UTC m=+227.351441707 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.610856 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/044d9d2a-6b99-49ee-b00d-728c7832815c-catalog-content\") pod \"community-operators-gzj5q\" (UID: \"044d9d2a-6b99-49ee-b00d-728c7832815c\") " pod="openshift-marketplace/community-operators-gzj5q" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.610912 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/044d9d2a-6b99-49ee-b00d-728c7832815c-utilities\") pod \"community-operators-gzj5q\" (UID: \"044d9d2a-6b99-49ee-b00d-728c7832815c\") " pod="openshift-marketplace/community-operators-gzj5q" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.611467 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fk4b9"] Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.612571 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fk4b9" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.695260 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fk4b9"] Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.698087 4950 patch_prober.go:28] interesting pod/router-default-5444994796-gsj4z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 20:10:13 crc kubenswrapper[4950]: [-]has-synced failed: reason withheld Mar 18 20:10:13 crc kubenswrapper[4950]: [+]process-running ok Mar 18 20:10:13 crc kubenswrapper[4950]: healthz check failed Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.698164 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsj4z" podUID="90ab43dd-38e6-4264-a086-5cef08691214" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.711664 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f54c83ec-47ad-4ced-93c3-d321460f40f3-utilities\") pod \"certified-operators-fk4b9\" (UID: \"f54c83ec-47ad-4ced-93c3-d321460f40f3\") " pod="openshift-marketplace/certified-operators-fk4b9" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.711723 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f54c83ec-47ad-4ced-93c3-d321460f40f3-catalog-content\") pod \"certified-operators-fk4b9\" (UID: \"f54c83ec-47ad-4ced-93c3-d321460f40f3\") " pod="openshift-marketplace/certified-operators-fk4b9" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.711760 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s64w\" (UniqueName: \"kubernetes.io/projected/f54c83ec-47ad-4ced-93c3-d321460f40f3-kube-api-access-9s64w\") pod \"certified-operators-fk4b9\" (UID: \"f54c83ec-47ad-4ced-93c3-d321460f40f3\") " pod="openshift-marketplace/certified-operators-fk4b9" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.711807 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.717729 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rwzv9"] Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.719357 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rwzv9" Mar 18 20:10:13 crc kubenswrapper[4950]: E0318 20:10:13.728257 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:14.228231369 +0000 UTC m=+227.469073247 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.782523 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rwzv9"] Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.813483 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.813928 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f54c83ec-47ad-4ced-93c3-d321460f40f3-catalog-content\") pod \"certified-operators-fk4b9\" (UID: \"f54c83ec-47ad-4ced-93c3-d321460f40f3\") " pod="openshift-marketplace/certified-operators-fk4b9" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.813965 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82c68303-8af4-41a0-af35-312cc8b0bd7e-utilities\") pod \"community-operators-rwzv9\" (UID: \"82c68303-8af4-41a0-af35-312cc8b0bd7e\") " pod="openshift-marketplace/community-operators-rwzv9" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.813986 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s64w\" (UniqueName: \"kubernetes.io/projected/f54c83ec-47ad-4ced-93c3-d321460f40f3-kube-api-access-9s64w\") pod \"certified-operators-fk4b9\" (UID: \"f54c83ec-47ad-4ced-93c3-d321460f40f3\") " pod="openshift-marketplace/certified-operators-fk4b9" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.814019 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swzmx\" (UniqueName: \"kubernetes.io/projected/82c68303-8af4-41a0-af35-312cc8b0bd7e-kube-api-access-swzmx\") pod \"community-operators-rwzv9\" (UID: \"82c68303-8af4-41a0-af35-312cc8b0bd7e\") " pod="openshift-marketplace/community-operators-rwzv9" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.814087 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f54c83ec-47ad-4ced-93c3-d321460f40f3-utilities\") pod \"certified-operators-fk4b9\" (UID: \"f54c83ec-47ad-4ced-93c3-d321460f40f3\") " pod="openshift-marketplace/certified-operators-fk4b9" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.814104 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82c68303-8af4-41a0-af35-312cc8b0bd7e-catalog-content\") pod \"community-operators-rwzv9\" (UID: \"82c68303-8af4-41a0-af35-312cc8b0bd7e\") " pod="openshift-marketplace/community-operators-rwzv9" Mar 18 20:10:13 crc kubenswrapper[4950]: E0318 20:10:13.814307 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:14.314282095 +0000 UTC m=+227.555123963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.814727 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f54c83ec-47ad-4ced-93c3-d321460f40f3-catalog-content\") pod \"certified-operators-fk4b9\" (UID: \"f54c83ec-47ad-4ced-93c3-d321460f40f3\") " pod="openshift-marketplace/certified-operators-fk4b9" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.814920 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f54c83ec-47ad-4ced-93c3-d321460f40f3-utilities\") pod \"certified-operators-fk4b9\" (UID: \"f54c83ec-47ad-4ced-93c3-d321460f40f3\") " pod="openshift-marketplace/certified-operators-fk4b9" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.885652 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slf9m\" (UniqueName: \"kubernetes.io/projected/044d9d2a-6b99-49ee-b00d-728c7832815c-kube-api-access-slf9m\") pod \"community-operators-gzj5q\" (UID: \"044d9d2a-6b99-49ee-b00d-728c7832815c\") " pod="openshift-marketplace/community-operators-gzj5q" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.888145 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s64w\" (UniqueName: \"kubernetes.io/projected/f54c83ec-47ad-4ced-93c3-d321460f40f3-kube-api-access-9s64w\") pod \"certified-operators-fk4b9\" (UID: \"f54c83ec-47ad-4ced-93c3-d321460f40f3\") " pod="openshift-marketplace/certified-operators-fk4b9" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.917225 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.917293 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82c68303-8af4-41a0-af35-312cc8b0bd7e-catalog-content\") pod \"community-operators-rwzv9\" (UID: \"82c68303-8af4-41a0-af35-312cc8b0bd7e\") " pod="openshift-marketplace/community-operators-rwzv9" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.917331 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82c68303-8af4-41a0-af35-312cc8b0bd7e-utilities\") pod \"community-operators-rwzv9\" (UID: \"82c68303-8af4-41a0-af35-312cc8b0bd7e\") " pod="openshift-marketplace/community-operators-rwzv9" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.917359 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swzmx\" (UniqueName: \"kubernetes.io/projected/82c68303-8af4-41a0-af35-312cc8b0bd7e-kube-api-access-swzmx\") pod \"community-operators-rwzv9\" (UID: \"82c68303-8af4-41a0-af35-312cc8b0bd7e\") " pod="openshift-marketplace/community-operators-rwzv9" Mar 18 20:10:13 crc kubenswrapper[4950]: E0318 20:10:13.917865 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:14.417850619 +0000 UTC m=+227.658692487 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.918317 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82c68303-8af4-41a0-af35-312cc8b0bd7e-catalog-content\") pod \"community-operators-rwzv9\" (UID: \"82c68303-8af4-41a0-af35-312cc8b0bd7e\") " pod="openshift-marketplace/community-operators-rwzv9" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.918536 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82c68303-8af4-41a0-af35-312cc8b0bd7e-utilities\") pod \"community-operators-rwzv9\" (UID: \"82c68303-8af4-41a0-af35-312cc8b0bd7e\") " pod="openshift-marketplace/community-operators-rwzv9" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.960323 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fk4b9" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.962151 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swzmx\" (UniqueName: \"kubernetes.io/projected/82c68303-8af4-41a0-af35-312cc8b0bd7e-kube-api-access-swzmx\") pod \"community-operators-rwzv9\" (UID: \"82c68303-8af4-41a0-af35-312cc8b0bd7e\") " pod="openshift-marketplace/community-operators-rwzv9" Mar 18 20:10:13 crc kubenswrapper[4950]: I0318 20:10:13.979141 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gzj5q" Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.013110 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-g2f49" Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.020099 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:14 crc kubenswrapper[4950]: E0318 20:10:14.020512 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:14.520496609 +0000 UTC m=+227.761338477 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.020724 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" event={"ID":"0560ae4e-4142-48c1-9829-2e770a3b56b4","Type":"ContainerStarted","Data":"fecc04ad7882f6c682279ac54416c952244708533696b7e424c0e17103c2877f"} Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.032133 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"451a99b6e1212dd212e84080fd1ebf6f79b59e11cf71de6d341937ce38239a15"} Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.041649 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"83e06a5e2ded3872dc6210d6a2408e3785bc40439287b136c94d3af9db5366e8"} Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.045285 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rwzv9" Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.077834 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"522f51b3cc54f5aaee9f763f2e43676e4a81e7fb4925ed361cffb7500ceb2110"} Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.094728 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.097451 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.115809 4950 patch_prober.go:28] interesting pod/apiserver-76f77b778f-27jdq container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Mar 18 20:10:14 crc kubenswrapper[4950]: [+]log ok Mar 18 20:10:14 crc kubenswrapper[4950]: [+]etcd ok Mar 18 20:10:14 crc kubenswrapper[4950]: [+]poststarthook/start-apiserver-admission-initializer ok Mar 18 20:10:14 crc kubenswrapper[4950]: [+]poststarthook/generic-apiserver-start-informers ok Mar 18 20:10:14 crc kubenswrapper[4950]: [+]poststarthook/max-in-flight-filter ok Mar 18 20:10:14 crc kubenswrapper[4950]: [+]poststarthook/storage-object-count-tracker-hook ok Mar 18 20:10:14 crc kubenswrapper[4950]: [+]poststarthook/image.openshift.io-apiserver-caches ok Mar 18 20:10:14 crc kubenswrapper[4950]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Mar 18 20:10:14 crc kubenswrapper[4950]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Mar 18 20:10:14 crc kubenswrapper[4950]: [+]poststarthook/project.openshift.io-projectcache ok Mar 18 20:10:14 crc kubenswrapper[4950]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Mar 18 20:10:14 crc kubenswrapper[4950]: [+]poststarthook/openshift.io-startinformers ok Mar 18 20:10:14 crc kubenswrapper[4950]: [+]poststarthook/openshift.io-restmapperupdater ok Mar 18 20:10:14 crc kubenswrapper[4950]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Mar 18 20:10:14 crc kubenswrapper[4950]: livez check failed Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.115863 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-27jdq" podUID="ff73cd60-bdf9-477a-8ede-1e781aaffd22" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.123457 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:14 crc kubenswrapper[4950]: E0318 20:10:14.124311 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:14.624299419 +0000 UTC m=+227.865141287 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.143343 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.144142 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.174878 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.226630 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:14 crc kubenswrapper[4950]: E0318 20:10:14.226731 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:14.726716373 +0000 UTC m=+227.967558231 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.227103 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:14 crc kubenswrapper[4950]: E0318 20:10:14.232562 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:14.732545915 +0000 UTC m=+227.973387783 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.328024 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:14 crc kubenswrapper[4950]: E0318 20:10:14.328687 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:14.828671265 +0000 UTC m=+228.069513133 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.409701 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8kwk7"] Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.430873 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:14 crc kubenswrapper[4950]: E0318 20:10:14.431157 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:14.93114466 +0000 UTC m=+228.171986518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.466254 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9l9mq"] Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.466470 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" podUID="70601ef2-ab17-4550-890b-b29bf1a7b85e" containerName="controller-manager" containerID="cri-o://6a6ac9f276afaa5327c38326539da8d4956e49fcf92a25f8ca48d1dfced47c93" gracePeriod=30 Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.537862 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:14 crc kubenswrapper[4950]: E0318 20:10:14.538219 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:15.038205646 +0000 UTC m=+228.279047514 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.570681 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54"] Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.570874 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" podUID="94495808-8600-45f6-a7fe-74ded4bd3ec4" containerName="route-controller-manager" containerID="cri-o://6554a4338c0d8074ba1cafa349754420b2a43f0b1a801ee6abad4f7293d58bb5" gracePeriod=30 Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.640511 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:14 crc kubenswrapper[4950]: E0318 20:10:14.641979 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:15.141967645 +0000 UTC m=+228.382809513 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.716170 4950 patch_prober.go:28] interesting pod/router-default-5444994796-gsj4z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 20:10:14 crc kubenswrapper[4950]: [-]has-synced failed: reason withheld Mar 18 20:10:14 crc kubenswrapper[4950]: [+]process-running ok Mar 18 20:10:14 crc kubenswrapper[4950]: healthz check failed Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.716218 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsj4z" podUID="90ab43dd-38e6-4264-a086-5cef08691214" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.742588 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:14 crc kubenswrapper[4950]: E0318 20:10:14.742994 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:15.242978962 +0000 UTC m=+228.483820830 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.743065 4950 patch_prober.go:28] interesting pod/downloads-7954f5f757-rf8k5 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.743095 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-rf8k5" podUID="290d1b89-d5c9-4928-8973-6c33f476b091" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.743176 4950 patch_prober.go:28] interesting pod/downloads-7954f5f757-rf8k5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.743195 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rf8k5" podUID="290d1b89-d5c9-4928-8973-6c33f476b091" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.852238 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:14 crc kubenswrapper[4950]: E0318 20:10:14.852817 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:15.352804099 +0000 UTC m=+228.593645967 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.901517 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.902517 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.903398 4950 patch_prober.go:28] interesting pod/console-f9d7485db-46hjn container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.903446 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-46hjn" podUID="1cec2292-f508-42b2-916f-9a5808045626" containerName="console" probeResult="failure" output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" Mar 18 20:10:14 crc kubenswrapper[4950]: I0318 20:10:14.956833 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:14 crc kubenswrapper[4950]: E0318 20:10:14.957180 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:15.457166294 +0000 UTC m=+228.698008162 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.049838 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7dzz6"] Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.056018 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7dzz6" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.058881 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.059694 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:15 crc kubenswrapper[4950]: E0318 20:10:15.059972 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:15.559961068 +0000 UTC m=+228.800802936 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.068917 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7dzz6"] Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.143305 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fk4b9"] Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.152024 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"9dcd79684fb4ec6f645d1843f010ca760df31ba18ab61459fc638a2b1bb95fbc"} Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.152097 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.156585 4950 generic.go:334] "Generic (PLEG): container finished" podID="70601ef2-ab17-4550-890b-b29bf1a7b85e" containerID="6a6ac9f276afaa5327c38326539da8d4956e49fcf92a25f8ca48d1dfced47c93" exitCode=0 Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.156639 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" event={"ID":"70601ef2-ab17-4550-890b-b29bf1a7b85e","Type":"ContainerDied","Data":"6a6ac9f276afaa5327c38326539da8d4956e49fcf92a25f8ca48d1dfced47c93"} Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.161329 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:15 crc kubenswrapper[4950]: E0318 20:10:15.161568 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:15.66154067 +0000 UTC m=+228.902382538 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.161615 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.161722 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbeed1b2-6abb-49df-94d9-7d9fe6e5a563-utilities\") pod \"redhat-marketplace-7dzz6\" (UID: \"bbeed1b2-6abb-49df-94d9-7d9fe6e5a563\") " pod="openshift-marketplace/redhat-marketplace-7dzz6" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.161811 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9n22\" (UniqueName: \"kubernetes.io/projected/bbeed1b2-6abb-49df-94d9-7d9fe6e5a563-kube-api-access-x9n22\") pod \"redhat-marketplace-7dzz6\" (UID: \"bbeed1b2-6abb-49df-94d9-7d9fe6e5a563\") " pod="openshift-marketplace/redhat-marketplace-7dzz6" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.161852 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbeed1b2-6abb-49df-94d9-7d9fe6e5a563-catalog-content\") pod \"redhat-marketplace-7dzz6\" (UID: \"bbeed1b2-6abb-49df-94d9-7d9fe6e5a563\") " pod="openshift-marketplace/redhat-marketplace-7dzz6" Mar 18 20:10:15 crc kubenswrapper[4950]: E0318 20:10:15.162266 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:15.662258668 +0000 UTC m=+228.903100536 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.164465 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"2ee244d7bc73c65a5dfb290d48c2eeb9d7d7d475adc41031ce755018aa97a418"} Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.187958 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"05b7c80db07aaa213cdb6cdadf2ca1f69cc930055b9d5a5890258eb2eda26e2f"} Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.219119 4950 generic.go:334] "Generic (PLEG): container finished" podID="94495808-8600-45f6-a7fe-74ded4bd3ec4" containerID="6554a4338c0d8074ba1cafa349754420b2a43f0b1a801ee6abad4f7293d58bb5" exitCode=0 Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.219244 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" event={"ID":"94495808-8600-45f6-a7fe-74ded4bd3ec4","Type":"ContainerDied","Data":"6554a4338c0d8074ba1cafa349754420b2a43f0b1a801ee6abad4f7293d58bb5"} Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.280436 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" event={"ID":"0560ae4e-4142-48c1-9829-2e770a3b56b4","Type":"ContainerStarted","Data":"673f8ae93ba0b27de661f5715cf5a2d5ac036c0ebc7c2a6b1b3d9b2924fff53d"} Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.281556 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.282198 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbeed1b2-6abb-49df-94d9-7d9fe6e5a563-utilities\") pod \"redhat-marketplace-7dzz6\" (UID: \"bbeed1b2-6abb-49df-94d9-7d9fe6e5a563\") " pod="openshift-marketplace/redhat-marketplace-7dzz6" Mar 18 20:10:15 crc kubenswrapper[4950]: E0318 20:10:15.284404 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:15.784375537 +0000 UTC m=+229.025217405 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.287204 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9n22\" (UniqueName: \"kubernetes.io/projected/bbeed1b2-6abb-49df-94d9-7d9fe6e5a563-kube-api-access-x9n22\") pod \"redhat-marketplace-7dzz6\" (UID: \"bbeed1b2-6abb-49df-94d9-7d9fe6e5a563\") " pod="openshift-marketplace/redhat-marketplace-7dzz6" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.287325 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbeed1b2-6abb-49df-94d9-7d9fe6e5a563-catalog-content\") pod \"redhat-marketplace-7dzz6\" (UID: \"bbeed1b2-6abb-49df-94d9-7d9fe6e5a563\") " pod="openshift-marketplace/redhat-marketplace-7dzz6" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.285167 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbeed1b2-6abb-49df-94d9-7d9fe6e5a563-utilities\") pod \"redhat-marketplace-7dzz6\" (UID: \"bbeed1b2-6abb-49df-94d9-7d9fe6e5a563\") " pod="openshift-marketplace/redhat-marketplace-7dzz6" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.288862 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbeed1b2-6abb-49df-94d9-7d9fe6e5a563-catalog-content\") pod \"redhat-marketplace-7dzz6\" (UID: \"bbeed1b2-6abb-49df-94d9-7d9fe6e5a563\") " pod="openshift-marketplace/redhat-marketplace-7dzz6" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.330941 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8kwk7" event={"ID":"4a386d81-3e2c-4df7-988d-5fb1758be616","Type":"ContainerStarted","Data":"c1062526df43c854543930497f0effd3048ed2c3014191448f91af0d4dd5f4a5"} Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.345663 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9n22\" (UniqueName: \"kubernetes.io/projected/bbeed1b2-6abb-49df-94d9-7d9fe6e5a563-kube-api-access-x9n22\") pod \"redhat-marketplace-7dzz6\" (UID: \"bbeed1b2-6abb-49df-94d9-7d9fe6e5a563\") " pod="openshift-marketplace/redhat-marketplace-7dzz6" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.352628 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-tnbzn" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.384876 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7dzz6" Mar 18 20:10:15 crc kubenswrapper[4950]: E0318 20:10:15.390360 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:15.890342513 +0000 UTC m=+229.131184381 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.421646 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.462146 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rwzv9"] Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.478309 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lhhrh"] Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.479379 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lhhrh" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.517578 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jq2zs" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.526639 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:15 crc kubenswrapper[4950]: E0318 20:10:15.527612 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:16.027585826 +0000 UTC m=+229.268427694 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.616964 4950 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.628728 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lhhrh"] Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.638494 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fe40723-62a6-4fe7-989c-b6f968467b4c-catalog-content\") pod \"redhat-marketplace-lhhrh\" (UID: \"8fe40723-62a6-4fe7-989c-b6f968467b4c\") " pod="openshift-marketplace/redhat-marketplace-lhhrh" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.638541 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvtkc\" (UniqueName: \"kubernetes.io/projected/8fe40723-62a6-4fe7-989c-b6f968467b4c-kube-api-access-mvtkc\") pod \"redhat-marketplace-lhhrh\" (UID: \"8fe40723-62a6-4fe7-989c-b6f968467b4c\") " pod="openshift-marketplace/redhat-marketplace-lhhrh" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.638603 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fe40723-62a6-4fe7-989c-b6f968467b4c-utilities\") pod \"redhat-marketplace-lhhrh\" (UID: \"8fe40723-62a6-4fe7-989c-b6f968467b4c\") " pod="openshift-marketplace/redhat-marketplace-lhhrh" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.638728 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:15 crc kubenswrapper[4950]: E0318 20:10:15.639139 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:16.139121498 +0000 UTC m=+229.379963366 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.681955 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-gsj4z" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.688740 4950 patch_prober.go:28] interesting pod/router-default-5444994796-gsj4z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 20:10:15 crc kubenswrapper[4950]: [-]has-synced failed: reason withheld Mar 18 20:10:15 crc kubenswrapper[4950]: [+]process-running ok Mar 18 20:10:15 crc kubenswrapper[4950]: healthz check failed Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.688788 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsj4z" podUID="90ab43dd-38e6-4264-a086-5cef08691214" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.701055 4950 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-18T20:10:15.616988971Z","Handler":null,"Name":""} Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.718112 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.740557 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:15 crc kubenswrapper[4950]: E0318 20:10:15.740825 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-18 20:10:16.240793653 +0000 UTC m=+229.481635521 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.740862 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fe40723-62a6-4fe7-989c-b6f968467b4c-utilities\") pod \"redhat-marketplace-lhhrh\" (UID: \"8fe40723-62a6-4fe7-989c-b6f968467b4c\") " pod="openshift-marketplace/redhat-marketplace-lhhrh" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.741088 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.741142 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fe40723-62a6-4fe7-989c-b6f968467b4c-catalog-content\") pod \"redhat-marketplace-lhhrh\" (UID: \"8fe40723-62a6-4fe7-989c-b6f968467b4c\") " pod="openshift-marketplace/redhat-marketplace-lhhrh" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.741161 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvtkc\" (UniqueName: \"kubernetes.io/projected/8fe40723-62a6-4fe7-989c-b6f968467b4c-kube-api-access-mvtkc\") pod \"redhat-marketplace-lhhrh\" (UID: \"8fe40723-62a6-4fe7-989c-b6f968467b4c\") " pod="openshift-marketplace/redhat-marketplace-lhhrh" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.742592 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fe40723-62a6-4fe7-989c-b6f968467b4c-utilities\") pod \"redhat-marketplace-lhhrh\" (UID: \"8fe40723-62a6-4fe7-989c-b6f968467b4c\") " pod="openshift-marketplace/redhat-marketplace-lhhrh" Mar 18 20:10:15 crc kubenswrapper[4950]: E0318 20:10:15.744360 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-18 20:10:16.244345756 +0000 UTC m=+229.485187624 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2kz74" (UID: "28e264f3-49f6-406a-b700-213e196574c1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.745049 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fe40723-62a6-4fe7-989c-b6f968467b4c-catalog-content\") pod \"redhat-marketplace-lhhrh\" (UID: \"8fe40723-62a6-4fe7-989c-b6f968467b4c\") " pod="openshift-marketplace/redhat-marketplace-lhhrh" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.776200 4950 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.776255 4950 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.793771 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvtkc\" (UniqueName: \"kubernetes.io/projected/8fe40723-62a6-4fe7-989c-b6f968467b4c-kube-api-access-mvtkc\") pod \"redhat-marketplace-lhhrh\" (UID: \"8fe40723-62a6-4fe7-989c-b6f968467b4c\") " pod="openshift-marketplace/redhat-marketplace-lhhrh" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.836789 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lhhrh" Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.842062 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.859723 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gzj5q"] Mar 18 20:10:15 crc kubenswrapper[4950]: W0318 20:10:15.922777 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod044d9d2a_6b99_49ee_b00d_728c7832815c.slice/crio-d7f50bce278012a7d1417f1bf553c4a3468fa213298eb3369f211c89c4dc71f0 WatchSource:0}: Error finding container d7f50bce278012a7d1417f1bf553c4a3468fa213298eb3369f211c89c4dc71f0: Status 404 returned error can't find the container with id d7f50bce278012a7d1417f1bf553c4a3468fa213298eb3369f211c89c4dc71f0 Mar 18 20:10:15 crc kubenswrapper[4950]: I0318 20:10:15.933115 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.011404 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.050212 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94495808-8600-45f6-a7fe-74ded4bd3ec4-config\") pod \"94495808-8600-45f6-a7fe-74ded4bd3ec4\" (UID: \"94495808-8600-45f6-a7fe-74ded4bd3ec4\") " Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.050257 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6dxl\" (UniqueName: \"kubernetes.io/projected/94495808-8600-45f6-a7fe-74ded4bd3ec4-kube-api-access-f6dxl\") pod \"94495808-8600-45f6-a7fe-74ded4bd3ec4\" (UID: \"94495808-8600-45f6-a7fe-74ded4bd3ec4\") " Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.050325 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94495808-8600-45f6-a7fe-74ded4bd3ec4-serving-cert\") pod \"94495808-8600-45f6-a7fe-74ded4bd3ec4\" (UID: \"94495808-8600-45f6-a7fe-74ded4bd3ec4\") " Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.050350 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/94495808-8600-45f6-a7fe-74ded4bd3ec4-client-ca\") pod \"94495808-8600-45f6-a7fe-74ded4bd3ec4\" (UID: \"94495808-8600-45f6-a7fe-74ded4bd3ec4\") " Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.051309 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.056729 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94495808-8600-45f6-a7fe-74ded4bd3ec4-client-ca" (OuterVolumeSpecName: "client-ca") pod "94495808-8600-45f6-a7fe-74ded4bd3ec4" (UID: "94495808-8600-45f6-a7fe-74ded4bd3ec4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.058179 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94495808-8600-45f6-a7fe-74ded4bd3ec4-config" (OuterVolumeSpecName: "config") pod "94495808-8600-45f6-a7fe-74ded4bd3ec4" (UID: "94495808-8600-45f6-a7fe-74ded4bd3ec4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.066023 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94495808-8600-45f6-a7fe-74ded4bd3ec4-kube-api-access-f6dxl" (OuterVolumeSpecName: "kube-api-access-f6dxl") pod "94495808-8600-45f6-a7fe-74ded4bd3ec4" (UID: "94495808-8600-45f6-a7fe-74ded4bd3ec4"). InnerVolumeSpecName "kube-api-access-f6dxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.091345 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94495808-8600-45f6-a7fe-74ded4bd3ec4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "94495808-8600-45f6-a7fe-74ded4bd3ec4" (UID: "94495808-8600-45f6-a7fe-74ded4bd3ec4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.152482 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94495808-8600-45f6-a7fe-74ded4bd3ec4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.152526 4950 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/94495808-8600-45f6-a7fe-74ded4bd3ec4-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.152538 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94495808-8600-45f6-a7fe-74ded4bd3ec4-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.152549 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6dxl\" (UniqueName: \"kubernetes.io/projected/94495808-8600-45f6-a7fe-74ded4bd3ec4-kube-api-access-f6dxl\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.157032 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.202526 4950 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.202577 4950 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.280592 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nb7fx"] Mar 18 20:10:16 crc kubenswrapper[4950]: E0318 20:10:16.280788 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94495808-8600-45f6-a7fe-74ded4bd3ec4" containerName="route-controller-manager" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.280799 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="94495808-8600-45f6-a7fe-74ded4bd3ec4" containerName="route-controller-manager" Mar 18 20:10:16 crc kubenswrapper[4950]: E0318 20:10:16.280814 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70601ef2-ab17-4550-890b-b29bf1a7b85e" containerName="controller-manager" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.280820 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="70601ef2-ab17-4550-890b-b29bf1a7b85e" containerName="controller-manager" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.280907 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="94495808-8600-45f6-a7fe-74ded4bd3ec4" containerName="route-controller-manager" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.280922 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="70601ef2-ab17-4550-890b-b29bf1a7b85e" containerName="controller-manager" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.281739 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nb7fx" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.299867 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nb7fx"] Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.300573 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.357213 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msbwr\" (UniqueName: \"kubernetes.io/projected/70601ef2-ab17-4550-890b-b29bf1a7b85e-kube-api-access-msbwr\") pod \"70601ef2-ab17-4550-890b-b29bf1a7b85e\" (UID: \"70601ef2-ab17-4550-890b-b29bf1a7b85e\") " Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.357653 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/70601ef2-ab17-4550-890b-b29bf1a7b85e-proxy-ca-bundles\") pod \"70601ef2-ab17-4550-890b-b29bf1a7b85e\" (UID: \"70601ef2-ab17-4550-890b-b29bf1a7b85e\") " Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.357676 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/70601ef2-ab17-4550-890b-b29bf1a7b85e-client-ca\") pod \"70601ef2-ab17-4550-890b-b29bf1a7b85e\" (UID: \"70601ef2-ab17-4550-890b-b29bf1a7b85e\") " Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.357740 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70601ef2-ab17-4550-890b-b29bf1a7b85e-serving-cert\") pod \"70601ef2-ab17-4550-890b-b29bf1a7b85e\" (UID: \"70601ef2-ab17-4550-890b-b29bf1a7b85e\") " Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.357811 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70601ef2-ab17-4550-890b-b29bf1a7b85e-config\") pod \"70601ef2-ab17-4550-890b-b29bf1a7b85e\" (UID: \"70601ef2-ab17-4550-890b-b29bf1a7b85e\") " Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.359040 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70601ef2-ab17-4550-890b-b29bf1a7b85e-config" (OuterVolumeSpecName: "config") pod "70601ef2-ab17-4550-890b-b29bf1a7b85e" (UID: "70601ef2-ab17-4550-890b-b29bf1a7b85e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.363424 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70601ef2-ab17-4550-890b-b29bf1a7b85e-client-ca" (OuterVolumeSpecName: "client-ca") pod "70601ef2-ab17-4550-890b-b29bf1a7b85e" (UID: "70601ef2-ab17-4550-890b-b29bf1a7b85e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.364025 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70601ef2-ab17-4550-890b-b29bf1a7b85e-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "70601ef2-ab17-4550-890b-b29bf1a7b85e" (UID: "70601ef2-ab17-4550-890b-b29bf1a7b85e"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.434804 4950 generic.go:334] "Generic (PLEG): container finished" podID="f54c83ec-47ad-4ced-93c3-d321460f40f3" containerID="8948c35590cdaa069e6a2360b8e142f7c41951aa6653788206319998f9803392" exitCode=0 Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.434880 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fk4b9" event={"ID":"f54c83ec-47ad-4ced-93c3-d321460f40f3","Type":"ContainerDied","Data":"8948c35590cdaa069e6a2360b8e142f7c41951aa6653788206319998f9803392"} Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.434906 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fk4b9" event={"ID":"f54c83ec-47ad-4ced-93c3-d321460f40f3","Type":"ContainerStarted","Data":"27272da5fedb4e5ec9815272afeb74c757e15e0cd40694777c5c91cbdb93776e"} Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.438734 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70601ef2-ab17-4550-890b-b29bf1a7b85e-kube-api-access-msbwr" (OuterVolumeSpecName: "kube-api-access-msbwr") pod "70601ef2-ab17-4550-890b-b29bf1a7b85e" (UID: "70601ef2-ab17-4550-890b-b29bf1a7b85e"). InnerVolumeSpecName "kube-api-access-msbwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.446203 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70601ef2-ab17-4550-890b-b29bf1a7b85e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "70601ef2-ab17-4550-890b-b29bf1a7b85e" (UID: "70601ef2-ab17-4550-890b-b29bf1a7b85e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.459192 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33a99c2b-09d6-4ee7-b875-509ef5e144eb-catalog-content\") pod \"redhat-operators-nb7fx\" (UID: \"33a99c2b-09d6-4ee7-b875-509ef5e144eb\") " pod="openshift-marketplace/redhat-operators-nb7fx" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.459256 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7zmg\" (UniqueName: \"kubernetes.io/projected/33a99c2b-09d6-4ee7-b875-509ef5e144eb-kube-api-access-k7zmg\") pod \"redhat-operators-nb7fx\" (UID: \"33a99c2b-09d6-4ee7-b875-509ef5e144eb\") " pod="openshift-marketplace/redhat-operators-nb7fx" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.459279 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33a99c2b-09d6-4ee7-b875-509ef5e144eb-utilities\") pod \"redhat-operators-nb7fx\" (UID: \"33a99c2b-09d6-4ee7-b875-509ef5e144eb\") " pod="openshift-marketplace/redhat-operators-nb7fx" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.459328 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70601ef2-ab17-4550-890b-b29bf1a7b85e-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.459342 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msbwr\" (UniqueName: \"kubernetes.io/projected/70601ef2-ab17-4550-890b-b29bf1a7b85e-kube-api-access-msbwr\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.459355 4950 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/70601ef2-ab17-4550-890b-b29bf1a7b85e-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.459366 4950 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/70601ef2-ab17-4550-890b-b29bf1a7b85e-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.459377 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70601ef2-ab17-4550-890b-b29bf1a7b85e-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.499621 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" event={"ID":"70601ef2-ab17-4550-890b-b29bf1a7b85e","Type":"ContainerDied","Data":"c40fe5f787730ada42f53a058b4a24f96d9949f2e312bb23c66a12e3063b094d"} Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.499681 4950 scope.go:117] "RemoveContainer" containerID="6a6ac9f276afaa5327c38326539da8d4956e49fcf92a25f8ca48d1dfced47c93" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.499800 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9l9mq" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.516293 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2kz74\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.545987 4950 generic.go:334] "Generic (PLEG): container finished" podID="82c68303-8af4-41a0-af35-312cc8b0bd7e" containerID="b95e7bb85d3a41c87de48a23686cc15d691b7ba08e6e8739a1d514ef1506a4a4" exitCode=0 Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.546101 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwzv9" event={"ID":"82c68303-8af4-41a0-af35-312cc8b0bd7e","Type":"ContainerDied","Data":"b95e7bb85d3a41c87de48a23686cc15d691b7ba08e6e8739a1d514ef1506a4a4"} Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.546134 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwzv9" event={"ID":"82c68303-8af4-41a0-af35-312cc8b0bd7e","Type":"ContainerStarted","Data":"2adb8218254954fff85e83b554f2aeff06ce47ac7fa13d106beaaf6aa8dc1682"} Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.572331 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33a99c2b-09d6-4ee7-b875-509ef5e144eb-catalog-content\") pod \"redhat-operators-nb7fx\" (UID: \"33a99c2b-09d6-4ee7-b875-509ef5e144eb\") " pod="openshift-marketplace/redhat-operators-nb7fx" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.577804 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7zmg\" (UniqueName: \"kubernetes.io/projected/33a99c2b-09d6-4ee7-b875-509ef5e144eb-kube-api-access-k7zmg\") pod \"redhat-operators-nb7fx\" (UID: \"33a99c2b-09d6-4ee7-b875-509ef5e144eb\") " pod="openshift-marketplace/redhat-operators-nb7fx" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.577899 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33a99c2b-09d6-4ee7-b875-509ef5e144eb-utilities\") pod \"redhat-operators-nb7fx\" (UID: \"33a99c2b-09d6-4ee7-b875-509ef5e144eb\") " pod="openshift-marketplace/redhat-operators-nb7fx" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.578602 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gzj5q" event={"ID":"044d9d2a-6b99-49ee-b00d-728c7832815c","Type":"ContainerStarted","Data":"4cceb3b27880b43731e742f438b4a81d0515623897cd0916352fcfac501a9f70"} Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.578648 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gzj5q" event={"ID":"044d9d2a-6b99-49ee-b00d-728c7832815c","Type":"ContainerStarted","Data":"d7f50bce278012a7d1417f1bf553c4a3468fa213298eb3369f211c89c4dc71f0"} Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.578998 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33a99c2b-09d6-4ee7-b875-509ef5e144eb-catalog-content\") pod \"redhat-operators-nb7fx\" (UID: \"33a99c2b-09d6-4ee7-b875-509ef5e144eb\") " pod="openshift-marketplace/redhat-operators-nb7fx" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.581962 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33a99c2b-09d6-4ee7-b875-509ef5e144eb-utilities\") pod \"redhat-operators-nb7fx\" (UID: \"33a99c2b-09d6-4ee7-b875-509ef5e144eb\") " pod="openshift-marketplace/redhat-operators-nb7fx" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.619146 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7dzz6"] Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.619926 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" event={"ID":"94495808-8600-45f6-a7fe-74ded4bd3ec4","Type":"ContainerDied","Data":"01c005459aac84dc7f80dbdbc25307698b3d7c74eed4d546d239541914d5cbf7"} Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.619956 4950 scope.go:117] "RemoveContainer" containerID="6554a4338c0d8074ba1cafa349754420b2a43f0b1a801ee6abad4f7293d58bb5" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.620049 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.635398 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7zmg\" (UniqueName: \"kubernetes.io/projected/33a99c2b-09d6-4ee7-b875-509ef5e144eb-kube-api-access-k7zmg\") pod \"redhat-operators-nb7fx\" (UID: \"33a99c2b-09d6-4ee7-b875-509ef5e144eb\") " pod="openshift-marketplace/redhat-operators-nb7fx" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.637403 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nb7fx" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.642905 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.657724 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q"] Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.658715 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.660753 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9l9mq"] Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.667028 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9l9mq"] Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.669194 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.669987 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.670099 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.670196 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.670305 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.670447 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.673923 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" event={"ID":"0560ae4e-4142-48c1-9829-2e770a3b56b4","Type":"ContainerStarted","Data":"23264bced8363d4360628b999f85c643db438adf78f35136f5dc3912090f7209"} Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.677235 4950 generic.go:334] "Generic (PLEG): container finished" podID="4a386d81-3e2c-4df7-988d-5fb1758be616" containerID="83b4b8d10e11d03ddade158312b75de67ffc876388cad834fc3fcdff654df10c" exitCode=0 Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.677998 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8kwk7" event={"ID":"4a386d81-3e2c-4df7-988d-5fb1758be616","Type":"ContainerDied","Data":"83b4b8d10e11d03ddade158312b75de67ffc876388cad834fc3fcdff654df10c"} Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.679335 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b888d329-77fd-4fdd-bf49-276493b342b1-config\") pod \"route-controller-manager-7cf6998584-w986q\" (UID: \"b888d329-77fd-4fdd-bf49-276493b342b1\") " pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.679403 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b888d329-77fd-4fdd-bf49-276493b342b1-client-ca\") pod \"route-controller-manager-7cf6998584-w986q\" (UID: \"b888d329-77fd-4fdd-bf49-276493b342b1\") " pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.679449 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b888d329-77fd-4fdd-bf49-276493b342b1-serving-cert\") pod \"route-controller-manager-7cf6998584-w986q\" (UID: \"b888d329-77fd-4fdd-bf49-276493b342b1\") " pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.679482 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szbc2\" (UniqueName: \"kubernetes.io/projected/b888d329-77fd-4fdd-bf49-276493b342b1-kube-api-access-szbc2\") pod \"route-controller-manager-7cf6998584-w986q\" (UID: \"b888d329-77fd-4fdd-bf49-276493b342b1\") " pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.679536 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q"] Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.683154 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5fnpx"] Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.684789 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5fnpx" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.685315 4950 patch_prober.go:28] interesting pod/router-default-5444994796-gsj4z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 20:10:16 crc kubenswrapper[4950]: [-]has-synced failed: reason withheld Mar 18 20:10:16 crc kubenswrapper[4950]: [+]process-running ok Mar 18 20:10:16 crc kubenswrapper[4950]: healthz check failed Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.685366 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsj4z" podUID="90ab43dd-38e6-4264-a086-5cef08691214" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.722817 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5fnpx"] Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.781895 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b888d329-77fd-4fdd-bf49-276493b342b1-client-ca\") pod \"route-controller-manager-7cf6998584-w986q\" (UID: \"b888d329-77fd-4fdd-bf49-276493b342b1\") " pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.781980 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b888d329-77fd-4fdd-bf49-276493b342b1-serving-cert\") pod \"route-controller-manager-7cf6998584-w986q\" (UID: \"b888d329-77fd-4fdd-bf49-276493b342b1\") " pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.782016 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szbc2\" (UniqueName: \"kubernetes.io/projected/b888d329-77fd-4fdd-bf49-276493b342b1-kube-api-access-szbc2\") pod \"route-controller-manager-7cf6998584-w986q\" (UID: \"b888d329-77fd-4fdd-bf49-276493b342b1\") " pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.782047 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3ae2e86-be5d-4546-97a0-88301381868b-catalog-content\") pod \"redhat-operators-5fnpx\" (UID: \"d3ae2e86-be5d-4546-97a0-88301381868b\") " pod="openshift-marketplace/redhat-operators-5fnpx" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.782085 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3ae2e86-be5d-4546-97a0-88301381868b-utilities\") pod \"redhat-operators-5fnpx\" (UID: \"d3ae2e86-be5d-4546-97a0-88301381868b\") " pod="openshift-marketplace/redhat-operators-5fnpx" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.782109 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbj6d\" (UniqueName: \"kubernetes.io/projected/d3ae2e86-be5d-4546-97a0-88301381868b-kube-api-access-lbj6d\") pod \"redhat-operators-5fnpx\" (UID: \"d3ae2e86-be5d-4546-97a0-88301381868b\") " pod="openshift-marketplace/redhat-operators-5fnpx" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.782174 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b888d329-77fd-4fdd-bf49-276493b342b1-config\") pod \"route-controller-manager-7cf6998584-w986q\" (UID: \"b888d329-77fd-4fdd-bf49-276493b342b1\") " pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.783291 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b888d329-77fd-4fdd-bf49-276493b342b1-config\") pod \"route-controller-manager-7cf6998584-w986q\" (UID: \"b888d329-77fd-4fdd-bf49-276493b342b1\") " pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.785119 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b888d329-77fd-4fdd-bf49-276493b342b1-client-ca\") pod \"route-controller-manager-7cf6998584-w986q\" (UID: \"b888d329-77fd-4fdd-bf49-276493b342b1\") " pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.798916 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-qqrpb" podStartSLOduration=14.798897148 podStartE2EDuration="14.798897148s" podCreationTimestamp="2026-03-18 20:10:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:16.798196769 +0000 UTC m=+230.039038637" watchObservedRunningTime="2026-03-18 20:10:16.798897148 +0000 UTC m=+230.039739006" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.812749 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b888d329-77fd-4fdd-bf49-276493b342b1-serving-cert\") pod \"route-controller-manager-7cf6998584-w986q\" (UID: \"b888d329-77fd-4fdd-bf49-276493b342b1\") " pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.856491 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54"] Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.862343 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wcd54"] Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.867815 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szbc2\" (UniqueName: \"kubernetes.io/projected/b888d329-77fd-4fdd-bf49-276493b342b1-kube-api-access-szbc2\") pod \"route-controller-manager-7cf6998584-w986q\" (UID: \"b888d329-77fd-4fdd-bf49-276493b342b1\") " pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.883804 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3ae2e86-be5d-4546-97a0-88301381868b-catalog-content\") pod \"redhat-operators-5fnpx\" (UID: \"d3ae2e86-be5d-4546-97a0-88301381868b\") " pod="openshift-marketplace/redhat-operators-5fnpx" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.883871 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3ae2e86-be5d-4546-97a0-88301381868b-utilities\") pod \"redhat-operators-5fnpx\" (UID: \"d3ae2e86-be5d-4546-97a0-88301381868b\") " pod="openshift-marketplace/redhat-operators-5fnpx" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.883899 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbj6d\" (UniqueName: \"kubernetes.io/projected/d3ae2e86-be5d-4546-97a0-88301381868b-kube-api-access-lbj6d\") pod \"redhat-operators-5fnpx\" (UID: \"d3ae2e86-be5d-4546-97a0-88301381868b\") " pod="openshift-marketplace/redhat-operators-5fnpx" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.885611 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3ae2e86-be5d-4546-97a0-88301381868b-catalog-content\") pod \"redhat-operators-5fnpx\" (UID: \"d3ae2e86-be5d-4546-97a0-88301381868b\") " pod="openshift-marketplace/redhat-operators-5fnpx" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.886034 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3ae2e86-be5d-4546-97a0-88301381868b-utilities\") pod \"redhat-operators-5fnpx\" (UID: \"d3ae2e86-be5d-4546-97a0-88301381868b\") " pod="openshift-marketplace/redhat-operators-5fnpx" Mar 18 20:10:16 crc kubenswrapper[4950]: I0318 20:10:16.909207 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbj6d\" (UniqueName: \"kubernetes.io/projected/d3ae2e86-be5d-4546-97a0-88301381868b-kube-api-access-lbj6d\") pod \"redhat-operators-5fnpx\" (UID: \"d3ae2e86-be5d-4546-97a0-88301381868b\") " pod="openshift-marketplace/redhat-operators-5fnpx" Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.053254 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.085337 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5fnpx" Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.158735 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lhhrh"] Mar 18 20:10:17 crc kubenswrapper[4950]: W0318 20:10:17.205381 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8fe40723_62a6_4fe7_989c_b6f968467b4c.slice/crio-0d2d3868583c0aabbdcdc6edea1bf63a62c608cfce81db97a2e4d1849b25beb7 WatchSource:0}: Error finding container 0d2d3868583c0aabbdcdc6edea1bf63a62c608cfce81db97a2e4d1849b25beb7: Status 404 returned error can't find the container with id 0d2d3868583c0aabbdcdc6edea1bf63a62c608cfce81db97a2e4d1849b25beb7 Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.261835 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2kz74"] Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.266760 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nb7fx"] Mar 18 20:10:17 crc kubenswrapper[4950]: W0318 20:10:17.282383 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28e264f3_49f6_406a_b700_213e196574c1.slice/crio-4de76fd78aa150421888d2bdca5e4a9c772694329c1d994d47ff4b7d136ada6a WatchSource:0}: Error finding container 4de76fd78aa150421888d2bdca5e4a9c772694329c1d994d47ff4b7d136ada6a: Status 404 returned error can't find the container with id 4de76fd78aa150421888d2bdca5e4a9c772694329c1d994d47ff4b7d136ada6a Mar 18 20:10:17 crc kubenswrapper[4950]: W0318 20:10:17.326096 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33a99c2b_09d6_4ee7_b875_509ef5e144eb.slice/crio-2afa4c113494611ccc861e335d5e09baf952c0b3c4f5f984e300a8f5521a735b WatchSource:0}: Error finding container 2afa4c113494611ccc861e335d5e09baf952c0b3c4f5f984e300a8f5521a735b: Status 404 returned error can't find the container with id 2afa4c113494611ccc861e335d5e09baf952c0b3c4f5f984e300a8f5521a735b Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.496878 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70601ef2-ab17-4550-890b-b29bf1a7b85e" path="/var/lib/kubelet/pods/70601ef2-ab17-4550-890b-b29bf1a7b85e/volumes" Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.498300 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.498909 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94495808-8600-45f6-a7fe-74ded4bd3ec4" path="/var/lib/kubelet/pods/94495808-8600-45f6-a7fe-74ded4bd3ec4/volumes" Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.691392 4950 patch_prober.go:28] interesting pod/router-default-5444994796-gsj4z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 20:10:17 crc kubenswrapper[4950]: [-]has-synced failed: reason withheld Mar 18 20:10:17 crc kubenswrapper[4950]: [+]process-running ok Mar 18 20:10:17 crc kubenswrapper[4950]: healthz check failed Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.691544 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsj4z" podUID="90ab43dd-38e6-4264-a086-5cef08691214" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.709583 4950 generic.go:334] "Generic (PLEG): container finished" podID="044d9d2a-6b99-49ee-b00d-728c7832815c" containerID="4cceb3b27880b43731e742f438b4a81d0515623897cd0916352fcfac501a9f70" exitCode=0 Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.709686 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gzj5q" event={"ID":"044d9d2a-6b99-49ee-b00d-728c7832815c","Type":"ContainerDied","Data":"4cceb3b27880b43731e742f438b4a81d0515623897cd0916352fcfac501a9f70"} Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.719507 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lhhrh" event={"ID":"8fe40723-62a6-4fe7-989c-b6f968467b4c","Type":"ContainerStarted","Data":"0d2d3868583c0aabbdcdc6edea1bf63a62c608cfce81db97a2e4d1849b25beb7"} Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.725850 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nb7fx" event={"ID":"33a99c2b-09d6-4ee7-b875-509ef5e144eb","Type":"ContainerStarted","Data":"2afa4c113494611ccc861e335d5e09baf952c0b3c4f5f984e300a8f5521a735b"} Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.727843 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" event={"ID":"28e264f3-49f6-406a-b700-213e196574c1","Type":"ContainerStarted","Data":"4de76fd78aa150421888d2bdca5e4a9c772694329c1d994d47ff4b7d136ada6a"} Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.731478 4950 generic.go:334] "Generic (PLEG): container finished" podID="bbeed1b2-6abb-49df-94d9-7d9fe6e5a563" containerID="e73af9bce8b03e6b8ea69a8eb5cba1504a5c451c568cb4b6135277a0623070f3" exitCode=0 Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.731577 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7dzz6" event={"ID":"bbeed1b2-6abb-49df-94d9-7d9fe6e5a563","Type":"ContainerDied","Data":"e73af9bce8b03e6b8ea69a8eb5cba1504a5c451c568cb4b6135277a0623070f3"} Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.731653 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7dzz6" event={"ID":"bbeed1b2-6abb-49df-94d9-7d9fe6e5a563","Type":"ContainerStarted","Data":"26e4af3d58e72ec29192677ecc2533933ed679944d6e39fcec71ae83e8109102"} Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.736463 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5fnpx"] Mar 18 20:10:17 crc kubenswrapper[4950]: W0318 20:10:17.750699 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3ae2e86_be5d_4546_97a0_88301381868b.slice/crio-7666b6d629d01dd8ff0cea990e91a38418786e06ba35febd7d3ec855d70babc3 WatchSource:0}: Error finding container 7666b6d629d01dd8ff0cea990e91a38418786e06ba35febd7d3ec855d70babc3: Status 404 returned error can't find the container with id 7666b6d629d01dd8ff0cea990e91a38418786e06ba35febd7d3ec855d70babc3 Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.847242 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q"] Mar 18 20:10:17 crc kubenswrapper[4950]: W0318 20:10:17.862560 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb888d329_77fd_4fdd_bf49_276493b342b1.slice/crio-0103e51dbd663e60a384abd6aa6b262fa7fcda60042408d1ae928c4e65593d53 WatchSource:0}: Error finding container 0103e51dbd663e60a384abd6aa6b262fa7fcda60042408d1ae928c4e65593d53: Status 404 returned error can't find the container with id 0103e51dbd663e60a384abd6aa6b262fa7fcda60042408d1ae928c4e65593d53 Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.890460 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.891153 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.891891 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.902078 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 18 20:10:17 crc kubenswrapper[4950]: I0318 20:10:17.902261 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.022223 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/01bd704c-cabf-438e-8deb-1fb33a59e95b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"01bd704c-cabf-438e-8deb-1fb33a59e95b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.022295 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01bd704c-cabf-438e-8deb-1fb33a59e95b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"01bd704c-cabf-438e-8deb-1fb33a59e95b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.129555 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs\") pod \"network-metrics-daemon-l2cjn\" (UID: \"d09a228c-313f-4b55-a298-51ebd404cfd5\") " pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.129695 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01bd704c-cabf-438e-8deb-1fb33a59e95b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"01bd704c-cabf-438e-8deb-1fb33a59e95b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.129836 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/01bd704c-cabf-438e-8deb-1fb33a59e95b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"01bd704c-cabf-438e-8deb-1fb33a59e95b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.130373 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/01bd704c-cabf-438e-8deb-1fb33a59e95b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"01bd704c-cabf-438e-8deb-1fb33a59e95b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.138522 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d09a228c-313f-4b55-a298-51ebd404cfd5-metrics-certs\") pod \"network-metrics-daemon-l2cjn\" (UID: \"d09a228c-313f-4b55-a298-51ebd404cfd5\") " pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.180132 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l2cjn" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.186024 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01bd704c-cabf-438e-8deb-1fb33a59e95b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"01bd704c-cabf-438e-8deb-1fb33a59e95b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.236879 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.241867 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.262802 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.263445 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.264184 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.282225 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.435656 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a2a617fa-97f4-4c7c-bc20-6234cc29e79a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a2a617fa-97f4-4c7c-bc20-6234cc29e79a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.436264 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a2a617fa-97f4-4c7c-bc20-6234cc29e79a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a2a617fa-97f4-4c7c-bc20-6234cc29e79a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.537085 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a2a617fa-97f4-4c7c-bc20-6234cc29e79a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a2a617fa-97f4-4c7c-bc20-6234cc29e79a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.537148 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a2a617fa-97f4-4c7c-bc20-6234cc29e79a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a2a617fa-97f4-4c7c-bc20-6234cc29e79a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.537293 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a2a617fa-97f4-4c7c-bc20-6234cc29e79a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a2a617fa-97f4-4c7c-bc20-6234cc29e79a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.571382 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a2a617fa-97f4-4c7c-bc20-6234cc29e79a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a2a617fa-97f4-4c7c-bc20-6234cc29e79a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.609335 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.636822 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6cc4b56544-l9lgg"] Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.639602 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.645706 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.645950 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.646096 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.646335 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.646483 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.646654 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.657214 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.669818 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6cc4b56544-l9lgg"] Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.688300 4950 patch_prober.go:28] interesting pod/router-default-5444994796-gsj4z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 20:10:18 crc kubenswrapper[4950]: [-]has-synced failed: reason withheld Mar 18 20:10:18 crc kubenswrapper[4950]: [+]process-running ok Mar 18 20:10:18 crc kubenswrapper[4950]: healthz check failed Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.688458 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsj4z" podUID="90ab43dd-38e6-4264-a086-5cef08691214" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.775633 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-l2cjn"] Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.805060 4950 ???:1] "http: TLS handshake error from 192.168.126.11:34522: no serving certificate available for the kubelet" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.810177 4950 generic.go:334] "Generic (PLEG): container finished" podID="d3ae2e86-be5d-4546-97a0-88301381868b" containerID="69309d7fec17ae4fbdf231e4dac66b9f7c29ca46eedc368111f5776bcd93cf45" exitCode=0 Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.810697 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fnpx" event={"ID":"d3ae2e86-be5d-4546-97a0-88301381868b","Type":"ContainerDied","Data":"69309d7fec17ae4fbdf231e4dac66b9f7c29ca46eedc368111f5776bcd93cf45"} Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.810814 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fnpx" event={"ID":"d3ae2e86-be5d-4546-97a0-88301381868b","Type":"ContainerStarted","Data":"7666b6d629d01dd8ff0cea990e91a38418786e06ba35febd7d3ec855d70babc3"} Mar 18 20:10:18 crc kubenswrapper[4950]: W0318 20:10:18.825803 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd09a228c_313f_4b55_a298_51ebd404cfd5.slice/crio-491de1fc49d533a8e5db2ce74007622c5a68c3a345d3456f9da59008ae516825 WatchSource:0}: Error finding container 491de1fc49d533a8e5db2ce74007622c5a68c3a345d3456f9da59008ae516825: Status 404 returned error can't find the container with id 491de1fc49d533a8e5db2ce74007622c5a68c3a345d3456f9da59008ae516825 Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.842454 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" event={"ID":"28e264f3-49f6-406a-b700-213e196574c1","Type":"ContainerStarted","Data":"3ceb89f25f617ed8ec72818a52205813cc4d8914af7fdcfad00cf1e9365d94ee"} Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.844064 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/19f3cffc-3f5f-4166-b282-cc1f352fde48-proxy-ca-bundles\") pod \"controller-manager-6cc4b56544-l9lgg\" (UID: \"19f3cffc-3f5f-4166-b282-cc1f352fde48\") " pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.846334 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/19f3cffc-3f5f-4166-b282-cc1f352fde48-client-ca\") pod \"controller-manager-6cc4b56544-l9lgg\" (UID: \"19f3cffc-3f5f-4166-b282-cc1f352fde48\") " pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.846380 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n2wn\" (UniqueName: \"kubernetes.io/projected/19f3cffc-3f5f-4166-b282-cc1f352fde48-kube-api-access-4n2wn\") pod \"controller-manager-6cc4b56544-l9lgg\" (UID: \"19f3cffc-3f5f-4166-b282-cc1f352fde48\") " pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.846451 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19f3cffc-3f5f-4166-b282-cc1f352fde48-config\") pod \"controller-manager-6cc4b56544-l9lgg\" (UID: \"19f3cffc-3f5f-4166-b282-cc1f352fde48\") " pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.846494 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19f3cffc-3f5f-4166-b282-cc1f352fde48-serving-cert\") pod \"controller-manager-6cc4b56544-l9lgg\" (UID: \"19f3cffc-3f5f-4166-b282-cc1f352fde48\") " pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.845220 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.859226 4950 generic.go:334] "Generic (PLEG): container finished" podID="8fe40723-62a6-4fe7-989c-b6f968467b4c" containerID="066cc63479a5ddbb896f736060d5c98157b3f0cc00f596abd47b8bf80f182383" exitCode=0 Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.859324 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lhhrh" event={"ID":"8fe40723-62a6-4fe7-989c-b6f968467b4c","Type":"ContainerDied","Data":"066cc63479a5ddbb896f736060d5c98157b3f0cc00f596abd47b8bf80f182383"} Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.895784 4950 generic.go:334] "Generic (PLEG): container finished" podID="33a99c2b-09d6-4ee7-b875-509ef5e144eb" containerID="14fc2e56f2022499c8bde7db5aa8430bf11cae9929d29b3e716e2ca6a0a8217c" exitCode=0 Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.896072 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nb7fx" event={"ID":"33a99c2b-09d6-4ee7-b875-509ef5e144eb","Type":"ContainerDied","Data":"14fc2e56f2022499c8bde7db5aa8430bf11cae9929d29b3e716e2ca6a0a8217c"} Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.902695 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" podStartSLOduration=183.902668432 podStartE2EDuration="3m3.902668432s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:18.883725918 +0000 UTC m=+232.124567786" watchObservedRunningTime="2026-03-18 20:10:18.902668432 +0000 UTC m=+232.143510300" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.940234 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" event={"ID":"b888d329-77fd-4fdd-bf49-276493b342b1","Type":"ContainerStarted","Data":"b2ffdb35fd9f78eff9a5838150733a90ae51886225a026b7b2d46c0aa05e5dfe"} Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.940315 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" event={"ID":"b888d329-77fd-4fdd-bf49-276493b342b1","Type":"ContainerStarted","Data":"0103e51dbd663e60a384abd6aa6b262fa7fcda60042408d1ae928c4e65593d53"} Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.941732 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.943466 4950 ???:1] "http: TLS handshake error from 192.168.126.11:34526: no serving certificate available for the kubelet" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.954762 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/19f3cffc-3f5f-4166-b282-cc1f352fde48-proxy-ca-bundles\") pod \"controller-manager-6cc4b56544-l9lgg\" (UID: \"19f3cffc-3f5f-4166-b282-cc1f352fde48\") " pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.962394 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/19f3cffc-3f5f-4166-b282-cc1f352fde48-proxy-ca-bundles\") pod \"controller-manager-6cc4b56544-l9lgg\" (UID: \"19f3cffc-3f5f-4166-b282-cc1f352fde48\") " pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.968374 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/19f3cffc-3f5f-4166-b282-cc1f352fde48-client-ca\") pod \"controller-manager-6cc4b56544-l9lgg\" (UID: \"19f3cffc-3f5f-4166-b282-cc1f352fde48\") " pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.968511 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n2wn\" (UniqueName: \"kubernetes.io/projected/19f3cffc-3f5f-4166-b282-cc1f352fde48-kube-api-access-4n2wn\") pod \"controller-manager-6cc4b56544-l9lgg\" (UID: \"19f3cffc-3f5f-4166-b282-cc1f352fde48\") " pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.968669 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19f3cffc-3f5f-4166-b282-cc1f352fde48-config\") pod \"controller-manager-6cc4b56544-l9lgg\" (UID: \"19f3cffc-3f5f-4166-b282-cc1f352fde48\") " pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.968753 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19f3cffc-3f5f-4166-b282-cc1f352fde48-serving-cert\") pod \"controller-manager-6cc4b56544-l9lgg\" (UID: \"19f3cffc-3f5f-4166-b282-cc1f352fde48\") " pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.970472 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/19f3cffc-3f5f-4166-b282-cc1f352fde48-client-ca\") pod \"controller-manager-6cc4b56544-l9lgg\" (UID: \"19f3cffc-3f5f-4166-b282-cc1f352fde48\") " pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.972075 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19f3cffc-3f5f-4166-b282-cc1f352fde48-config\") pod \"controller-manager-6cc4b56544-l9lgg\" (UID: \"19f3cffc-3f5f-4166-b282-cc1f352fde48\") " pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" Mar 18 20:10:18 crc kubenswrapper[4950]: I0318 20:10:18.978971 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" Mar 18 20:10:19 crc kubenswrapper[4950]: I0318 20:10:19.001574 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n2wn\" (UniqueName: \"kubernetes.io/projected/19f3cffc-3f5f-4166-b282-cc1f352fde48-kube-api-access-4n2wn\") pod \"controller-manager-6cc4b56544-l9lgg\" (UID: \"19f3cffc-3f5f-4166-b282-cc1f352fde48\") " pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" Mar 18 20:10:19 crc kubenswrapper[4950]: I0318 20:10:19.008440 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19f3cffc-3f5f-4166-b282-cc1f352fde48-serving-cert\") pod \"controller-manager-6cc4b56544-l9lgg\" (UID: \"19f3cffc-3f5f-4166-b282-cc1f352fde48\") " pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" Mar 18 20:10:19 crc kubenswrapper[4950]: I0318 20:10:19.058873 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" podStartSLOduration=4.05885957 podStartE2EDuration="4.05885957s" podCreationTimestamp="2026-03-18 20:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:19.057763301 +0000 UTC m=+232.298605169" watchObservedRunningTime="2026-03-18 20:10:19.05885957 +0000 UTC m=+232.299701438" Mar 18 20:10:19 crc kubenswrapper[4950]: I0318 20:10:19.113625 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:19 crc kubenswrapper[4950]: I0318 20:10:19.136202 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-27jdq" Mar 18 20:10:19 crc kubenswrapper[4950]: I0318 20:10:19.281892 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" Mar 18 20:10:19 crc kubenswrapper[4950]: I0318 20:10:19.352058 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 18 20:10:19 crc kubenswrapper[4950]: I0318 20:10:19.510685 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 18 20:10:19 crc kubenswrapper[4950]: W0318 20:10:19.539546 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod01bd704c_cabf_438e_8deb_1fb33a59e95b.slice/crio-236a0fdc6ff92fd1d3407521f9bb86ab56036db5f22dd0f7125a3dd77e5221c6 WatchSource:0}: Error finding container 236a0fdc6ff92fd1d3407521f9bb86ab56036db5f22dd0f7125a3dd77e5221c6: Status 404 returned error can't find the container with id 236a0fdc6ff92fd1d3407521f9bb86ab56036db5f22dd0f7125a3dd77e5221c6 Mar 18 20:10:19 crc kubenswrapper[4950]: I0318 20:10:19.685081 4950 patch_prober.go:28] interesting pod/router-default-5444994796-gsj4z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 20:10:19 crc kubenswrapper[4950]: [-]has-synced failed: reason withheld Mar 18 20:10:19 crc kubenswrapper[4950]: [+]process-running ok Mar 18 20:10:19 crc kubenswrapper[4950]: healthz check failed Mar 18 20:10:19 crc kubenswrapper[4950]: I0318 20:10:19.685585 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsj4z" podUID="90ab43dd-38e6-4264-a086-5cef08691214" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 20:10:19 crc kubenswrapper[4950]: I0318 20:10:19.974705 4950 generic.go:334] "Generic (PLEG): container finished" podID="164efb7c-0b83-42ed-8d30-acc8a1ff1794" containerID="25e62c249dbad8da15635226eb6e8947d89177ad45c9ad3022cf2b7827bfcaff" exitCode=0 Mar 18 20:10:19 crc kubenswrapper[4950]: I0318 20:10:19.974946 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564400-d9vqv" event={"ID":"164efb7c-0b83-42ed-8d30-acc8a1ff1794","Type":"ContainerDied","Data":"25e62c249dbad8da15635226eb6e8947d89177ad45c9ad3022cf2b7827bfcaff"} Mar 18 20:10:20 crc kubenswrapper[4950]: I0318 20:10:20.008110 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6cc4b56544-l9lgg"] Mar 18 20:10:20 crc kubenswrapper[4950]: I0318 20:10:20.021049 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" event={"ID":"d09a228c-313f-4b55-a298-51ebd404cfd5","Type":"ContainerStarted","Data":"749582399dee3b6b75f202a36fb8a021cc7fca7bbacd079fa96f08ab035b80c9"} Mar 18 20:10:20 crc kubenswrapper[4950]: I0318 20:10:20.021085 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" event={"ID":"d09a228c-313f-4b55-a298-51ebd404cfd5","Type":"ContainerStarted","Data":"491de1fc49d533a8e5db2ce74007622c5a68c3a345d3456f9da59008ae516825"} Mar 18 20:10:20 crc kubenswrapper[4950]: I0318 20:10:20.107821 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"01bd704c-cabf-438e-8deb-1fb33a59e95b","Type":"ContainerStarted","Data":"236a0fdc6ff92fd1d3407521f9bb86ab56036db5f22dd0f7125a3dd77e5221c6"} Mar 18 20:10:20 crc kubenswrapper[4950]: I0318 20:10:20.130604 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a2a617fa-97f4-4c7c-bc20-6234cc29e79a","Type":"ContainerStarted","Data":"380cdd2e7341221f2a046e2d21b0e9dd18174740ae44a32642df6d062eaf0290"} Mar 18 20:10:20 crc kubenswrapper[4950]: I0318 20:10:20.556084 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-cdpgz" Mar 18 20:10:20 crc kubenswrapper[4950]: I0318 20:10:20.684435 4950 patch_prober.go:28] interesting pod/router-default-5444994796-gsj4z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 20:10:20 crc kubenswrapper[4950]: [-]has-synced failed: reason withheld Mar 18 20:10:20 crc kubenswrapper[4950]: [+]process-running ok Mar 18 20:10:20 crc kubenswrapper[4950]: healthz check failed Mar 18 20:10:20 crc kubenswrapper[4950]: I0318 20:10:20.684511 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsj4z" podUID="90ab43dd-38e6-4264-a086-5cef08691214" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 20:10:21 crc kubenswrapper[4950]: I0318 20:10:21.224052 4950 generic.go:334] "Generic (PLEG): container finished" podID="01bd704c-cabf-438e-8deb-1fb33a59e95b" containerID="e473552c0aadfbfe9cfc0fe2e9edeb9d5bae3cd7d805b9eca7a0da557e2e9406" exitCode=0 Mar 18 20:10:21 crc kubenswrapper[4950]: I0318 20:10:21.224174 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"01bd704c-cabf-438e-8deb-1fb33a59e95b","Type":"ContainerDied","Data":"e473552c0aadfbfe9cfc0fe2e9edeb9d5bae3cd7d805b9eca7a0da557e2e9406"} Mar 18 20:10:21 crc kubenswrapper[4950]: I0318 20:10:21.234269 4950 generic.go:334] "Generic (PLEG): container finished" podID="a2a617fa-97f4-4c7c-bc20-6234cc29e79a" containerID="e0136c543df50d9fd16c571befabaa017c45e412ae707a9b16a5cba56801d653" exitCode=0 Mar 18 20:10:21 crc kubenswrapper[4950]: I0318 20:10:21.234347 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a2a617fa-97f4-4c7c-bc20-6234cc29e79a","Type":"ContainerDied","Data":"e0136c543df50d9fd16c571befabaa017c45e412ae707a9b16a5cba56801d653"} Mar 18 20:10:21 crc kubenswrapper[4950]: I0318 20:10:21.266514 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" event={"ID":"19f3cffc-3f5f-4166-b282-cc1f352fde48","Type":"ContainerStarted","Data":"64cf7de40223a2111e01f2c5dd2405deb1c51e5d0e6b1e65dca597ff2c235e62"} Mar 18 20:10:21 crc kubenswrapper[4950]: I0318 20:10:21.266605 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" event={"ID":"19f3cffc-3f5f-4166-b282-cc1f352fde48","Type":"ContainerStarted","Data":"f0ac3d91d6d9cae262049da4cb96530c59baec099614b9f14af0935d87340c49"} Mar 18 20:10:21 crc kubenswrapper[4950]: I0318 20:10:21.270044 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" Mar 18 20:10:21 crc kubenswrapper[4950]: I0318 20:10:21.283693 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" Mar 18 20:10:21 crc kubenswrapper[4950]: I0318 20:10:21.294582 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-l2cjn" event={"ID":"d09a228c-313f-4b55-a298-51ebd404cfd5","Type":"ContainerStarted","Data":"d82aebbf1dd808bba2b5a08d358316cbf5b197dae3d3dd195d50eed3f5c9344c"} Mar 18 20:10:21 crc kubenswrapper[4950]: I0318 20:10:21.304506 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" podStartSLOduration=7.304471967 podStartE2EDuration="7.304471967s" podCreationTimestamp="2026-03-18 20:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:21.295833892 +0000 UTC m=+234.536675770" watchObservedRunningTime="2026-03-18 20:10:21.304471967 +0000 UTC m=+234.545313835" Mar 18 20:10:21 crc kubenswrapper[4950]: I0318 20:10:21.685240 4950 patch_prober.go:28] interesting pod/router-default-5444994796-gsj4z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 20:10:21 crc kubenswrapper[4950]: [-]has-synced failed: reason withheld Mar 18 20:10:21 crc kubenswrapper[4950]: [+]process-running ok Mar 18 20:10:21 crc kubenswrapper[4950]: healthz check failed Mar 18 20:10:21 crc kubenswrapper[4950]: I0318 20:10:21.685300 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsj4z" podUID="90ab43dd-38e6-4264-a086-5cef08691214" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 20:10:21 crc kubenswrapper[4950]: I0318 20:10:21.751867 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564400-d9vqv" Mar 18 20:10:21 crc kubenswrapper[4950]: I0318 20:10:21.774680 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-l2cjn" podStartSLOduration=186.774624722 podStartE2EDuration="3m6.774624722s" podCreationTimestamp="2026-03-18 20:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:10:21.372548745 +0000 UTC m=+234.613390613" watchObservedRunningTime="2026-03-18 20:10:21.774624722 +0000 UTC m=+235.015466590" Mar 18 20:10:21 crc kubenswrapper[4950]: I0318 20:10:21.869840 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/164efb7c-0b83-42ed-8d30-acc8a1ff1794-config-volume\") pod \"164efb7c-0b83-42ed-8d30-acc8a1ff1794\" (UID: \"164efb7c-0b83-42ed-8d30-acc8a1ff1794\") " Mar 18 20:10:21 crc kubenswrapper[4950]: I0318 20:10:21.869933 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rprg\" (UniqueName: \"kubernetes.io/projected/164efb7c-0b83-42ed-8d30-acc8a1ff1794-kube-api-access-2rprg\") pod \"164efb7c-0b83-42ed-8d30-acc8a1ff1794\" (UID: \"164efb7c-0b83-42ed-8d30-acc8a1ff1794\") " Mar 18 20:10:21 crc kubenswrapper[4950]: I0318 20:10:21.870286 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/164efb7c-0b83-42ed-8d30-acc8a1ff1794-secret-volume\") pod \"164efb7c-0b83-42ed-8d30-acc8a1ff1794\" (UID: \"164efb7c-0b83-42ed-8d30-acc8a1ff1794\") " Mar 18 20:10:21 crc kubenswrapper[4950]: I0318 20:10:21.873854 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/164efb7c-0b83-42ed-8d30-acc8a1ff1794-config-volume" (OuterVolumeSpecName: "config-volume") pod "164efb7c-0b83-42ed-8d30-acc8a1ff1794" (UID: "164efb7c-0b83-42ed-8d30-acc8a1ff1794"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:10:21 crc kubenswrapper[4950]: I0318 20:10:21.884140 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/164efb7c-0b83-42ed-8d30-acc8a1ff1794-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "164efb7c-0b83-42ed-8d30-acc8a1ff1794" (UID: "164efb7c-0b83-42ed-8d30-acc8a1ff1794"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:10:21 crc kubenswrapper[4950]: I0318 20:10:21.885123 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/164efb7c-0b83-42ed-8d30-acc8a1ff1794-kube-api-access-2rprg" (OuterVolumeSpecName: "kube-api-access-2rprg") pod "164efb7c-0b83-42ed-8d30-acc8a1ff1794" (UID: "164efb7c-0b83-42ed-8d30-acc8a1ff1794"). InnerVolumeSpecName "kube-api-access-2rprg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:10:21 crc kubenswrapper[4950]: I0318 20:10:21.974108 4950 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/164efb7c-0b83-42ed-8d30-acc8a1ff1794-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:21 crc kubenswrapper[4950]: I0318 20:10:21.974279 4950 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/164efb7c-0b83-42ed-8d30-acc8a1ff1794-config-volume\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:21 crc kubenswrapper[4950]: I0318 20:10:21.974293 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rprg\" (UniqueName: \"kubernetes.io/projected/164efb7c-0b83-42ed-8d30-acc8a1ff1794-kube-api-access-2rprg\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:22 crc kubenswrapper[4950]: I0318 20:10:22.360237 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564400-d9vqv" Mar 18 20:10:22 crc kubenswrapper[4950]: I0318 20:10:22.361458 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564400-d9vqv" event={"ID":"164efb7c-0b83-42ed-8d30-acc8a1ff1794","Type":"ContainerDied","Data":"bbefef7236c11b25684d8d2d54cfe0434ffc226bf5beaa23de40c535693387cc"} Mar 18 20:10:22 crc kubenswrapper[4950]: I0318 20:10:22.361519 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbefef7236c11b25684d8d2d54cfe0434ffc226bf5beaa23de40c535693387cc" Mar 18 20:10:22 crc kubenswrapper[4950]: I0318 20:10:22.690462 4950 patch_prober.go:28] interesting pod/router-default-5444994796-gsj4z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 20:10:22 crc kubenswrapper[4950]: [-]has-synced failed: reason withheld Mar 18 20:10:22 crc kubenswrapper[4950]: [+]process-running ok Mar 18 20:10:22 crc kubenswrapper[4950]: healthz check failed Mar 18 20:10:22 crc kubenswrapper[4950]: I0318 20:10:22.693069 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsj4z" podUID="90ab43dd-38e6-4264-a086-5cef08691214" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 20:10:22 crc kubenswrapper[4950]: I0318 20:10:22.835473 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 20:10:22 crc kubenswrapper[4950]: I0318 20:10:22.893224 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01bd704c-cabf-438e-8deb-1fb33a59e95b-kube-api-access\") pod \"01bd704c-cabf-438e-8deb-1fb33a59e95b\" (UID: \"01bd704c-cabf-438e-8deb-1fb33a59e95b\") " Mar 18 20:10:22 crc kubenswrapper[4950]: I0318 20:10:22.893268 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/01bd704c-cabf-438e-8deb-1fb33a59e95b-kubelet-dir\") pod \"01bd704c-cabf-438e-8deb-1fb33a59e95b\" (UID: \"01bd704c-cabf-438e-8deb-1fb33a59e95b\") " Mar 18 20:10:22 crc kubenswrapper[4950]: I0318 20:10:22.893548 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/01bd704c-cabf-438e-8deb-1fb33a59e95b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "01bd704c-cabf-438e-8deb-1fb33a59e95b" (UID: "01bd704c-cabf-438e-8deb-1fb33a59e95b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:10:22 crc kubenswrapper[4950]: I0318 20:10:22.922786 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01bd704c-cabf-438e-8deb-1fb33a59e95b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "01bd704c-cabf-438e-8deb-1fb33a59e95b" (UID: "01bd704c-cabf-438e-8deb-1fb33a59e95b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:10:22 crc kubenswrapper[4950]: I0318 20:10:22.952734 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 20:10:22 crc kubenswrapper[4950]: I0318 20:10:22.994073 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a2a617fa-97f4-4c7c-bc20-6234cc29e79a-kubelet-dir\") pod \"a2a617fa-97f4-4c7c-bc20-6234cc29e79a\" (UID: \"a2a617fa-97f4-4c7c-bc20-6234cc29e79a\") " Mar 18 20:10:22 crc kubenswrapper[4950]: I0318 20:10:22.994129 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a2a617fa-97f4-4c7c-bc20-6234cc29e79a-kube-api-access\") pod \"a2a617fa-97f4-4c7c-bc20-6234cc29e79a\" (UID: \"a2a617fa-97f4-4c7c-bc20-6234cc29e79a\") " Mar 18 20:10:22 crc kubenswrapper[4950]: I0318 20:10:22.994310 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01bd704c-cabf-438e-8deb-1fb33a59e95b-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:22 crc kubenswrapper[4950]: I0318 20:10:22.994322 4950 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/01bd704c-cabf-438e-8deb-1fb33a59e95b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:22 crc kubenswrapper[4950]: I0318 20:10:22.994883 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a2a617fa-97f4-4c7c-bc20-6234cc29e79a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a2a617fa-97f4-4c7c-bc20-6234cc29e79a" (UID: "a2a617fa-97f4-4c7c-bc20-6234cc29e79a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:10:22 crc kubenswrapper[4950]: I0318 20:10:22.998202 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2a617fa-97f4-4c7c-bc20-6234cc29e79a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a2a617fa-97f4-4c7c-bc20-6234cc29e79a" (UID: "a2a617fa-97f4-4c7c-bc20-6234cc29e79a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:10:23 crc kubenswrapper[4950]: I0318 20:10:23.095079 4950 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a2a617fa-97f4-4c7c-bc20-6234cc29e79a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:23 crc kubenswrapper[4950]: I0318 20:10:23.095112 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a2a617fa-97f4-4c7c-bc20-6234cc29e79a-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:23 crc kubenswrapper[4950]: I0318 20:10:23.417460 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"01bd704c-cabf-438e-8deb-1fb33a59e95b","Type":"ContainerDied","Data":"236a0fdc6ff92fd1d3407521f9bb86ab56036db5f22dd0f7125a3dd77e5221c6"} Mar 18 20:10:23 crc kubenswrapper[4950]: I0318 20:10:23.417717 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="236a0fdc6ff92fd1d3407521f9bb86ab56036db5f22dd0f7125a3dd77e5221c6" Mar 18 20:10:23 crc kubenswrapper[4950]: I0318 20:10:23.417553 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 18 20:10:23 crc kubenswrapper[4950]: I0318 20:10:23.424508 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a2a617fa-97f4-4c7c-bc20-6234cc29e79a","Type":"ContainerDied","Data":"380cdd2e7341221f2a046e2d21b0e9dd18174740ae44a32642df6d062eaf0290"} Mar 18 20:10:23 crc kubenswrapper[4950]: I0318 20:10:23.426012 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="380cdd2e7341221f2a046e2d21b0e9dd18174740ae44a32642df6d062eaf0290" Mar 18 20:10:23 crc kubenswrapper[4950]: I0318 20:10:23.425471 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 18 20:10:23 crc kubenswrapper[4950]: I0318 20:10:23.684519 4950 patch_prober.go:28] interesting pod/router-default-5444994796-gsj4z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 20:10:23 crc kubenswrapper[4950]: [-]has-synced failed: reason withheld Mar 18 20:10:23 crc kubenswrapper[4950]: [+]process-running ok Mar 18 20:10:23 crc kubenswrapper[4950]: healthz check failed Mar 18 20:10:23 crc kubenswrapper[4950]: I0318 20:10:23.684575 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsj4z" podUID="90ab43dd-38e6-4264-a086-5cef08691214" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 20:10:24 crc kubenswrapper[4950]: I0318 20:10:24.682690 4950 patch_prober.go:28] interesting pod/router-default-5444994796-gsj4z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 20:10:24 crc kubenswrapper[4950]: [-]has-synced failed: reason withheld Mar 18 20:10:24 crc kubenswrapper[4950]: [+]process-running ok Mar 18 20:10:24 crc kubenswrapper[4950]: healthz check failed Mar 18 20:10:24 crc kubenswrapper[4950]: I0318 20:10:24.682761 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsj4z" podUID="90ab43dd-38e6-4264-a086-5cef08691214" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 20:10:24 crc kubenswrapper[4950]: I0318 20:10:24.723641 4950 patch_prober.go:28] interesting pod/downloads-7954f5f757-rf8k5 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Mar 18 20:10:24 crc kubenswrapper[4950]: I0318 20:10:24.723711 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-rf8k5" podUID="290d1b89-d5c9-4928-8973-6c33f476b091" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Mar 18 20:10:24 crc kubenswrapper[4950]: I0318 20:10:24.723716 4950 patch_prober.go:28] interesting pod/downloads-7954f5f757-rf8k5 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Mar 18 20:10:24 crc kubenswrapper[4950]: I0318 20:10:24.723766 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rf8k5" podUID="290d1b89-d5c9-4928-8973-6c33f476b091" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Mar 18 20:10:24 crc kubenswrapper[4950]: I0318 20:10:24.902838 4950 patch_prober.go:28] interesting pod/console-f9d7485db-46hjn container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Mar 18 20:10:24 crc kubenswrapper[4950]: I0318 20:10:24.902895 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-46hjn" podUID="1cec2292-f508-42b2-916f-9a5808045626" containerName="console" probeResult="failure" output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" Mar 18 20:10:25 crc kubenswrapper[4950]: I0318 20:10:25.682914 4950 patch_prober.go:28] interesting pod/router-default-5444994796-gsj4z container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 18 20:10:25 crc kubenswrapper[4950]: [-]has-synced failed: reason withheld Mar 18 20:10:25 crc kubenswrapper[4950]: [+]process-running ok Mar 18 20:10:25 crc kubenswrapper[4950]: healthz check failed Mar 18 20:10:25 crc kubenswrapper[4950]: I0318 20:10:25.682975 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsj4z" podUID="90ab43dd-38e6-4264-a086-5cef08691214" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 18 20:10:26 crc kubenswrapper[4950]: I0318 20:10:26.683665 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-gsj4z" Mar 18 20:10:26 crc kubenswrapper[4950]: I0318 20:10:26.685635 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-gsj4z" Mar 18 20:10:29 crc kubenswrapper[4950]: I0318 20:10:29.073532 4950 ???:1] "http: TLS handshake error from 192.168.126.11:59978: no serving certificate available for the kubelet" Mar 18 20:10:32 crc kubenswrapper[4950]: I0318 20:10:32.959067 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6cc4b56544-l9lgg"] Mar 18 20:10:32 crc kubenswrapper[4950]: I0318 20:10:32.960371 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" podUID="19f3cffc-3f5f-4166-b282-cc1f352fde48" containerName="controller-manager" containerID="cri-o://64cf7de40223a2111e01f2c5dd2405deb1c51e5d0e6b1e65dca597ff2c235e62" gracePeriod=30 Mar 18 20:10:32 crc kubenswrapper[4950]: I0318 20:10:32.980012 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q"] Mar 18 20:10:32 crc kubenswrapper[4950]: I0318 20:10:32.980281 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" podUID="b888d329-77fd-4fdd-bf49-276493b342b1" containerName="route-controller-manager" containerID="cri-o://b2ffdb35fd9f78eff9a5838150733a90ae51886225a026b7b2d46c0aa05e5dfe" gracePeriod=30 Mar 18 20:10:33 crc kubenswrapper[4950]: I0318 20:10:33.767663 4950 generic.go:334] "Generic (PLEG): container finished" podID="19f3cffc-3f5f-4166-b282-cc1f352fde48" containerID="64cf7de40223a2111e01f2c5dd2405deb1c51e5d0e6b1e65dca597ff2c235e62" exitCode=0 Mar 18 20:10:33 crc kubenswrapper[4950]: I0318 20:10:33.767737 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" event={"ID":"19f3cffc-3f5f-4166-b282-cc1f352fde48","Type":"ContainerDied","Data":"64cf7de40223a2111e01f2c5dd2405deb1c51e5d0e6b1e65dca597ff2c235e62"} Mar 18 20:10:33 crc kubenswrapper[4950]: I0318 20:10:33.836165 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:10:33 crc kubenswrapper[4950]: I0318 20:10:33.836268 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:10:34 crc kubenswrapper[4950]: I0318 20:10:34.730847 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-rf8k5" Mar 18 20:10:34 crc kubenswrapper[4950]: I0318 20:10:34.773804 4950 generic.go:334] "Generic (PLEG): container finished" podID="b888d329-77fd-4fdd-bf49-276493b342b1" containerID="b2ffdb35fd9f78eff9a5838150733a90ae51886225a026b7b2d46c0aa05e5dfe" exitCode=0 Mar 18 20:10:34 crc kubenswrapper[4950]: I0318 20:10:34.773847 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" event={"ID":"b888d329-77fd-4fdd-bf49-276493b342b1","Type":"ContainerDied","Data":"b2ffdb35fd9f78eff9a5838150733a90ae51886225a026b7b2d46c0aa05e5dfe"} Mar 18 20:10:34 crc kubenswrapper[4950]: I0318 20:10:34.906215 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:34 crc kubenswrapper[4950]: I0318 20:10:34.910450 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:10:36 crc kubenswrapper[4950]: I0318 20:10:36.647892 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:10:37 crc kubenswrapper[4950]: I0318 20:10:37.054625 4950 patch_prober.go:28] interesting pod/route-controller-manager-7cf6998584-w986q container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.53:8443/healthz\": dial tcp 10.217.0.53:8443: connect: connection refused" start-of-body= Mar 18 20:10:37 crc kubenswrapper[4950]: I0318 20:10:37.054684 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" podUID="b888d329-77fd-4fdd-bf49-276493b342b1" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.53:8443/healthz\": dial tcp 10.217.0.53:8443: connect: connection refused" Mar 18 20:10:39 crc kubenswrapper[4950]: I0318 20:10:39.282638 4950 patch_prober.go:28] interesting pod/controller-manager-6cc4b56544-l9lgg container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" start-of-body= Mar 18 20:10:39 crc kubenswrapper[4950]: I0318 20:10:39.282697 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" podUID="19f3cffc-3f5f-4166-b282-cc1f352fde48" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.349184 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.361313 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b888d329-77fd-4fdd-bf49-276493b342b1-config\") pod \"b888d329-77fd-4fdd-bf49-276493b342b1\" (UID: \"b888d329-77fd-4fdd-bf49-276493b342b1\") " Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.361377 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b888d329-77fd-4fdd-bf49-276493b342b1-client-ca\") pod \"b888d329-77fd-4fdd-bf49-276493b342b1\" (UID: \"b888d329-77fd-4fdd-bf49-276493b342b1\") " Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.361648 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b888d329-77fd-4fdd-bf49-276493b342b1-serving-cert\") pod \"b888d329-77fd-4fdd-bf49-276493b342b1\" (UID: \"b888d329-77fd-4fdd-bf49-276493b342b1\") " Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.361693 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szbc2\" (UniqueName: \"kubernetes.io/projected/b888d329-77fd-4fdd-bf49-276493b342b1-kube-api-access-szbc2\") pod \"b888d329-77fd-4fdd-bf49-276493b342b1\" (UID: \"b888d329-77fd-4fdd-bf49-276493b342b1\") " Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.362165 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b888d329-77fd-4fdd-bf49-276493b342b1-client-ca" (OuterVolumeSpecName: "client-ca") pod "b888d329-77fd-4fdd-bf49-276493b342b1" (UID: "b888d329-77fd-4fdd-bf49-276493b342b1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.362302 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b888d329-77fd-4fdd-bf49-276493b342b1-config" (OuterVolumeSpecName: "config") pod "b888d329-77fd-4fdd-bf49-276493b342b1" (UID: "b888d329-77fd-4fdd-bf49-276493b342b1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.375802 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b888d329-77fd-4fdd-bf49-276493b342b1-kube-api-access-szbc2" (OuterVolumeSpecName: "kube-api-access-szbc2") pod "b888d329-77fd-4fdd-bf49-276493b342b1" (UID: "b888d329-77fd-4fdd-bf49-276493b342b1"). InnerVolumeSpecName "kube-api-access-szbc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.381952 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq"] Mar 18 20:10:42 crc kubenswrapper[4950]: E0318 20:10:42.382267 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="164efb7c-0b83-42ed-8d30-acc8a1ff1794" containerName="collect-profiles" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.382341 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="164efb7c-0b83-42ed-8d30-acc8a1ff1794" containerName="collect-profiles" Mar 18 20:10:42 crc kubenswrapper[4950]: E0318 20:10:42.382685 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01bd704c-cabf-438e-8deb-1fb33a59e95b" containerName="pruner" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.382800 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="01bd704c-cabf-438e-8deb-1fb33a59e95b" containerName="pruner" Mar 18 20:10:42 crc kubenswrapper[4950]: E0318 20:10:42.382884 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b888d329-77fd-4fdd-bf49-276493b342b1" containerName="route-controller-manager" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.382955 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="b888d329-77fd-4fdd-bf49-276493b342b1" containerName="route-controller-manager" Mar 18 20:10:42 crc kubenswrapper[4950]: E0318 20:10:42.383018 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2a617fa-97f4-4c7c-bc20-6234cc29e79a" containerName="pruner" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.383078 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2a617fa-97f4-4c7c-bc20-6234cc29e79a" containerName="pruner" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.383281 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="b888d329-77fd-4fdd-bf49-276493b342b1" containerName="route-controller-manager" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.383377 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="01bd704c-cabf-438e-8deb-1fb33a59e95b" containerName="pruner" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.383521 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2a617fa-97f4-4c7c-bc20-6234cc29e79a" containerName="pruner" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.383613 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="164efb7c-0b83-42ed-8d30-acc8a1ff1794" containerName="collect-profiles" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.384150 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.382632 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b888d329-77fd-4fdd-bf49-276493b342b1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b888d329-77fd-4fdd-bf49-276493b342b1" (UID: "b888d329-77fd-4fdd-bf49-276493b342b1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.403714 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq"] Mar 18 20:10:42 crc kubenswrapper[4950]: E0318 20:10:42.428478 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: reading blob sha256:2086b7801d96d309e48e1c678789d95541de89bbae905e6f5a8de845927ca051: fetching blob: received unexpected HTTP status: 502 Bad Gateway" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 18 20:10:42 crc kubenswrapper[4950]: E0318 20:10:42.428677 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9s64w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-fk4b9_openshift-marketplace(f54c83ec-47ad-4ced-93c3-d321460f40f3): ErrImagePull: copying system image from manifest list: reading blob sha256:2086b7801d96d309e48e1c678789d95541de89bbae905e6f5a8de845927ca051: fetching blob: received unexpected HTTP status: 502 Bad Gateway" logger="UnhandledError" Mar 18 20:10:42 crc kubenswrapper[4950]: E0318 20:10:42.429873 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"copying system image from manifest list: reading blob sha256:2086b7801d96d309e48e1c678789d95541de89bbae905e6f5a8de845927ca051: fetching blob: received unexpected HTTP status: 502 Bad Gateway\"" pod="openshift-marketplace/certified-operators-fk4b9" podUID="f54c83ec-47ad-4ced-93c3-d321460f40f3" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.464352 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-serving-cert\") pod \"route-controller-manager-6bcd56cc6f-ddbbq\" (UID: \"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1\") " pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.464400 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k68qr\" (UniqueName: \"kubernetes.io/projected/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-kube-api-access-k68qr\") pod \"route-controller-manager-6bcd56cc6f-ddbbq\" (UID: \"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1\") " pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.464433 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-config\") pod \"route-controller-manager-6bcd56cc6f-ddbbq\" (UID: \"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1\") " pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.464478 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-client-ca\") pod \"route-controller-manager-6bcd56cc6f-ddbbq\" (UID: \"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1\") " pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.464520 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b888d329-77fd-4fdd-bf49-276493b342b1-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.464531 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szbc2\" (UniqueName: \"kubernetes.io/projected/b888d329-77fd-4fdd-bf49-276493b342b1-kube-api-access-szbc2\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.464539 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b888d329-77fd-4fdd-bf49-276493b342b1-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.464547 4950 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b888d329-77fd-4fdd-bf49-276493b342b1-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.566269 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-client-ca\") pod \"route-controller-manager-6bcd56cc6f-ddbbq\" (UID: \"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1\") " pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.566754 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-serving-cert\") pod \"route-controller-manager-6bcd56cc6f-ddbbq\" (UID: \"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1\") " pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.566787 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k68qr\" (UniqueName: \"kubernetes.io/projected/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-kube-api-access-k68qr\") pod \"route-controller-manager-6bcd56cc6f-ddbbq\" (UID: \"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1\") " pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.566890 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-config\") pod \"route-controller-manager-6bcd56cc6f-ddbbq\" (UID: \"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1\") " pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.568456 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-client-ca\") pod \"route-controller-manager-6bcd56cc6f-ddbbq\" (UID: \"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1\") " pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.569015 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-config\") pod \"route-controller-manager-6bcd56cc6f-ddbbq\" (UID: \"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1\") " pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.573752 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-serving-cert\") pod \"route-controller-manager-6bcd56cc6f-ddbbq\" (UID: \"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1\") " pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.587166 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k68qr\" (UniqueName: \"kubernetes.io/projected/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-kube-api-access-k68qr\") pod \"route-controller-manager-6bcd56cc6f-ddbbq\" (UID: \"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1\") " pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.709222 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.831447 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" event={"ID":"b888d329-77fd-4fdd-bf49-276493b342b1","Type":"ContainerDied","Data":"0103e51dbd663e60a384abd6aa6b262fa7fcda60042408d1ae928c4e65593d53"} Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.831528 4950 scope.go:117] "RemoveContainer" containerID="b2ffdb35fd9f78eff9a5838150733a90ae51886225a026b7b2d46c0aa05e5dfe" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.831475 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q" Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.880385 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q"] Mar 18 20:10:42 crc kubenswrapper[4950]: I0318 20:10:42.887164 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cf6998584-w986q"] Mar 18 20:10:43 crc kubenswrapper[4950]: I0318 20:10:43.488072 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b888d329-77fd-4fdd-bf49-276493b342b1" path="/var/lib/kubelet/pods/b888d329-77fd-4fdd-bf49-276493b342b1/volumes" Mar 18 20:10:43 crc kubenswrapper[4950]: E0318 20:10:43.602582 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-fk4b9" podUID="f54c83ec-47ad-4ced-93c3-d321460f40f3" Mar 18 20:10:43 crc kubenswrapper[4950]: E0318 20:10:43.634152 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-cli:latest" Mar 18 20:10:43 crc kubenswrapper[4950]: E0318 20:10:43.634322 4950 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 20:10:43 crc kubenswrapper[4950]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Mar 18 20:10:43 crc kubenswrapper[4950]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sjg8b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29564408-2658d_openshift-infra(97cb884c-5f55-4ca0-8639-193c2f1df139): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled Mar 18 20:10:43 crc kubenswrapper[4950]: > logger="UnhandledError" Mar 18 20:10:43 crc kubenswrapper[4950]: E0318 20:10:43.636072 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-infra/auto-csr-approver-29564408-2658d" podUID="97cb884c-5f55-4ca0-8639-193c2f1df139" Mar 18 20:10:43 crc kubenswrapper[4950]: E0318 20:10:43.837678 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29564408-2658d" podUID="97cb884c-5f55-4ca0-8639-193c2f1df139" Mar 18 20:10:43 crc kubenswrapper[4950]: E0318 20:10:43.967734 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-cli:latest" Mar 18 20:10:43 crc kubenswrapper[4950]: E0318 20:10:43.968315 4950 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 18 20:10:43 crc kubenswrapper[4950]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Mar 18 20:10:43 crc kubenswrapper[4950]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cjf7l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29564410-9hqrd_openshift-infra(627e6f3d-cde5-4849-bd02-6437622f4978): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled Mar 18 20:10:43 crc kubenswrapper[4950]: > logger="UnhandledError" Mar 18 20:10:43 crc kubenswrapper[4950]: E0318 20:10:43.969508 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-infra/auto-csr-approver-29564410-9hqrd" podUID="627e6f3d-cde5-4849-bd02-6437622f4978" Mar 18 20:10:43 crc kubenswrapper[4950]: I0318 20:10:43.979143 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.089450 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19f3cffc-3f5f-4166-b282-cc1f352fde48-serving-cert\") pod \"19f3cffc-3f5f-4166-b282-cc1f352fde48\" (UID: \"19f3cffc-3f5f-4166-b282-cc1f352fde48\") " Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.089518 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19f3cffc-3f5f-4166-b282-cc1f352fde48-config\") pod \"19f3cffc-3f5f-4166-b282-cc1f352fde48\" (UID: \"19f3cffc-3f5f-4166-b282-cc1f352fde48\") " Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.089576 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/19f3cffc-3f5f-4166-b282-cc1f352fde48-proxy-ca-bundles\") pod \"19f3cffc-3f5f-4166-b282-cc1f352fde48\" (UID: \"19f3cffc-3f5f-4166-b282-cc1f352fde48\") " Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.089631 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/19f3cffc-3f5f-4166-b282-cc1f352fde48-client-ca\") pod \"19f3cffc-3f5f-4166-b282-cc1f352fde48\" (UID: \"19f3cffc-3f5f-4166-b282-cc1f352fde48\") " Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.089690 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4n2wn\" (UniqueName: \"kubernetes.io/projected/19f3cffc-3f5f-4166-b282-cc1f352fde48-kube-api-access-4n2wn\") pod \"19f3cffc-3f5f-4166-b282-cc1f352fde48\" (UID: \"19f3cffc-3f5f-4166-b282-cc1f352fde48\") " Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.090891 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19f3cffc-3f5f-4166-b282-cc1f352fde48-config" (OuterVolumeSpecName: "config") pod "19f3cffc-3f5f-4166-b282-cc1f352fde48" (UID: "19f3cffc-3f5f-4166-b282-cc1f352fde48"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.091476 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19f3cffc-3f5f-4166-b282-cc1f352fde48-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "19f3cffc-3f5f-4166-b282-cc1f352fde48" (UID: "19f3cffc-3f5f-4166-b282-cc1f352fde48"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.091870 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19f3cffc-3f5f-4166-b282-cc1f352fde48-client-ca" (OuterVolumeSpecName: "client-ca") pod "19f3cffc-3f5f-4166-b282-cc1f352fde48" (UID: "19f3cffc-3f5f-4166-b282-cc1f352fde48"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.096820 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19f3cffc-3f5f-4166-b282-cc1f352fde48-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "19f3cffc-3f5f-4166-b282-cc1f352fde48" (UID: "19f3cffc-3f5f-4166-b282-cc1f352fde48"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.098338 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19f3cffc-3f5f-4166-b282-cc1f352fde48-kube-api-access-4n2wn" (OuterVolumeSpecName: "kube-api-access-4n2wn") pod "19f3cffc-3f5f-4166-b282-cc1f352fde48" (UID: "19f3cffc-3f5f-4166-b282-cc1f352fde48"). InnerVolumeSpecName "kube-api-access-4n2wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.191297 4950 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/19f3cffc-3f5f-4166-b282-cc1f352fde48-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.191338 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4n2wn\" (UniqueName: \"kubernetes.io/projected/19f3cffc-3f5f-4166-b282-cc1f352fde48-kube-api-access-4n2wn\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.191350 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19f3cffc-3f5f-4166-b282-cc1f352fde48-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.191359 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19f3cffc-3f5f-4166-b282-cc1f352fde48-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.191368 4950 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/19f3cffc-3f5f-4166-b282-cc1f352fde48-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.655996 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7"] Mar 18 20:10:44 crc kubenswrapper[4950]: E0318 20:10:44.656655 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19f3cffc-3f5f-4166-b282-cc1f352fde48" containerName="controller-manager" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.656672 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="19f3cffc-3f5f-4166-b282-cc1f352fde48" containerName="controller-manager" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.656815 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="19f3cffc-3f5f-4166-b282-cc1f352fde48" containerName="controller-manager" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.657315 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.667592 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7"] Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.698075 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/331ad71b-521e-4b17-995b-0711b41d3b0e-config\") pod \"controller-manager-6c5b68fd4f-9qpg7\" (UID: \"331ad71b-521e-4b17-995b-0711b41d3b0e\") " pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.698190 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/331ad71b-521e-4b17-995b-0711b41d3b0e-serving-cert\") pod \"controller-manager-6c5b68fd4f-9qpg7\" (UID: \"331ad71b-521e-4b17-995b-0711b41d3b0e\") " pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.698218 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl6kg\" (UniqueName: \"kubernetes.io/projected/331ad71b-521e-4b17-995b-0711b41d3b0e-kube-api-access-xl6kg\") pod \"controller-manager-6c5b68fd4f-9qpg7\" (UID: \"331ad71b-521e-4b17-995b-0711b41d3b0e\") " pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.698263 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/331ad71b-521e-4b17-995b-0711b41d3b0e-proxy-ca-bundles\") pod \"controller-manager-6c5b68fd4f-9qpg7\" (UID: \"331ad71b-521e-4b17-995b-0711b41d3b0e\") " pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.698293 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/331ad71b-521e-4b17-995b-0711b41d3b0e-client-ca\") pod \"controller-manager-6c5b68fd4f-9qpg7\" (UID: \"331ad71b-521e-4b17-995b-0711b41d3b0e\") " pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.799163 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/331ad71b-521e-4b17-995b-0711b41d3b0e-config\") pod \"controller-manager-6c5b68fd4f-9qpg7\" (UID: \"331ad71b-521e-4b17-995b-0711b41d3b0e\") " pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.799227 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/331ad71b-521e-4b17-995b-0711b41d3b0e-serving-cert\") pod \"controller-manager-6c5b68fd4f-9qpg7\" (UID: \"331ad71b-521e-4b17-995b-0711b41d3b0e\") " pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.799251 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl6kg\" (UniqueName: \"kubernetes.io/projected/331ad71b-521e-4b17-995b-0711b41d3b0e-kube-api-access-xl6kg\") pod \"controller-manager-6c5b68fd4f-9qpg7\" (UID: \"331ad71b-521e-4b17-995b-0711b41d3b0e\") " pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.799276 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/331ad71b-521e-4b17-995b-0711b41d3b0e-proxy-ca-bundles\") pod \"controller-manager-6c5b68fd4f-9qpg7\" (UID: \"331ad71b-521e-4b17-995b-0711b41d3b0e\") " pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.799305 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/331ad71b-521e-4b17-995b-0711b41d3b0e-client-ca\") pod \"controller-manager-6c5b68fd4f-9qpg7\" (UID: \"331ad71b-521e-4b17-995b-0711b41d3b0e\") " pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.800859 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/331ad71b-521e-4b17-995b-0711b41d3b0e-proxy-ca-bundles\") pod \"controller-manager-6c5b68fd4f-9qpg7\" (UID: \"331ad71b-521e-4b17-995b-0711b41d3b0e\") " pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.801252 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/331ad71b-521e-4b17-995b-0711b41d3b0e-client-ca\") pod \"controller-manager-6c5b68fd4f-9qpg7\" (UID: \"331ad71b-521e-4b17-995b-0711b41d3b0e\") " pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.802059 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/331ad71b-521e-4b17-995b-0711b41d3b0e-config\") pod \"controller-manager-6c5b68fd4f-9qpg7\" (UID: \"331ad71b-521e-4b17-995b-0711b41d3b0e\") " pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.819845 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/331ad71b-521e-4b17-995b-0711b41d3b0e-serving-cert\") pod \"controller-manager-6c5b68fd4f-9qpg7\" (UID: \"331ad71b-521e-4b17-995b-0711b41d3b0e\") " pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.823703 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl6kg\" (UniqueName: \"kubernetes.io/projected/331ad71b-521e-4b17-995b-0711b41d3b0e-kube-api-access-xl6kg\") pod \"controller-manager-6c5b68fd4f-9qpg7\" (UID: \"331ad71b-521e-4b17-995b-0711b41d3b0e\") " pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.844399 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.844799 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6cc4b56544-l9lgg" event={"ID":"19f3cffc-3f5f-4166-b282-cc1f352fde48","Type":"ContainerDied","Data":"f0ac3d91d6d9cae262049da4cb96530c59baec099614b9f14af0935d87340c49"} Mar 18 20:10:44 crc kubenswrapper[4950]: E0318 20:10:44.848694 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29564410-9hqrd" podUID="627e6f3d-cde5-4849-bd02-6437622f4978" Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.873374 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6cc4b56544-l9lgg"] Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.881515 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6cc4b56544-l9lgg"] Mar 18 20:10:44 crc kubenswrapper[4950]: I0318 20:10:44.977019 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" Mar 18 20:10:45 crc kubenswrapper[4950]: I0318 20:10:45.425874 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qqfjh" Mar 18 20:10:45 crc kubenswrapper[4950]: I0318 20:10:45.495041 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19f3cffc-3f5f-4166-b282-cc1f352fde48" path="/var/lib/kubelet/pods/19f3cffc-3f5f-4166-b282-cc1f352fde48/volumes" Mar 18 20:10:47 crc kubenswrapper[4950]: I0318 20:10:47.566552 4950 scope.go:117] "RemoveContainer" containerID="64cf7de40223a2111e01f2c5dd2405deb1c51e5d0e6b1e65dca597ff2c235e62" Mar 18 20:10:49 crc kubenswrapper[4950]: I0318 20:10:49.488696 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 18 20:10:49 crc kubenswrapper[4950]: I0318 20:10:49.489548 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 20:10:49 crc kubenswrapper[4950]: I0318 20:10:49.491803 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 18 20:10:49 crc kubenswrapper[4950]: I0318 20:10:49.494150 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 18 20:10:49 crc kubenswrapper[4950]: I0318 20:10:49.497030 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 18 20:10:49 crc kubenswrapper[4950]: I0318 20:10:49.560442 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0509451c-e8c8-4122-bf76-e51f16acbcc5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0509451c-e8c8-4122-bf76-e51f16acbcc5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 20:10:49 crc kubenswrapper[4950]: I0318 20:10:49.560523 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0509451c-e8c8-4122-bf76-e51f16acbcc5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0509451c-e8c8-4122-bf76-e51f16acbcc5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 20:10:49 crc kubenswrapper[4950]: I0318 20:10:49.662619 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0509451c-e8c8-4122-bf76-e51f16acbcc5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0509451c-e8c8-4122-bf76-e51f16acbcc5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 20:10:49 crc kubenswrapper[4950]: I0318 20:10:49.663062 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0509451c-e8c8-4122-bf76-e51f16acbcc5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0509451c-e8c8-4122-bf76-e51f16acbcc5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 20:10:49 crc kubenswrapper[4950]: I0318 20:10:49.663260 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0509451c-e8c8-4122-bf76-e51f16acbcc5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0509451c-e8c8-4122-bf76-e51f16acbcc5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 20:10:49 crc kubenswrapper[4950]: I0318 20:10:49.681272 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0509451c-e8c8-4122-bf76-e51f16acbcc5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0509451c-e8c8-4122-bf76-e51f16acbcc5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 20:10:49 crc kubenswrapper[4950]: I0318 20:10:49.825895 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 20:10:51 crc kubenswrapper[4950]: I0318 20:10:51.515750 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 18 20:10:52 crc kubenswrapper[4950]: E0318 20:10:52.070294 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 18 20:10:52 crc kubenswrapper[4950]: E0318 20:10:52.070502 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k7zmg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-nb7fx_openshift-marketplace(33a99c2b-09d6-4ee7-b875-509ef5e144eb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 18 20:10:52 crc kubenswrapper[4950]: E0318 20:10:52.071678 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-nb7fx" podUID="33a99c2b-09d6-4ee7-b875-509ef5e144eb" Mar 18 20:10:52 crc kubenswrapper[4950]: I0318 20:10:52.964976 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7"] Mar 18 20:10:53 crc kubenswrapper[4950]: I0318 20:10:53.068743 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq"] Mar 18 20:10:53 crc kubenswrapper[4950]: I0318 20:10:53.471955 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 18 20:10:53 crc kubenswrapper[4950]: I0318 20:10:53.481698 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 18 20:10:53 crc kubenswrapper[4950]: I0318 20:10:53.497205 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 18 20:10:53 crc kubenswrapper[4950]: I0318 20:10:53.509500 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e7e1cabf-6954-4269-a638-16fb70a90850-var-lock\") pod \"installer-9-crc\" (UID: \"e7e1cabf-6954-4269-a638-16fb70a90850\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 20:10:53 crc kubenswrapper[4950]: I0318 20:10:53.509569 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e1cabf-6954-4269-a638-16fb70a90850-kube-api-access\") pod \"installer-9-crc\" (UID: \"e7e1cabf-6954-4269-a638-16fb70a90850\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 20:10:53 crc kubenswrapper[4950]: I0318 20:10:53.509608 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e7e1cabf-6954-4269-a638-16fb70a90850-kubelet-dir\") pod \"installer-9-crc\" (UID: \"e7e1cabf-6954-4269-a638-16fb70a90850\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 20:10:53 crc kubenswrapper[4950]: I0318 20:10:53.610797 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e1cabf-6954-4269-a638-16fb70a90850-kube-api-access\") pod \"installer-9-crc\" (UID: \"e7e1cabf-6954-4269-a638-16fb70a90850\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 20:10:53 crc kubenswrapper[4950]: I0318 20:10:53.610847 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e7e1cabf-6954-4269-a638-16fb70a90850-kubelet-dir\") pod \"installer-9-crc\" (UID: \"e7e1cabf-6954-4269-a638-16fb70a90850\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 20:10:53 crc kubenswrapper[4950]: I0318 20:10:53.610903 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e7e1cabf-6954-4269-a638-16fb70a90850-var-lock\") pod \"installer-9-crc\" (UID: \"e7e1cabf-6954-4269-a638-16fb70a90850\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 20:10:53 crc kubenswrapper[4950]: I0318 20:10:53.610976 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e7e1cabf-6954-4269-a638-16fb70a90850-var-lock\") pod \"installer-9-crc\" (UID: \"e7e1cabf-6954-4269-a638-16fb70a90850\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 20:10:53 crc kubenswrapper[4950]: I0318 20:10:53.611302 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e7e1cabf-6954-4269-a638-16fb70a90850-kubelet-dir\") pod \"installer-9-crc\" (UID: \"e7e1cabf-6954-4269-a638-16fb70a90850\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 20:10:53 crc kubenswrapper[4950]: I0318 20:10:53.632169 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e1cabf-6954-4269-a638-16fb70a90850-kube-api-access\") pod \"installer-9-crc\" (UID: \"e7e1cabf-6954-4269-a638-16fb70a90850\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 18 20:10:53 crc kubenswrapper[4950]: I0318 20:10:53.813346 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 18 20:10:53 crc kubenswrapper[4950]: E0318 20:10:53.914021 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-nb7fx" podUID="33a99c2b-09d6-4ee7-b875-509ef5e144eb" Mar 18 20:10:53 crc kubenswrapper[4950]: E0318 20:10:53.992229 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 18 20:10:53 crc kubenswrapper[4950]: E0318 20:10:53.992459 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x9n22,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-7dzz6_openshift-marketplace(bbeed1b2-6abb-49df-94d9-7d9fe6e5a563): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 18 20:10:53 crc kubenswrapper[4950]: E0318 20:10:53.997799 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-7dzz6" podUID="bbeed1b2-6abb-49df-94d9-7d9fe6e5a563" Mar 18 20:10:54 crc kubenswrapper[4950]: E0318 20:10:54.034361 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 18 20:10:54 crc kubenswrapper[4950]: E0318 20:10:54.034622 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mvtkc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-lhhrh_openshift-marketplace(8fe40723-62a6-4fe7-989c-b6f968467b4c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 18 20:10:54 crc kubenswrapper[4950]: E0318 20:10:54.036346 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-lhhrh" podUID="8fe40723-62a6-4fe7-989c-b6f968467b4c" Mar 18 20:10:56 crc kubenswrapper[4950]: E0318 20:10:56.163270 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7dzz6" podUID="bbeed1b2-6abb-49df-94d9-7d9fe6e5a563" Mar 18 20:10:56 crc kubenswrapper[4950]: E0318 20:10:56.163965 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-lhhrh" podUID="8fe40723-62a6-4fe7-989c-b6f968467b4c" Mar 18 20:10:56 crc kubenswrapper[4950]: E0318 20:10:56.259829 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 18 20:10:56 crc kubenswrapper[4950]: E0318 20:10:56.260024 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-swzmx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-rwzv9_openshift-marketplace(82c68303-8af4-41a0-af35-312cc8b0bd7e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 18 20:10:56 crc kubenswrapper[4950]: E0318 20:10:56.261539 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-rwzv9" podUID="82c68303-8af4-41a0-af35-312cc8b0bd7e" Mar 18 20:10:56 crc kubenswrapper[4950]: E0318 20:10:56.283946 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 18 20:10:56 crc kubenswrapper[4950]: E0318 20:10:56.284184 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-slf9m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-gzj5q_openshift-marketplace(044d9d2a-6b99-49ee-b00d-728c7832815c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 18 20:10:56 crc kubenswrapper[4950]: E0318 20:10:56.285534 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-gzj5q" podUID="044d9d2a-6b99-49ee-b00d-728c7832815c" Mar 18 20:10:56 crc kubenswrapper[4950]: E0318 20:10:56.288229 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 18 20:10:56 crc kubenswrapper[4950]: E0318 20:10:56.288331 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lbj6d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-5fnpx_openshift-marketplace(d3ae2e86-be5d-4546-97a0-88301381868b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 18 20:10:56 crc kubenswrapper[4950]: E0318 20:10:56.289599 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-5fnpx" podUID="d3ae2e86-be5d-4546-97a0-88301381868b" Mar 18 20:10:59 crc kubenswrapper[4950]: E0318 20:10:59.797856 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-gzj5q" podUID="044d9d2a-6b99-49ee-b00d-728c7832815c" Mar 18 20:10:59 crc kubenswrapper[4950]: E0318 20:10:59.797856 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-rwzv9" podUID="82c68303-8af4-41a0-af35-312cc8b0bd7e" Mar 18 20:10:59 crc kubenswrapper[4950]: E0318 20:10:59.798050 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-5fnpx" podUID="d3ae2e86-be5d-4546-97a0-88301381868b" Mar 18 20:11:00 crc kubenswrapper[4950]: I0318 20:11:00.056698 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq"] Mar 18 20:11:00 crc kubenswrapper[4950]: W0318 20:11:00.061938 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4325fb8_e169_4c5a_9b31_fa46cb8cb8a1.slice/crio-0aa2f1245c9cb0702c3e2868ca9f5435b72c9527010dc1cd396147011a1aabc4 WatchSource:0}: Error finding container 0aa2f1245c9cb0702c3e2868ca9f5435b72c9527010dc1cd396147011a1aabc4: Status 404 returned error can't find the container with id 0aa2f1245c9cb0702c3e2868ca9f5435b72c9527010dc1cd396147011a1aabc4 Mar 18 20:11:00 crc kubenswrapper[4950]: I0318 20:11:00.365093 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 18 20:11:00 crc kubenswrapper[4950]: I0318 20:11:00.368571 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7"] Mar 18 20:11:00 crc kubenswrapper[4950]: I0318 20:11:00.381567 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 18 20:11:00 crc kubenswrapper[4950]: W0318 20:11:00.386726 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod331ad71b_521e_4b17_995b_0711b41d3b0e.slice/crio-b1b0b87723316764de411e8c98a78dfd1d24fb9d48d89b962b7f19d02669ff99 WatchSource:0}: Error finding container b1b0b87723316764de411e8c98a78dfd1d24fb9d48d89b962b7f19d02669ff99: Status 404 returned error can't find the container with id b1b0b87723316764de411e8c98a78dfd1d24fb9d48d89b962b7f19d02669ff99 Mar 18 20:11:00 crc kubenswrapper[4950]: W0318 20:11:00.395801 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod0509451c_e8c8_4122_bf76_e51f16acbcc5.slice/crio-a9dadff189717861421e11c8df7d4dc7caac40f42a20c40824135999e77e2015 WatchSource:0}: Error finding container a9dadff189717861421e11c8df7d4dc7caac40f42a20c40824135999e77e2015: Status 404 returned error can't find the container with id a9dadff189717861421e11c8df7d4dc7caac40f42a20c40824135999e77e2015 Mar 18 20:11:00 crc kubenswrapper[4950]: E0318 20:11:00.466910 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 18 20:11:00 crc kubenswrapper[4950]: E0318 20:11:00.467074 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bgcwj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-8kwk7_openshift-marketplace(4a386d81-3e2c-4df7-988d-5fb1758be616): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 18 20:11:00 crc kubenswrapper[4950]: E0318 20:11:00.468263 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-8kwk7" podUID="4a386d81-3e2c-4df7-988d-5fb1758be616" Mar 18 20:11:00 crc kubenswrapper[4950]: I0318 20:11:00.957933 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" event={"ID":"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1","Type":"ContainerStarted","Data":"867bc39c90e6be3e7d28a7c43b2ac342c37101586533e172e7281a164fcbd070"} Mar 18 20:11:00 crc kubenswrapper[4950]: I0318 20:11:00.958296 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" Mar 18 20:11:00 crc kubenswrapper[4950]: I0318 20:11:00.958329 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" event={"ID":"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1","Type":"ContainerStarted","Data":"0aa2f1245c9cb0702c3e2868ca9f5435b72c9527010dc1cd396147011a1aabc4"} Mar 18 20:11:00 crc kubenswrapper[4950]: I0318 20:11:00.958053 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" podUID="c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1" containerName="route-controller-manager" containerID="cri-o://867bc39c90e6be3e7d28a7c43b2ac342c37101586533e172e7281a164fcbd070" gracePeriod=30 Mar 18 20:11:00 crc kubenswrapper[4950]: I0318 20:11:00.962683 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"0509451c-e8c8-4122-bf76-e51f16acbcc5","Type":"ContainerStarted","Data":"d232e03212007cd9aa092924970118e81dcb88ded57dbdf034acbbc5627929ac"} Mar 18 20:11:00 crc kubenswrapper[4950]: I0318 20:11:00.962734 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"0509451c-e8c8-4122-bf76-e51f16acbcc5","Type":"ContainerStarted","Data":"a9dadff189717861421e11c8df7d4dc7caac40f42a20c40824135999e77e2015"} Mar 18 20:11:00 crc kubenswrapper[4950]: I0318 20:11:00.967621 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" event={"ID":"331ad71b-521e-4b17-995b-0711b41d3b0e","Type":"ContainerStarted","Data":"e005608daa13eb6b654aa3e3c023165e2978653d6cfb8251c9a417e6a95ba96a"} Mar 18 20:11:00 crc kubenswrapper[4950]: I0318 20:11:00.967693 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" event={"ID":"331ad71b-521e-4b17-995b-0711b41d3b0e","Type":"ContainerStarted","Data":"b1b0b87723316764de411e8c98a78dfd1d24fb9d48d89b962b7f19d02669ff99"} Mar 18 20:11:00 crc kubenswrapper[4950]: I0318 20:11:00.969056 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e7e1cabf-6954-4269-a638-16fb70a90850","Type":"ContainerStarted","Data":"1b098843c103958ad8366dd7296e3874a197d70b5b5e7762b091b24705d5ca80"} Mar 18 20:11:00 crc kubenswrapper[4950]: I0318 20:11:00.969089 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e7e1cabf-6954-4269-a638-16fb70a90850","Type":"ContainerStarted","Data":"2cf6261e65c92df022363e1bc50e9bf0c8424f4c5df276f7ca9365c4b90d2f79"} Mar 18 20:11:00 crc kubenswrapper[4950]: E0318 20:11:00.970579 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-8kwk7" podUID="4a386d81-3e2c-4df7-988d-5fb1758be616" Mar 18 20:11:00 crc kubenswrapper[4950]: I0318 20:11:00.986888 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" podStartSLOduration=27.986866546 podStartE2EDuration="27.986866546s" podCreationTimestamp="2026-03-18 20:10:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:11:00.983999478 +0000 UTC m=+274.224841336" watchObservedRunningTime="2026-03-18 20:11:00.986866546 +0000 UTC m=+274.227708424" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.078202 4950 patch_prober.go:28] interesting pod/route-controller-manager-6bcd56cc6f-ddbbq container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.58:8443/healthz\": read tcp 10.217.0.2:56284->10.217.0.58:8443: read: connection reset by peer" start-of-body= Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.078292 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" podUID="c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.58:8443/healthz\": read tcp 10.217.0.2:56284->10.217.0.58:8443: read: connection reset by peer" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.635596 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-6bcd56cc6f-ddbbq_c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1/route-controller-manager/0.log" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.635860 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.663854 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd"] Mar 18 20:11:01 crc kubenswrapper[4950]: E0318 20:11:01.664087 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1" containerName="route-controller-manager" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.664100 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1" containerName="route-controller-manager" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.664397 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1" containerName="route-controller-manager" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.666537 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.673030 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd"] Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.753388 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-config\") pod \"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1\" (UID: \"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1\") " Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.753431 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-client-ca\") pod \"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1\" (UID: \"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1\") " Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.753462 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-serving-cert\") pod \"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1\" (UID: \"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1\") " Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.753508 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k68qr\" (UniqueName: \"kubernetes.io/projected/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-kube-api-access-k68qr\") pod \"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1\" (UID: \"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1\") " Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.753672 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxzkk\" (UniqueName: \"kubernetes.io/projected/21200e40-6c66-43d0-b8a2-1206361409ed-kube-api-access-xxzkk\") pod \"route-controller-manager-cc5465bc6-b9kcd\" (UID: \"21200e40-6c66-43d0-b8a2-1206361409ed\") " pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.753713 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21200e40-6c66-43d0-b8a2-1206361409ed-client-ca\") pod \"route-controller-manager-cc5465bc6-b9kcd\" (UID: \"21200e40-6c66-43d0-b8a2-1206361409ed\") " pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.753738 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21200e40-6c66-43d0-b8a2-1206361409ed-serving-cert\") pod \"route-controller-manager-cc5465bc6-b9kcd\" (UID: \"21200e40-6c66-43d0-b8a2-1206361409ed\") " pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.753754 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21200e40-6c66-43d0-b8a2-1206361409ed-config\") pod \"route-controller-manager-cc5465bc6-b9kcd\" (UID: \"21200e40-6c66-43d0-b8a2-1206361409ed\") " pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.754232 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-config" (OuterVolumeSpecName: "config") pod "c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1" (UID: "c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.754250 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-client-ca" (OuterVolumeSpecName: "client-ca") pod "c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1" (UID: "c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.761045 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1" (UID: "c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.761797 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-kube-api-access-k68qr" (OuterVolumeSpecName: "kube-api-access-k68qr") pod "c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1" (UID: "c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1"). InnerVolumeSpecName "kube-api-access-k68qr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.855277 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxzkk\" (UniqueName: \"kubernetes.io/projected/21200e40-6c66-43d0-b8a2-1206361409ed-kube-api-access-xxzkk\") pod \"route-controller-manager-cc5465bc6-b9kcd\" (UID: \"21200e40-6c66-43d0-b8a2-1206361409ed\") " pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.855337 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21200e40-6c66-43d0-b8a2-1206361409ed-client-ca\") pod \"route-controller-manager-cc5465bc6-b9kcd\" (UID: \"21200e40-6c66-43d0-b8a2-1206361409ed\") " pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.855366 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21200e40-6c66-43d0-b8a2-1206361409ed-serving-cert\") pod \"route-controller-manager-cc5465bc6-b9kcd\" (UID: \"21200e40-6c66-43d0-b8a2-1206361409ed\") " pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.855383 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21200e40-6c66-43d0-b8a2-1206361409ed-config\") pod \"route-controller-manager-cc5465bc6-b9kcd\" (UID: \"21200e40-6c66-43d0-b8a2-1206361409ed\") " pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.855435 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.855445 4950 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.855453 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.855462 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k68qr\" (UniqueName: \"kubernetes.io/projected/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1-kube-api-access-k68qr\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.856545 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21200e40-6c66-43d0-b8a2-1206361409ed-client-ca\") pod \"route-controller-manager-cc5465bc6-b9kcd\" (UID: \"21200e40-6c66-43d0-b8a2-1206361409ed\") " pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.856649 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21200e40-6c66-43d0-b8a2-1206361409ed-config\") pod \"route-controller-manager-cc5465bc6-b9kcd\" (UID: \"21200e40-6c66-43d0-b8a2-1206361409ed\") " pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.862166 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21200e40-6c66-43d0-b8a2-1206361409ed-serving-cert\") pod \"route-controller-manager-cc5465bc6-b9kcd\" (UID: \"21200e40-6c66-43d0-b8a2-1206361409ed\") " pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.886164 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxzkk\" (UniqueName: \"kubernetes.io/projected/21200e40-6c66-43d0-b8a2-1206361409ed-kube-api-access-xxzkk\") pod \"route-controller-manager-cc5465bc6-b9kcd\" (UID: \"21200e40-6c66-43d0-b8a2-1206361409ed\") " pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.977698 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564408-2658d" event={"ID":"97cb884c-5f55-4ca0-8639-193c2f1df139","Type":"ContainerStarted","Data":"d790f769ee6aced97a1576b8d64131e51abc022caf2a585431da9cc291585aeb"} Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.982028 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-6bcd56cc6f-ddbbq_c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1/route-controller-manager/0.log" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.982068 4950 generic.go:334] "Generic (PLEG): container finished" podID="c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1" containerID="867bc39c90e6be3e7d28a7c43b2ac342c37101586533e172e7281a164fcbd070" exitCode=255 Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.982157 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.982803 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" event={"ID":"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1","Type":"ContainerDied","Data":"867bc39c90e6be3e7d28a7c43b2ac342c37101586533e172e7281a164fcbd070"} Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.982869 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq" event={"ID":"c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1","Type":"ContainerDied","Data":"0aa2f1245c9cb0702c3e2868ca9f5435b72c9527010dc1cd396147011a1aabc4"} Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.982886 4950 scope.go:117] "RemoveContainer" containerID="867bc39c90e6be3e7d28a7c43b2ac342c37101586533e172e7281a164fcbd070" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.987527 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.988918 4950 generic.go:334] "Generic (PLEG): container finished" podID="0509451c-e8c8-4122-bf76-e51f16acbcc5" containerID="d232e03212007cd9aa092924970118e81dcb88ded57dbdf034acbbc5627929ac" exitCode=0 Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.988978 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"0509451c-e8c8-4122-bf76-e51f16acbcc5","Type":"ContainerDied","Data":"d232e03212007cd9aa092924970118e81dcb88ded57dbdf034acbbc5627929ac"} Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.991756 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29564408-2658d" podStartSLOduration=128.366888857 podStartE2EDuration="3m1.99174433s" podCreationTimestamp="2026-03-18 20:08:00 +0000 UTC" firstStartedPulling="2026-03-18 20:10:07.835180095 +0000 UTC m=+221.076021963" lastFinishedPulling="2026-03-18 20:11:01.460035568 +0000 UTC m=+274.700877436" observedRunningTime="2026-03-18 20:11:01.990418564 +0000 UTC m=+275.231260432" watchObservedRunningTime="2026-03-18 20:11:01.99174433 +0000 UTC m=+275.232586198" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.994033 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" podUID="331ad71b-521e-4b17-995b-0711b41d3b0e" containerName="controller-manager" containerID="cri-o://e005608daa13eb6b654aa3e3c023165e2978653d6cfb8251c9a417e6a95ba96a" gracePeriod=30 Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.994466 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564410-9hqrd" event={"ID":"627e6f3d-cde5-4849-bd02-6437622f4978","Type":"ContainerStarted","Data":"00dffb6edda0aae584ee703172b0a41786cf67c059f901a0afd2d75aeedf585d"} Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.994522 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" Mar 18 20:11:01 crc kubenswrapper[4950]: I0318 20:11:01.998387 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.020872 4950 scope.go:117] "RemoveContainer" containerID="867bc39c90e6be3e7d28a7c43b2ac342c37101586533e172e7281a164fcbd070" Mar 18 20:11:02 crc kubenswrapper[4950]: E0318 20:11:02.028838 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"867bc39c90e6be3e7d28a7c43b2ac342c37101586533e172e7281a164fcbd070\": container with ID starting with 867bc39c90e6be3e7d28a7c43b2ac342c37101586533e172e7281a164fcbd070 not found: ID does not exist" containerID="867bc39c90e6be3e7d28a7c43b2ac342c37101586533e172e7281a164fcbd070" Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.028888 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"867bc39c90e6be3e7d28a7c43b2ac342c37101586533e172e7281a164fcbd070"} err="failed to get container status \"867bc39c90e6be3e7d28a7c43b2ac342c37101586533e172e7281a164fcbd070\": rpc error: code = NotFound desc = could not find container \"867bc39c90e6be3e7d28a7c43b2ac342c37101586533e172e7281a164fcbd070\": container with ID starting with 867bc39c90e6be3e7d28a7c43b2ac342c37101586533e172e7281a164fcbd070 not found: ID does not exist" Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.051781 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" podStartSLOduration=30.05176151 podStartE2EDuration="30.05176151s" podCreationTimestamp="2026-03-18 20:10:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:11:02.049032806 +0000 UTC m=+275.289874674" watchObservedRunningTime="2026-03-18 20:11:02.05176151 +0000 UTC m=+275.292603378" Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.070886 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=9.070865142 podStartE2EDuration="9.070865142s" podCreationTimestamp="2026-03-18 20:10:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:11:02.063618424 +0000 UTC m=+275.304460292" watchObservedRunningTime="2026-03-18 20:11:02.070865142 +0000 UTC m=+275.311707010" Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.131167 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29564410-9hqrd" podStartSLOduration=7.561369004 podStartE2EDuration="1m2.13114428s" podCreationTimestamp="2026-03-18 20:10:00 +0000 UTC" firstStartedPulling="2026-03-18 20:10:06.892346119 +0000 UTC m=+220.133187987" lastFinishedPulling="2026-03-18 20:11:01.462121395 +0000 UTC m=+274.702963263" observedRunningTime="2026-03-18 20:11:02.116941972 +0000 UTC m=+275.357783840" watchObservedRunningTime="2026-03-18 20:11:02.13114428 +0000 UTC m=+275.371986158" Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.144339 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq"] Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.150399 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6bcd56cc6f-ddbbq"] Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.296020 4950 csr.go:261] certificate signing request csr-ppmnj is approved, waiting to be issued Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.301664 4950 csr.go:257] certificate signing request csr-ppmnj is issued Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.475816 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd"] Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.482219 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" Mar 18 20:11:02 crc kubenswrapper[4950]: W0318 20:11:02.484472 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21200e40_6c66_43d0_b8a2_1206361409ed.slice/crio-d87b134b60500cb2c32ee2b83ff97162fdebaec4c7707cb7c249673bb9867685 WatchSource:0}: Error finding container d87b134b60500cb2c32ee2b83ff97162fdebaec4c7707cb7c249673bb9867685: Status 404 returned error can't find the container with id d87b134b60500cb2c32ee2b83ff97162fdebaec4c7707cb7c249673bb9867685 Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.612749 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/331ad71b-521e-4b17-995b-0711b41d3b0e-serving-cert\") pod \"331ad71b-521e-4b17-995b-0711b41d3b0e\" (UID: \"331ad71b-521e-4b17-995b-0711b41d3b0e\") " Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.612808 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/331ad71b-521e-4b17-995b-0711b41d3b0e-config\") pod \"331ad71b-521e-4b17-995b-0711b41d3b0e\" (UID: \"331ad71b-521e-4b17-995b-0711b41d3b0e\") " Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.612837 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xl6kg\" (UniqueName: \"kubernetes.io/projected/331ad71b-521e-4b17-995b-0711b41d3b0e-kube-api-access-xl6kg\") pod \"331ad71b-521e-4b17-995b-0711b41d3b0e\" (UID: \"331ad71b-521e-4b17-995b-0711b41d3b0e\") " Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.612901 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/331ad71b-521e-4b17-995b-0711b41d3b0e-client-ca\") pod \"331ad71b-521e-4b17-995b-0711b41d3b0e\" (UID: \"331ad71b-521e-4b17-995b-0711b41d3b0e\") " Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.612916 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/331ad71b-521e-4b17-995b-0711b41d3b0e-proxy-ca-bundles\") pod \"331ad71b-521e-4b17-995b-0711b41d3b0e\" (UID: \"331ad71b-521e-4b17-995b-0711b41d3b0e\") " Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.613782 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/331ad71b-521e-4b17-995b-0711b41d3b0e-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "331ad71b-521e-4b17-995b-0711b41d3b0e" (UID: "331ad71b-521e-4b17-995b-0711b41d3b0e"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.614189 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/331ad71b-521e-4b17-995b-0711b41d3b0e-client-ca" (OuterVolumeSpecName: "client-ca") pod "331ad71b-521e-4b17-995b-0711b41d3b0e" (UID: "331ad71b-521e-4b17-995b-0711b41d3b0e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.614500 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/331ad71b-521e-4b17-995b-0711b41d3b0e-config" (OuterVolumeSpecName: "config") pod "331ad71b-521e-4b17-995b-0711b41d3b0e" (UID: "331ad71b-521e-4b17-995b-0711b41d3b0e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.618504 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/331ad71b-521e-4b17-995b-0711b41d3b0e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "331ad71b-521e-4b17-995b-0711b41d3b0e" (UID: "331ad71b-521e-4b17-995b-0711b41d3b0e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.618788 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/331ad71b-521e-4b17-995b-0711b41d3b0e-kube-api-access-xl6kg" (OuterVolumeSpecName: "kube-api-access-xl6kg") pod "331ad71b-521e-4b17-995b-0711b41d3b0e" (UID: "331ad71b-521e-4b17-995b-0711b41d3b0e"). InnerVolumeSpecName "kube-api-access-xl6kg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.714651 4950 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/331ad71b-521e-4b17-995b-0711b41d3b0e-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.714687 4950 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/331ad71b-521e-4b17-995b-0711b41d3b0e-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.714698 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/331ad71b-521e-4b17-995b-0711b41d3b0e-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.714707 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/331ad71b-521e-4b17-995b-0711b41d3b0e-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:02 crc kubenswrapper[4950]: I0318 20:11:02.714717 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xl6kg\" (UniqueName: \"kubernetes.io/projected/331ad71b-521e-4b17-995b-0711b41d3b0e-kube-api-access-xl6kg\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.008987 4950 generic.go:334] "Generic (PLEG): container finished" podID="f54c83ec-47ad-4ced-93c3-d321460f40f3" containerID="fd9b6b6486e4d512fa928e89ee3f4376f75210c39b053248471ee72768a8c097" exitCode=0 Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.009044 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fk4b9" event={"ID":"f54c83ec-47ad-4ced-93c3-d321460f40f3","Type":"ContainerDied","Data":"fd9b6b6486e4d512fa928e89ee3f4376f75210c39b053248471ee72768a8c097"} Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.013110 4950 generic.go:334] "Generic (PLEG): container finished" podID="97cb884c-5f55-4ca0-8639-193c2f1df139" containerID="d790f769ee6aced97a1576b8d64131e51abc022caf2a585431da9cc291585aeb" exitCode=0 Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.013176 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564408-2658d" event={"ID":"97cb884c-5f55-4ca0-8639-193c2f1df139","Type":"ContainerDied","Data":"d790f769ee6aced97a1576b8d64131e51abc022caf2a585431da9cc291585aeb"} Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.022008 4950 generic.go:334] "Generic (PLEG): container finished" podID="627e6f3d-cde5-4849-bd02-6437622f4978" containerID="00dffb6edda0aae584ee703172b0a41786cf67c059f901a0afd2d75aeedf585d" exitCode=0 Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.022066 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564410-9hqrd" event={"ID":"627e6f3d-cde5-4849-bd02-6437622f4978","Type":"ContainerDied","Data":"00dffb6edda0aae584ee703172b0a41786cf67c059f901a0afd2d75aeedf585d"} Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.030348 4950 generic.go:334] "Generic (PLEG): container finished" podID="331ad71b-521e-4b17-995b-0711b41d3b0e" containerID="e005608daa13eb6b654aa3e3c023165e2978653d6cfb8251c9a417e6a95ba96a" exitCode=0 Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.030984 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.031878 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" event={"ID":"331ad71b-521e-4b17-995b-0711b41d3b0e","Type":"ContainerDied","Data":"e005608daa13eb6b654aa3e3c023165e2978653d6cfb8251c9a417e6a95ba96a"} Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.031982 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7" event={"ID":"331ad71b-521e-4b17-995b-0711b41d3b0e","Type":"ContainerDied","Data":"b1b0b87723316764de411e8c98a78dfd1d24fb9d48d89b962b7f19d02669ff99"} Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.032005 4950 scope.go:117] "RemoveContainer" containerID="e005608daa13eb6b654aa3e3c023165e2978653d6cfb8251c9a417e6a95ba96a" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.048717 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" event={"ID":"21200e40-6c66-43d0-b8a2-1206361409ed","Type":"ContainerStarted","Data":"4f993329d64266f90dc7c7fd589333a0837921faa2d7d825b0bbc874fbbe8fae"} Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.049004 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" event={"ID":"21200e40-6c66-43d0-b8a2-1206361409ed","Type":"ContainerStarted","Data":"d87b134b60500cb2c32ee2b83ff97162fdebaec4c7707cb7c249673bb9867685"} Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.049075 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.075468 4950 scope.go:117] "RemoveContainer" containerID="e005608daa13eb6b654aa3e3c023165e2978653d6cfb8251c9a417e6a95ba96a" Mar 18 20:11:03 crc kubenswrapper[4950]: E0318 20:11:03.076116 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e005608daa13eb6b654aa3e3c023165e2978653d6cfb8251c9a417e6a95ba96a\": container with ID starting with e005608daa13eb6b654aa3e3c023165e2978653d6cfb8251c9a417e6a95ba96a not found: ID does not exist" containerID="e005608daa13eb6b654aa3e3c023165e2978653d6cfb8251c9a417e6a95ba96a" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.076147 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e005608daa13eb6b654aa3e3c023165e2978653d6cfb8251c9a417e6a95ba96a"} err="failed to get container status \"e005608daa13eb6b654aa3e3c023165e2978653d6cfb8251c9a417e6a95ba96a\": rpc error: code = NotFound desc = could not find container \"e005608daa13eb6b654aa3e3c023165e2978653d6cfb8251c9a417e6a95ba96a\": container with ID starting with e005608daa13eb6b654aa3e3c023165e2978653d6cfb8251c9a417e6a95ba96a not found: ID does not exist" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.090828 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" podStartSLOduration=10.090803288 podStartE2EDuration="10.090803288s" podCreationTimestamp="2026-03-18 20:10:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:11:03.084178717 +0000 UTC m=+276.325020585" watchObservedRunningTime="2026-03-18 20:11:03.090803288 +0000 UTC m=+276.331645166" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.102941 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.113710 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7"] Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.132304 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6c5b68fd4f-9qpg7"] Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.303157 4950 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-21 00:21:22.308696625 +0000 UTC Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.303184 4950 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 5932h10m19.005515143s for next certificate rotation Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.368210 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.489333 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="331ad71b-521e-4b17-995b-0711b41d3b0e" path="/var/lib/kubelet/pods/331ad71b-521e-4b17-995b-0711b41d3b0e/volumes" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.496900 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1" path="/var/lib/kubelet/pods/c4325fb8-e169-4c5a-9b31-fa46cb8cb8a1/volumes" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.527098 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0509451c-e8c8-4122-bf76-e51f16acbcc5-kube-api-access\") pod \"0509451c-e8c8-4122-bf76-e51f16acbcc5\" (UID: \"0509451c-e8c8-4122-bf76-e51f16acbcc5\") " Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.527195 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0509451c-e8c8-4122-bf76-e51f16acbcc5-kubelet-dir\") pod \"0509451c-e8c8-4122-bf76-e51f16acbcc5\" (UID: \"0509451c-e8c8-4122-bf76-e51f16acbcc5\") " Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.527528 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0509451c-e8c8-4122-bf76-e51f16acbcc5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0509451c-e8c8-4122-bf76-e51f16acbcc5" (UID: "0509451c-e8c8-4122-bf76-e51f16acbcc5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.532110 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0509451c-e8c8-4122-bf76-e51f16acbcc5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0509451c-e8c8-4122-bf76-e51f16acbcc5" (UID: "0509451c-e8c8-4122-bf76-e51f16acbcc5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.585612 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wv7kr"] Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.628152 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0509451c-e8c8-4122-bf76-e51f16acbcc5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.628184 4950 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0509451c-e8c8-4122-bf76-e51f16acbcc5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.675808 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5557678cd8-zwr5b"] Mar 18 20:11:03 crc kubenswrapper[4950]: E0318 20:11:03.676028 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="331ad71b-521e-4b17-995b-0711b41d3b0e" containerName="controller-manager" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.676041 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="331ad71b-521e-4b17-995b-0711b41d3b0e" containerName="controller-manager" Mar 18 20:11:03 crc kubenswrapper[4950]: E0318 20:11:03.676052 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0509451c-e8c8-4122-bf76-e51f16acbcc5" containerName="pruner" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.676058 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="0509451c-e8c8-4122-bf76-e51f16acbcc5" containerName="pruner" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.676161 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="0509451c-e8c8-4122-bf76-e51f16acbcc5" containerName="pruner" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.676175 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="331ad71b-521e-4b17-995b-0711b41d3b0e" containerName="controller-manager" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.676545 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.682585 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.684092 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.685735 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.693447 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.693922 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.694388 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.704779 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5557678cd8-zwr5b"] Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.705777 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.830914 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-proxy-ca-bundles\") pod \"controller-manager-5557678cd8-zwr5b\" (UID: \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\") " pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.830975 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnsqt\" (UniqueName: \"kubernetes.io/projected/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-kube-api-access-wnsqt\") pod \"controller-manager-5557678cd8-zwr5b\" (UID: \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\") " pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.831130 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-client-ca\") pod \"controller-manager-5557678cd8-zwr5b\" (UID: \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\") " pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.831226 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-serving-cert\") pod \"controller-manager-5557678cd8-zwr5b\" (UID: \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\") " pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.831268 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-config\") pod \"controller-manager-5557678cd8-zwr5b\" (UID: \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\") " pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.835847 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.835899 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.933394 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-proxy-ca-bundles\") pod \"controller-manager-5557678cd8-zwr5b\" (UID: \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\") " pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.933472 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnsqt\" (UniqueName: \"kubernetes.io/projected/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-kube-api-access-wnsqt\") pod \"controller-manager-5557678cd8-zwr5b\" (UID: \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\") " pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.933512 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-client-ca\") pod \"controller-manager-5557678cd8-zwr5b\" (UID: \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\") " pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.933539 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-serving-cert\") pod \"controller-manager-5557678cd8-zwr5b\" (UID: \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\") " pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.933566 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-config\") pod \"controller-manager-5557678cd8-zwr5b\" (UID: \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\") " pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.934757 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-config\") pod \"controller-manager-5557678cd8-zwr5b\" (UID: \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\") " pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.935476 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-proxy-ca-bundles\") pod \"controller-manager-5557678cd8-zwr5b\" (UID: \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\") " pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.936211 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-client-ca\") pod \"controller-manager-5557678cd8-zwr5b\" (UID: \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\") " pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.939857 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-serving-cert\") pod \"controller-manager-5557678cd8-zwr5b\" (UID: \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\") " pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.954509 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnsqt\" (UniqueName: \"kubernetes.io/projected/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-kube-api-access-wnsqt\") pod \"controller-manager-5557678cd8-zwr5b\" (UID: \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\") " pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" Mar 18 20:11:03 crc kubenswrapper[4950]: I0318 20:11:03.992334 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" Mar 18 20:11:04 crc kubenswrapper[4950]: I0318 20:11:04.063290 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fk4b9" event={"ID":"f54c83ec-47ad-4ced-93c3-d321460f40f3","Type":"ContainerStarted","Data":"f6e0189ec82ee1f0253a21b17a8af6e71e6310e4a6b96384395acf947e72497d"} Mar 18 20:11:04 crc kubenswrapper[4950]: I0318 20:11:04.067943 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 18 20:11:04 crc kubenswrapper[4950]: I0318 20:11:04.068174 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"0509451c-e8c8-4122-bf76-e51f16acbcc5","Type":"ContainerDied","Data":"a9dadff189717861421e11c8df7d4dc7caac40f42a20c40824135999e77e2015"} Mar 18 20:11:04 crc kubenswrapper[4950]: I0318 20:11:04.068212 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9dadff189717861421e11c8df7d4dc7caac40f42a20c40824135999e77e2015" Mar 18 20:11:04 crc kubenswrapper[4950]: I0318 20:11:04.087640 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fk4b9" podStartSLOduration=4.078123269 podStartE2EDuration="51.087471146s" podCreationTimestamp="2026-03-18 20:10:13 +0000 UTC" firstStartedPulling="2026-03-18 20:10:16.437850761 +0000 UTC m=+229.678692629" lastFinishedPulling="2026-03-18 20:11:03.447198648 +0000 UTC m=+276.688040506" observedRunningTime="2026-03-18 20:11:04.086536151 +0000 UTC m=+277.327378019" watchObservedRunningTime="2026-03-18 20:11:04.087471146 +0000 UTC m=+277.328313014" Mar 18 20:11:04 crc kubenswrapper[4950]: I0318 20:11:04.303921 4950 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-17 01:13:43.952831807 +0000 UTC Mar 18 20:11:04 crc kubenswrapper[4950]: I0318 20:11:04.303991 4950 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6557h2m39.648843763s for next certificate rotation Mar 18 20:11:04 crc kubenswrapper[4950]: I0318 20:11:04.350182 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564408-2658d" Mar 18 20:11:04 crc kubenswrapper[4950]: I0318 20:11:04.373493 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564410-9hqrd" Mar 18 20:11:04 crc kubenswrapper[4950]: I0318 20:11:04.444148 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjg8b\" (UniqueName: \"kubernetes.io/projected/97cb884c-5f55-4ca0-8639-193c2f1df139-kube-api-access-sjg8b\") pod \"97cb884c-5f55-4ca0-8639-193c2f1df139\" (UID: \"97cb884c-5f55-4ca0-8639-193c2f1df139\") " Mar 18 20:11:04 crc kubenswrapper[4950]: I0318 20:11:04.449815 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97cb884c-5f55-4ca0-8639-193c2f1df139-kube-api-access-sjg8b" (OuterVolumeSpecName: "kube-api-access-sjg8b") pod "97cb884c-5f55-4ca0-8639-193c2f1df139" (UID: "97cb884c-5f55-4ca0-8639-193c2f1df139"). InnerVolumeSpecName "kube-api-access-sjg8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:11:04 crc kubenswrapper[4950]: I0318 20:11:04.545398 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjf7l\" (UniqueName: \"kubernetes.io/projected/627e6f3d-cde5-4849-bd02-6437622f4978-kube-api-access-cjf7l\") pod \"627e6f3d-cde5-4849-bd02-6437622f4978\" (UID: \"627e6f3d-cde5-4849-bd02-6437622f4978\") " Mar 18 20:11:04 crc kubenswrapper[4950]: I0318 20:11:04.545786 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjg8b\" (UniqueName: \"kubernetes.io/projected/97cb884c-5f55-4ca0-8639-193c2f1df139-kube-api-access-sjg8b\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:04 crc kubenswrapper[4950]: I0318 20:11:04.654950 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/627e6f3d-cde5-4849-bd02-6437622f4978-kube-api-access-cjf7l" (OuterVolumeSpecName: "kube-api-access-cjf7l") pod "627e6f3d-cde5-4849-bd02-6437622f4978" (UID: "627e6f3d-cde5-4849-bd02-6437622f4978"). InnerVolumeSpecName "kube-api-access-cjf7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:11:04 crc kubenswrapper[4950]: I0318 20:11:04.682891 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5557678cd8-zwr5b"] Mar 18 20:11:04 crc kubenswrapper[4950]: W0318 20:11:04.689906 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c66a209_1abc_4b65_9e65_2d3ca0b807cf.slice/crio-49c4a68bc3be62a0f6c81131d27197f4fdbeb364aad469b07896c4dbda197a3f WatchSource:0}: Error finding container 49c4a68bc3be62a0f6c81131d27197f4fdbeb364aad469b07896c4dbda197a3f: Status 404 returned error can't find the container with id 49c4a68bc3be62a0f6c81131d27197f4fdbeb364aad469b07896c4dbda197a3f Mar 18 20:11:04 crc kubenswrapper[4950]: I0318 20:11:04.749661 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjf7l\" (UniqueName: \"kubernetes.io/projected/627e6f3d-cde5-4849-bd02-6437622f4978-kube-api-access-cjf7l\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:05 crc kubenswrapper[4950]: I0318 20:11:05.080238 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" event={"ID":"4c66a209-1abc-4b65-9e65-2d3ca0b807cf","Type":"ContainerStarted","Data":"ea3a073c628cb2df5d0872ab0355923e70b5c09b9822c861dd887a23d9c197d3"} Mar 18 20:11:05 crc kubenswrapper[4950]: I0318 20:11:05.080287 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" event={"ID":"4c66a209-1abc-4b65-9e65-2d3ca0b807cf","Type":"ContainerStarted","Data":"49c4a68bc3be62a0f6c81131d27197f4fdbeb364aad469b07896c4dbda197a3f"} Mar 18 20:11:05 crc kubenswrapper[4950]: I0318 20:11:05.080640 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" Mar 18 20:11:05 crc kubenswrapper[4950]: I0318 20:11:05.085305 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564408-2658d" Mar 18 20:11:05 crc kubenswrapper[4950]: I0318 20:11:05.085349 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564408-2658d" event={"ID":"97cb884c-5f55-4ca0-8639-193c2f1df139","Type":"ContainerDied","Data":"cb0d25f2e7d0a0dbdeebd645a69c14dd59c8f100deff662973fdcc0747e297ad"} Mar 18 20:11:05 crc kubenswrapper[4950]: I0318 20:11:05.085378 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb0d25f2e7d0a0dbdeebd645a69c14dd59c8f100deff662973fdcc0747e297ad" Mar 18 20:11:05 crc kubenswrapper[4950]: I0318 20:11:05.087786 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564410-9hqrd" event={"ID":"627e6f3d-cde5-4849-bd02-6437622f4978","Type":"ContainerDied","Data":"586f0b3af3e8f89f832874751a5f050777edfe372bea7a93c8e4a7e2a5768640"} Mar 18 20:11:05 crc kubenswrapper[4950]: I0318 20:11:05.087825 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="586f0b3af3e8f89f832874751a5f050777edfe372bea7a93c8e4a7e2a5768640" Mar 18 20:11:05 crc kubenswrapper[4950]: I0318 20:11:05.087794 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564410-9hqrd" Mar 18 20:11:05 crc kubenswrapper[4950]: I0318 20:11:05.123304 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" Mar 18 20:11:05 crc kubenswrapper[4950]: I0318 20:11:05.128718 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" podStartSLOduration=13.128680413 podStartE2EDuration="13.128680413s" podCreationTimestamp="2026-03-18 20:10:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:11:05.104512543 +0000 UTC m=+278.345354401" watchObservedRunningTime="2026-03-18 20:11:05.128680413 +0000 UTC m=+278.369522281" Mar 18 20:11:10 crc kubenswrapper[4950]: I0318 20:11:10.118583 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nb7fx" event={"ID":"33a99c2b-09d6-4ee7-b875-509ef5e144eb","Type":"ContainerStarted","Data":"786e809c5bbec43680ae9e1cb62efe0945f0c7a82d0c1f4e42c481597f4b2820"} Mar 18 20:11:11 crc kubenswrapper[4950]: I0318 20:11:11.124772 4950 generic.go:334] "Generic (PLEG): container finished" podID="33a99c2b-09d6-4ee7-b875-509ef5e144eb" containerID="786e809c5bbec43680ae9e1cb62efe0945f0c7a82d0c1f4e42c481597f4b2820" exitCode=0 Mar 18 20:11:11 crc kubenswrapper[4950]: I0318 20:11:11.124841 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nb7fx" event={"ID":"33a99c2b-09d6-4ee7-b875-509ef5e144eb","Type":"ContainerDied","Data":"786e809c5bbec43680ae9e1cb62efe0945f0c7a82d0c1f4e42c481597f4b2820"} Mar 18 20:11:11 crc kubenswrapper[4950]: I0318 20:11:11.127476 4950 generic.go:334] "Generic (PLEG): container finished" podID="bbeed1b2-6abb-49df-94d9-7d9fe6e5a563" containerID="633af43d8823abb96bf78609a47d9f3ff3e8c5ce993efdca967ceb91288247a5" exitCode=0 Mar 18 20:11:11 crc kubenswrapper[4950]: I0318 20:11:11.127519 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7dzz6" event={"ID":"bbeed1b2-6abb-49df-94d9-7d9fe6e5a563","Type":"ContainerDied","Data":"633af43d8823abb96bf78609a47d9f3ff3e8c5ce993efdca967ceb91288247a5"} Mar 18 20:11:11 crc kubenswrapper[4950]: I0318 20:11:11.129740 4950 generic.go:334] "Generic (PLEG): container finished" podID="8fe40723-62a6-4fe7-989c-b6f968467b4c" containerID="4858a2c5928016a18fcb073654870f8c2593355ffb79728629abb9f88912a4c0" exitCode=0 Mar 18 20:11:11 crc kubenswrapper[4950]: I0318 20:11:11.129771 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lhhrh" event={"ID":"8fe40723-62a6-4fe7-989c-b6f968467b4c","Type":"ContainerDied","Data":"4858a2c5928016a18fcb073654870f8c2593355ffb79728629abb9f88912a4c0"} Mar 18 20:11:12 crc kubenswrapper[4950]: I0318 20:11:12.143389 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwzv9" event={"ID":"82c68303-8af4-41a0-af35-312cc8b0bd7e","Type":"ContainerStarted","Data":"38907306b830da3b6e55507900c1a6932b547eca7618ff03aef91df92c2300b3"} Mar 18 20:11:12 crc kubenswrapper[4950]: I0318 20:11:12.981610 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5557678cd8-zwr5b"] Mar 18 20:11:12 crc kubenswrapper[4950]: I0318 20:11:12.982074 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" podUID="4c66a209-1abc-4b65-9e65-2d3ca0b807cf" containerName="controller-manager" containerID="cri-o://ea3a073c628cb2df5d0872ab0355923e70b5c09b9822c861dd887a23d9c197d3" gracePeriod=30 Mar 18 20:11:12 crc kubenswrapper[4950]: I0318 20:11:12.991313 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd"] Mar 18 20:11:12 crc kubenswrapper[4950]: I0318 20:11:12.991772 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" podUID="21200e40-6c66-43d0-b8a2-1206361409ed" containerName="route-controller-manager" containerID="cri-o://4f993329d64266f90dc7c7fd589333a0837921faa2d7d825b0bbc874fbbe8fae" gracePeriod=30 Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.151229 4950 generic.go:334] "Generic (PLEG): container finished" podID="4c66a209-1abc-4b65-9e65-2d3ca0b807cf" containerID="ea3a073c628cb2df5d0872ab0355923e70b5c09b9822c861dd887a23d9c197d3" exitCode=0 Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.151286 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" event={"ID":"4c66a209-1abc-4b65-9e65-2d3ca0b807cf","Type":"ContainerDied","Data":"ea3a073c628cb2df5d0872ab0355923e70b5c09b9822c861dd887a23d9c197d3"} Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.156057 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lhhrh" event={"ID":"8fe40723-62a6-4fe7-989c-b6f968467b4c","Type":"ContainerStarted","Data":"f8dcebf5f3960e992a7a9abf7429c75530bc03f926813c5fd243b4f6f03df668"} Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.161270 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nb7fx" event={"ID":"33a99c2b-09d6-4ee7-b875-509ef5e144eb","Type":"ContainerStarted","Data":"866301b46a0fa42f5ddd84e80734a99c7495f2e22720901af0847f2ada691dcb"} Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.163720 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7dzz6" event={"ID":"bbeed1b2-6abb-49df-94d9-7d9fe6e5a563","Type":"ContainerStarted","Data":"1fd9ccb388e588608b540567d2013892106bacdc049f5bcd80b2066229e558f3"} Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.169905 4950 generic.go:334] "Generic (PLEG): container finished" podID="82c68303-8af4-41a0-af35-312cc8b0bd7e" containerID="38907306b830da3b6e55507900c1a6932b547eca7618ff03aef91df92c2300b3" exitCode=0 Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.169955 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwzv9" event={"ID":"82c68303-8af4-41a0-af35-312cc8b0bd7e","Type":"ContainerDied","Data":"38907306b830da3b6e55507900c1a6932b547eca7618ff03aef91df92c2300b3"} Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.175266 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lhhrh" podStartSLOduration=5.076080114 podStartE2EDuration="58.175251278s" podCreationTimestamp="2026-03-18 20:10:15 +0000 UTC" firstStartedPulling="2026-03-18 20:10:18.942082751 +0000 UTC m=+232.182924619" lastFinishedPulling="2026-03-18 20:11:12.041253895 +0000 UTC m=+285.282095783" observedRunningTime="2026-03-18 20:11:13.173945312 +0000 UTC m=+286.414787180" watchObservedRunningTime="2026-03-18 20:11:13.175251278 +0000 UTC m=+286.416093146" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.176595 4950 generic.go:334] "Generic (PLEG): container finished" podID="21200e40-6c66-43d0-b8a2-1206361409ed" containerID="4f993329d64266f90dc7c7fd589333a0837921faa2d7d825b0bbc874fbbe8fae" exitCode=0 Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.176630 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" event={"ID":"21200e40-6c66-43d0-b8a2-1206361409ed","Type":"ContainerDied","Data":"4f993329d64266f90dc7c7fd589333a0837921faa2d7d825b0bbc874fbbe8fae"} Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.192690 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7dzz6" podStartSLOduration=3.939113193 podStartE2EDuration="58.192676714s" podCreationTimestamp="2026-03-18 20:10:15 +0000 UTC" firstStartedPulling="2026-03-18 20:10:17.736960958 +0000 UTC m=+230.977802826" lastFinishedPulling="2026-03-18 20:11:11.990524479 +0000 UTC m=+285.231366347" observedRunningTime="2026-03-18 20:11:13.190133305 +0000 UTC m=+286.430975173" watchObservedRunningTime="2026-03-18 20:11:13.192676714 +0000 UTC m=+286.433518572" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.234118 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nb7fx" podStartSLOduration=3.9806429420000002 podStartE2EDuration="57.234103516s" podCreationTimestamp="2026-03-18 20:10:16 +0000 UTC" firstStartedPulling="2026-03-18 20:10:18.953212532 +0000 UTC m=+232.194054400" lastFinishedPulling="2026-03-18 20:11:12.206673106 +0000 UTC m=+285.447514974" observedRunningTime="2026-03-18 20:11:13.217900533 +0000 UTC m=+286.458742401" watchObservedRunningTime="2026-03-18 20:11:13.234103516 +0000 UTC m=+286.474945384" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.595125 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.648656 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.672184 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxzkk\" (UniqueName: \"kubernetes.io/projected/21200e40-6c66-43d0-b8a2-1206361409ed-kube-api-access-xxzkk\") pod \"21200e40-6c66-43d0-b8a2-1206361409ed\" (UID: \"21200e40-6c66-43d0-b8a2-1206361409ed\") " Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.676905 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21200e40-6c66-43d0-b8a2-1206361409ed-client-ca" (OuterVolumeSpecName: "client-ca") pod "21200e40-6c66-43d0-b8a2-1206361409ed" (UID: "21200e40-6c66-43d0-b8a2-1206361409ed"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.672258 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21200e40-6c66-43d0-b8a2-1206361409ed-client-ca\") pod \"21200e40-6c66-43d0-b8a2-1206361409ed\" (UID: \"21200e40-6c66-43d0-b8a2-1206361409ed\") " Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.698440 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21200e40-6c66-43d0-b8a2-1206361409ed-serving-cert\") pod \"21200e40-6c66-43d0-b8a2-1206361409ed\" (UID: \"21200e40-6c66-43d0-b8a2-1206361409ed\") " Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.698506 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21200e40-6c66-43d0-b8a2-1206361409ed-config\") pod \"21200e40-6c66-43d0-b8a2-1206361409ed\" (UID: \"21200e40-6c66-43d0-b8a2-1206361409ed\") " Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.699135 4950 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21200e40-6c66-43d0-b8a2-1206361409ed-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.705920 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21200e40-6c66-43d0-b8a2-1206361409ed-kube-api-access-xxzkk" (OuterVolumeSpecName: "kube-api-access-xxzkk") pod "21200e40-6c66-43d0-b8a2-1206361409ed" (UID: "21200e40-6c66-43d0-b8a2-1206361409ed"). InnerVolumeSpecName "kube-api-access-xxzkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.706546 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21200e40-6c66-43d0-b8a2-1206361409ed-config" (OuterVolumeSpecName: "config") pod "21200e40-6c66-43d0-b8a2-1206361409ed" (UID: "21200e40-6c66-43d0-b8a2-1206361409ed"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.712673 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21200e40-6c66-43d0-b8a2-1206361409ed-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "21200e40-6c66-43d0-b8a2-1206361409ed" (UID: "21200e40-6c66-43d0-b8a2-1206361409ed"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.802180 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-serving-cert\") pod \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\" (UID: \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\") " Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.802243 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-client-ca\") pod \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\" (UID: \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\") " Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.802306 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-config\") pod \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\" (UID: \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\") " Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.802324 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-proxy-ca-bundles\") pod \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\" (UID: \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\") " Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.802391 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnsqt\" (UniqueName: \"kubernetes.io/projected/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-kube-api-access-wnsqt\") pod \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\" (UID: \"4c66a209-1abc-4b65-9e65-2d3ca0b807cf\") " Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.803171 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-client-ca" (OuterVolumeSpecName: "client-ca") pod "4c66a209-1abc-4b65-9e65-2d3ca0b807cf" (UID: "4c66a209-1abc-4b65-9e65-2d3ca0b807cf"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.803267 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-config" (OuterVolumeSpecName: "config") pod "4c66a209-1abc-4b65-9e65-2d3ca0b807cf" (UID: "4c66a209-1abc-4b65-9e65-2d3ca0b807cf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.803529 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "4c66a209-1abc-4b65-9e65-2d3ca0b807cf" (UID: "4c66a209-1abc-4b65-9e65-2d3ca0b807cf"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.806636 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.806797 4950 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.806891 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxzkk\" (UniqueName: \"kubernetes.io/projected/21200e40-6c66-43d0-b8a2-1206361409ed-kube-api-access-xxzkk\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.806973 4950 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.807056 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21200e40-6c66-43d0-b8a2-1206361409ed-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.807140 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21200e40-6c66-43d0-b8a2-1206361409ed-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.807197 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4c66a209-1abc-4b65-9e65-2d3ca0b807cf" (UID: "4c66a209-1abc-4b65-9e65-2d3ca0b807cf"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.808978 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-kube-api-access-wnsqt" (OuterVolumeSpecName: "kube-api-access-wnsqt") pod "4c66a209-1abc-4b65-9e65-2d3ca0b807cf" (UID: "4c66a209-1abc-4b65-9e65-2d3ca0b807cf"). InnerVolumeSpecName "kube-api-access-wnsqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.908521 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnsqt\" (UniqueName: \"kubernetes.io/projected/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-kube-api-access-wnsqt\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.908710 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c66a209-1abc-4b65-9e65-2d3ca0b807cf-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.961557 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fk4b9" Mar 18 20:11:13 crc kubenswrapper[4950]: I0318 20:11:13.962041 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fk4b9" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.138309 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fk4b9" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.183570 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.183597 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5557678cd8-zwr5b" event={"ID":"4c66a209-1abc-4b65-9e65-2d3ca0b807cf","Type":"ContainerDied","Data":"49c4a68bc3be62a0f6c81131d27197f4fdbeb364aad469b07896c4dbda197a3f"} Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.183652 4950 scope.go:117] "RemoveContainer" containerID="ea3a073c628cb2df5d0872ab0355923e70b5c09b9822c861dd887a23d9c197d3" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.190465 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwzv9" event={"ID":"82c68303-8af4-41a0-af35-312cc8b0bd7e","Type":"ContainerStarted","Data":"6526d85d276087eb39ac4ab3d4a0cae6048b554628fa67a4a9f6f1f70721cdd0"} Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.196078 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" event={"ID":"21200e40-6c66-43d0-b8a2-1206361409ed","Type":"ContainerDied","Data":"d87b134b60500cb2c32ee2b83ff97162fdebaec4c7707cb7c249673bb9867685"} Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.196197 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.198306 4950 scope.go:117] "RemoveContainer" containerID="4f993329d64266f90dc7c7fd589333a0837921faa2d7d825b0bbc874fbbe8fae" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.227878 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rwzv9" podStartSLOduration=3.991574367 podStartE2EDuration="1m1.227859116s" podCreationTimestamp="2026-03-18 20:10:13 +0000 UTC" firstStartedPulling="2026-03-18 20:10:16.567022224 +0000 UTC m=+229.807864092" lastFinishedPulling="2026-03-18 20:11:13.803306973 +0000 UTC m=+287.044148841" observedRunningTime="2026-03-18 20:11:14.219009854 +0000 UTC m=+287.459851722" watchObservedRunningTime="2026-03-18 20:11:14.227859116 +0000 UTC m=+287.468700984" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.229720 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd"] Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.237456 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cc5465bc6-b9kcd"] Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.240142 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5557678cd8-zwr5b"] Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.243715 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5557678cd8-zwr5b"] Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.244594 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fk4b9" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.669617 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-664567ccff-7rdnp"] Mar 18 20:11:14 crc kubenswrapper[4950]: E0318 20:11:14.669822 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97cb884c-5f55-4ca0-8639-193c2f1df139" containerName="oc" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.669837 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="97cb884c-5f55-4ca0-8639-193c2f1df139" containerName="oc" Mar 18 20:11:14 crc kubenswrapper[4950]: E0318 20:11:14.669852 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c66a209-1abc-4b65-9e65-2d3ca0b807cf" containerName="controller-manager" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.669861 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c66a209-1abc-4b65-9e65-2d3ca0b807cf" containerName="controller-manager" Mar 18 20:11:14 crc kubenswrapper[4950]: E0318 20:11:14.669876 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21200e40-6c66-43d0-b8a2-1206361409ed" containerName="route-controller-manager" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.669883 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="21200e40-6c66-43d0-b8a2-1206361409ed" containerName="route-controller-manager" Mar 18 20:11:14 crc kubenswrapper[4950]: E0318 20:11:14.669897 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="627e6f3d-cde5-4849-bd02-6437622f4978" containerName="oc" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.669903 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="627e6f3d-cde5-4849-bd02-6437622f4978" containerName="oc" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.669988 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="21200e40-6c66-43d0-b8a2-1206361409ed" containerName="route-controller-manager" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.670004 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="627e6f3d-cde5-4849-bd02-6437622f4978" containerName="oc" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.670011 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c66a209-1abc-4b65-9e65-2d3ca0b807cf" containerName="controller-manager" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.670018 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="97cb884c-5f55-4ca0-8639-193c2f1df139" containerName="oc" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.670359 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.672074 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.674247 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.674317 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.674502 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.674568 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.674954 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.675654 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr"] Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.676281 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.679856 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.681395 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.681499 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.681582 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.681796 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.681808 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.687036 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.689036 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-664567ccff-7rdnp"] Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.697855 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr"] Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.819365 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h42ds\" (UniqueName: \"kubernetes.io/projected/26289d96-c077-4b10-9139-09231241e9d0-kube-api-access-h42ds\") pod \"controller-manager-664567ccff-7rdnp\" (UID: \"26289d96-c077-4b10-9139-09231241e9d0\") " pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.819467 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26289d96-c077-4b10-9139-09231241e9d0-serving-cert\") pod \"controller-manager-664567ccff-7rdnp\" (UID: \"26289d96-c077-4b10-9139-09231241e9d0\") " pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.819495 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa7ec310-f833-410f-9184-0c9b910fe20f-client-ca\") pod \"route-controller-manager-7cf7cfb578-tpqwr\" (UID: \"fa7ec310-f833-410f-9184-0c9b910fe20f\") " pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.819511 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwsq2\" (UniqueName: \"kubernetes.io/projected/fa7ec310-f833-410f-9184-0c9b910fe20f-kube-api-access-cwsq2\") pod \"route-controller-manager-7cf7cfb578-tpqwr\" (UID: \"fa7ec310-f833-410f-9184-0c9b910fe20f\") " pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.819549 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/26289d96-c077-4b10-9139-09231241e9d0-proxy-ca-bundles\") pod \"controller-manager-664567ccff-7rdnp\" (UID: \"26289d96-c077-4b10-9139-09231241e9d0\") " pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.819569 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/26289d96-c077-4b10-9139-09231241e9d0-client-ca\") pod \"controller-manager-664567ccff-7rdnp\" (UID: \"26289d96-c077-4b10-9139-09231241e9d0\") " pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.819676 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26289d96-c077-4b10-9139-09231241e9d0-config\") pod \"controller-manager-664567ccff-7rdnp\" (UID: \"26289d96-c077-4b10-9139-09231241e9d0\") " pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.819704 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa7ec310-f833-410f-9184-0c9b910fe20f-config\") pod \"route-controller-manager-7cf7cfb578-tpqwr\" (UID: \"fa7ec310-f833-410f-9184-0c9b910fe20f\") " pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.819753 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa7ec310-f833-410f-9184-0c9b910fe20f-serving-cert\") pod \"route-controller-manager-7cf7cfb578-tpqwr\" (UID: \"fa7ec310-f833-410f-9184-0c9b910fe20f\") " pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.921380 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h42ds\" (UniqueName: \"kubernetes.io/projected/26289d96-c077-4b10-9139-09231241e9d0-kube-api-access-h42ds\") pod \"controller-manager-664567ccff-7rdnp\" (UID: \"26289d96-c077-4b10-9139-09231241e9d0\") " pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.921464 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26289d96-c077-4b10-9139-09231241e9d0-serving-cert\") pod \"controller-manager-664567ccff-7rdnp\" (UID: \"26289d96-c077-4b10-9139-09231241e9d0\") " pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.921487 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa7ec310-f833-410f-9184-0c9b910fe20f-client-ca\") pod \"route-controller-manager-7cf7cfb578-tpqwr\" (UID: \"fa7ec310-f833-410f-9184-0c9b910fe20f\") " pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.921502 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwsq2\" (UniqueName: \"kubernetes.io/projected/fa7ec310-f833-410f-9184-0c9b910fe20f-kube-api-access-cwsq2\") pod \"route-controller-manager-7cf7cfb578-tpqwr\" (UID: \"fa7ec310-f833-410f-9184-0c9b910fe20f\") " pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.921541 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/26289d96-c077-4b10-9139-09231241e9d0-proxy-ca-bundles\") pod \"controller-manager-664567ccff-7rdnp\" (UID: \"26289d96-c077-4b10-9139-09231241e9d0\") " pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.921564 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/26289d96-c077-4b10-9139-09231241e9d0-client-ca\") pod \"controller-manager-664567ccff-7rdnp\" (UID: \"26289d96-c077-4b10-9139-09231241e9d0\") " pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.921927 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26289d96-c077-4b10-9139-09231241e9d0-config\") pod \"controller-manager-664567ccff-7rdnp\" (UID: \"26289d96-c077-4b10-9139-09231241e9d0\") " pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.921974 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa7ec310-f833-410f-9184-0c9b910fe20f-config\") pod \"route-controller-manager-7cf7cfb578-tpqwr\" (UID: \"fa7ec310-f833-410f-9184-0c9b910fe20f\") " pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.922617 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa7ec310-f833-410f-9184-0c9b910fe20f-client-ca\") pod \"route-controller-manager-7cf7cfb578-tpqwr\" (UID: \"fa7ec310-f833-410f-9184-0c9b910fe20f\") " pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.922630 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/26289d96-c077-4b10-9139-09231241e9d0-client-ca\") pod \"controller-manager-664567ccff-7rdnp\" (UID: \"26289d96-c077-4b10-9139-09231241e9d0\") " pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.922922 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/26289d96-c077-4b10-9139-09231241e9d0-proxy-ca-bundles\") pod \"controller-manager-664567ccff-7rdnp\" (UID: \"26289d96-c077-4b10-9139-09231241e9d0\") " pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.923326 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26289d96-c077-4b10-9139-09231241e9d0-config\") pod \"controller-manager-664567ccff-7rdnp\" (UID: \"26289d96-c077-4b10-9139-09231241e9d0\") " pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.923475 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa7ec310-f833-410f-9184-0c9b910fe20f-serving-cert\") pod \"route-controller-manager-7cf7cfb578-tpqwr\" (UID: \"fa7ec310-f833-410f-9184-0c9b910fe20f\") " pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.924328 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa7ec310-f833-410f-9184-0c9b910fe20f-config\") pod \"route-controller-manager-7cf7cfb578-tpqwr\" (UID: \"fa7ec310-f833-410f-9184-0c9b910fe20f\") " pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.930065 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa7ec310-f833-410f-9184-0c9b910fe20f-serving-cert\") pod \"route-controller-manager-7cf7cfb578-tpqwr\" (UID: \"fa7ec310-f833-410f-9184-0c9b910fe20f\") " pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.939944 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26289d96-c077-4b10-9139-09231241e9d0-serving-cert\") pod \"controller-manager-664567ccff-7rdnp\" (UID: \"26289d96-c077-4b10-9139-09231241e9d0\") " pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.944045 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h42ds\" (UniqueName: \"kubernetes.io/projected/26289d96-c077-4b10-9139-09231241e9d0-kube-api-access-h42ds\") pod \"controller-manager-664567ccff-7rdnp\" (UID: \"26289d96-c077-4b10-9139-09231241e9d0\") " pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.945052 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwsq2\" (UniqueName: \"kubernetes.io/projected/fa7ec310-f833-410f-9184-0c9b910fe20f-kube-api-access-cwsq2\") pod \"route-controller-manager-7cf7cfb578-tpqwr\" (UID: \"fa7ec310-f833-410f-9184-0c9b910fe20f\") " pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.986992 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" Mar 18 20:11:14 crc kubenswrapper[4950]: I0318 20:11:14.997052 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" Mar 18 20:11:15 crc kubenswrapper[4950]: I0318 20:11:15.238426 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-664567ccff-7rdnp"] Mar 18 20:11:15 crc kubenswrapper[4950]: W0318 20:11:15.253007 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26289d96_c077_4b10_9139_09231241e9d0.slice/crio-8d42d04b351183ca696848df231c2926996d7e1cafda3b5c825d66fdfb915055 WatchSource:0}: Error finding container 8d42d04b351183ca696848df231c2926996d7e1cafda3b5c825d66fdfb915055: Status 404 returned error can't find the container with id 8d42d04b351183ca696848df231c2926996d7e1cafda3b5c825d66fdfb915055 Mar 18 20:11:15 crc kubenswrapper[4950]: I0318 20:11:15.386056 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7dzz6" Mar 18 20:11:15 crc kubenswrapper[4950]: I0318 20:11:15.386370 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7dzz6" Mar 18 20:11:15 crc kubenswrapper[4950]: I0318 20:11:15.425130 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7dzz6" Mar 18 20:11:15 crc kubenswrapper[4950]: I0318 20:11:15.485720 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21200e40-6c66-43d0-b8a2-1206361409ed" path="/var/lib/kubelet/pods/21200e40-6c66-43d0-b8a2-1206361409ed/volumes" Mar 18 20:11:15 crc kubenswrapper[4950]: I0318 20:11:15.486559 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c66a209-1abc-4b65-9e65-2d3ca0b807cf" path="/var/lib/kubelet/pods/4c66a209-1abc-4b65-9e65-2d3ca0b807cf/volumes" Mar 18 20:11:15 crc kubenswrapper[4950]: I0318 20:11:15.494852 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr"] Mar 18 20:11:15 crc kubenswrapper[4950]: I0318 20:11:15.840446 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lhhrh" Mar 18 20:11:15 crc kubenswrapper[4950]: I0318 20:11:15.840770 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lhhrh" Mar 18 20:11:15 crc kubenswrapper[4950]: I0318 20:11:15.942650 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lhhrh" Mar 18 20:11:16 crc kubenswrapper[4950]: I0318 20:11:16.218101 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fnpx" event={"ID":"d3ae2e86-be5d-4546-97a0-88301381868b","Type":"ContainerStarted","Data":"7308adae68168391c9ddefc29042a8b6c7245e1cda57a4b8f90ea857c8fe20a3"} Mar 18 20:11:16 crc kubenswrapper[4950]: I0318 20:11:16.220563 4950 generic.go:334] "Generic (PLEG): container finished" podID="044d9d2a-6b99-49ee-b00d-728c7832815c" containerID="e2be46f0ac137f088abf60e3ad3627ab19697959bf03f1e4948fe205b6b7016c" exitCode=0 Mar 18 20:11:16 crc kubenswrapper[4950]: I0318 20:11:16.220632 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gzj5q" event={"ID":"044d9d2a-6b99-49ee-b00d-728c7832815c","Type":"ContainerDied","Data":"e2be46f0ac137f088abf60e3ad3627ab19697959bf03f1e4948fe205b6b7016c"} Mar 18 20:11:16 crc kubenswrapper[4950]: I0318 20:11:16.225289 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" event={"ID":"fa7ec310-f833-410f-9184-0c9b910fe20f","Type":"ContainerStarted","Data":"c1e84b4e561754ff798344621025273750eb6771b74ff9fd4ba00dadd87421bf"} Mar 18 20:11:16 crc kubenswrapper[4950]: I0318 20:11:16.225337 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" event={"ID":"fa7ec310-f833-410f-9184-0c9b910fe20f","Type":"ContainerStarted","Data":"d389c536f83e1c5012c761678d56f6b15a647db002901c23716dc059707e465a"} Mar 18 20:11:16 crc kubenswrapper[4950]: I0318 20:11:16.226223 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" Mar 18 20:11:16 crc kubenswrapper[4950]: I0318 20:11:16.229020 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" event={"ID":"26289d96-c077-4b10-9139-09231241e9d0","Type":"ContainerStarted","Data":"0c33b6169085d2c06985cd0bc6eff75cbf956e5cf0a877495954750461b8ef53"} Mar 18 20:11:16 crc kubenswrapper[4950]: I0318 20:11:16.229056 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" event={"ID":"26289d96-c077-4b10-9139-09231241e9d0","Type":"ContainerStarted","Data":"8d42d04b351183ca696848df231c2926996d7e1cafda3b5c825d66fdfb915055"} Mar 18 20:11:16 crc kubenswrapper[4950]: I0318 20:11:16.279957 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" podStartSLOduration=3.2799422 podStartE2EDuration="3.2799422s" podCreationTimestamp="2026-03-18 20:11:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:11:16.276630589 +0000 UTC m=+289.517472457" watchObservedRunningTime="2026-03-18 20:11:16.2799422 +0000 UTC m=+289.520784068" Mar 18 20:11:16 crc kubenswrapper[4950]: I0318 20:11:16.504955 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" podStartSLOduration=3.504938619 podStartE2EDuration="3.504938619s" podCreationTimestamp="2026-03-18 20:11:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:11:16.301505429 +0000 UTC m=+289.542347317" watchObservedRunningTime="2026-03-18 20:11:16.504938619 +0000 UTC m=+289.745780487" Mar 18 20:11:16 crc kubenswrapper[4950]: I0318 20:11:16.614907 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" Mar 18 20:11:16 crc kubenswrapper[4950]: I0318 20:11:16.638432 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nb7fx" Mar 18 20:11:16 crc kubenswrapper[4950]: I0318 20:11:16.639247 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nb7fx" Mar 18 20:11:17 crc kubenswrapper[4950]: I0318 20:11:17.239150 4950 generic.go:334] "Generic (PLEG): container finished" podID="d3ae2e86-be5d-4546-97a0-88301381868b" containerID="7308adae68168391c9ddefc29042a8b6c7245e1cda57a4b8f90ea857c8fe20a3" exitCode=0 Mar 18 20:11:17 crc kubenswrapper[4950]: I0318 20:11:17.239304 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fnpx" event={"ID":"d3ae2e86-be5d-4546-97a0-88301381868b","Type":"ContainerDied","Data":"7308adae68168391c9ddefc29042a8b6c7245e1cda57a4b8f90ea857c8fe20a3"} Mar 18 20:11:17 crc kubenswrapper[4950]: I0318 20:11:17.239944 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" Mar 18 20:11:17 crc kubenswrapper[4950]: I0318 20:11:17.246655 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" Mar 18 20:11:17 crc kubenswrapper[4950]: I0318 20:11:17.525761 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fk4b9"] Mar 18 20:11:17 crc kubenswrapper[4950]: I0318 20:11:17.526177 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fk4b9" podUID="f54c83ec-47ad-4ced-93c3-d321460f40f3" containerName="registry-server" containerID="cri-o://f6e0189ec82ee1f0253a21b17a8af6e71e6310e4a6b96384395acf947e72497d" gracePeriod=2 Mar 18 20:11:17 crc kubenswrapper[4950]: I0318 20:11:17.696388 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nb7fx" podUID="33a99c2b-09d6-4ee7-b875-509ef5e144eb" containerName="registry-server" probeResult="failure" output=< Mar 18 20:11:17 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 20:11:17 crc kubenswrapper[4950]: > Mar 18 20:11:18 crc kubenswrapper[4950]: I0318 20:11:18.249778 4950 generic.go:334] "Generic (PLEG): container finished" podID="f54c83ec-47ad-4ced-93c3-d321460f40f3" containerID="f6e0189ec82ee1f0253a21b17a8af6e71e6310e4a6b96384395acf947e72497d" exitCode=0 Mar 18 20:11:18 crc kubenswrapper[4950]: I0318 20:11:18.249877 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fk4b9" event={"ID":"f54c83ec-47ad-4ced-93c3-d321460f40f3","Type":"ContainerDied","Data":"f6e0189ec82ee1f0253a21b17a8af6e71e6310e4a6b96384395acf947e72497d"} Mar 18 20:11:18 crc kubenswrapper[4950]: I0318 20:11:18.997642 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fk4b9" Mar 18 20:11:19 crc kubenswrapper[4950]: I0318 20:11:19.089054 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9s64w\" (UniqueName: \"kubernetes.io/projected/f54c83ec-47ad-4ced-93c3-d321460f40f3-kube-api-access-9s64w\") pod \"f54c83ec-47ad-4ced-93c3-d321460f40f3\" (UID: \"f54c83ec-47ad-4ced-93c3-d321460f40f3\") " Mar 18 20:11:19 crc kubenswrapper[4950]: I0318 20:11:19.089112 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f54c83ec-47ad-4ced-93c3-d321460f40f3-utilities\") pod \"f54c83ec-47ad-4ced-93c3-d321460f40f3\" (UID: \"f54c83ec-47ad-4ced-93c3-d321460f40f3\") " Mar 18 20:11:19 crc kubenswrapper[4950]: I0318 20:11:19.089190 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f54c83ec-47ad-4ced-93c3-d321460f40f3-catalog-content\") pod \"f54c83ec-47ad-4ced-93c3-d321460f40f3\" (UID: \"f54c83ec-47ad-4ced-93c3-d321460f40f3\") " Mar 18 20:11:19 crc kubenswrapper[4950]: I0318 20:11:19.090544 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f54c83ec-47ad-4ced-93c3-d321460f40f3-utilities" (OuterVolumeSpecName: "utilities") pod "f54c83ec-47ad-4ced-93c3-d321460f40f3" (UID: "f54c83ec-47ad-4ced-93c3-d321460f40f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:11:19 crc kubenswrapper[4950]: I0318 20:11:19.109450 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f54c83ec-47ad-4ced-93c3-d321460f40f3-kube-api-access-9s64w" (OuterVolumeSpecName: "kube-api-access-9s64w") pod "f54c83ec-47ad-4ced-93c3-d321460f40f3" (UID: "f54c83ec-47ad-4ced-93c3-d321460f40f3"). InnerVolumeSpecName "kube-api-access-9s64w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:11:19 crc kubenswrapper[4950]: I0318 20:11:19.150576 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f54c83ec-47ad-4ced-93c3-d321460f40f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f54c83ec-47ad-4ced-93c3-d321460f40f3" (UID: "f54c83ec-47ad-4ced-93c3-d321460f40f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:11:19 crc kubenswrapper[4950]: I0318 20:11:19.190769 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f54c83ec-47ad-4ced-93c3-d321460f40f3-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:19 crc kubenswrapper[4950]: I0318 20:11:19.191070 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f54c83ec-47ad-4ced-93c3-d321460f40f3-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:19 crc kubenswrapper[4950]: I0318 20:11:19.191146 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9s64w\" (UniqueName: \"kubernetes.io/projected/f54c83ec-47ad-4ced-93c3-d321460f40f3-kube-api-access-9s64w\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:19 crc kubenswrapper[4950]: I0318 20:11:19.260253 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fk4b9" Mar 18 20:11:19 crc kubenswrapper[4950]: I0318 20:11:19.263528 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fk4b9" event={"ID":"f54c83ec-47ad-4ced-93c3-d321460f40f3","Type":"ContainerDied","Data":"27272da5fedb4e5ec9815272afeb74c757e15e0cd40694777c5c91cbdb93776e"} Mar 18 20:11:19 crc kubenswrapper[4950]: I0318 20:11:19.263683 4950 scope.go:117] "RemoveContainer" containerID="f6e0189ec82ee1f0253a21b17a8af6e71e6310e4a6b96384395acf947e72497d" Mar 18 20:11:19 crc kubenswrapper[4950]: I0318 20:11:19.294182 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fk4b9"] Mar 18 20:11:19 crc kubenswrapper[4950]: I0318 20:11:19.299750 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fk4b9"] Mar 18 20:11:19 crc kubenswrapper[4950]: I0318 20:11:19.486380 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f54c83ec-47ad-4ced-93c3-d321460f40f3" path="/var/lib/kubelet/pods/f54c83ec-47ad-4ced-93c3-d321460f40f3/volumes" Mar 18 20:11:20 crc kubenswrapper[4950]: I0318 20:11:20.258003 4950 scope.go:117] "RemoveContainer" containerID="fd9b6b6486e4d512fa928e89ee3f4376f75210c39b053248471ee72768a8c097" Mar 18 20:11:20 crc kubenswrapper[4950]: I0318 20:11:20.690283 4950 scope.go:117] "RemoveContainer" containerID="8948c35590cdaa069e6a2360b8e142f7c41951aa6653788206319998f9803392" Mar 18 20:11:24 crc kubenswrapper[4950]: I0318 20:11:24.046763 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rwzv9" Mar 18 20:11:24 crc kubenswrapper[4950]: I0318 20:11:24.047442 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rwzv9" Mar 18 20:11:24 crc kubenswrapper[4950]: I0318 20:11:24.107937 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rwzv9" Mar 18 20:11:24 crc kubenswrapper[4950]: I0318 20:11:24.314830 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gzj5q" event={"ID":"044d9d2a-6b99-49ee-b00d-728c7832815c","Type":"ContainerStarted","Data":"a9132437997afe87402a5c278ffd8f59b66a8f2a8fa686e53a6c74092124a735"} Mar 18 20:11:24 crc kubenswrapper[4950]: I0318 20:11:24.331378 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gzj5q" podStartSLOduration=5.579331627 podStartE2EDuration="1m11.331361481s" podCreationTimestamp="2026-03-18 20:10:13 +0000 UTC" firstStartedPulling="2026-03-18 20:10:16.602915641 +0000 UTC m=+229.843757499" lastFinishedPulling="2026-03-18 20:11:22.354945485 +0000 UTC m=+295.595787353" observedRunningTime="2026-03-18 20:11:24.32985426 +0000 UTC m=+297.570696138" watchObservedRunningTime="2026-03-18 20:11:24.331361481 +0000 UTC m=+297.572203349" Mar 18 20:11:24 crc kubenswrapper[4950]: I0318 20:11:24.360614 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rwzv9" Mar 18 20:11:25 crc kubenswrapper[4950]: I0318 20:11:25.344445 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rwzv9"] Mar 18 20:11:25 crc kubenswrapper[4950]: I0318 20:11:25.439936 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7dzz6" Mar 18 20:11:25 crc kubenswrapper[4950]: I0318 20:11:25.884318 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lhhrh" Mar 18 20:11:26 crc kubenswrapper[4950]: I0318 20:11:26.325176 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rwzv9" podUID="82c68303-8af4-41a0-af35-312cc8b0bd7e" containerName="registry-server" containerID="cri-o://6526d85d276087eb39ac4ab3d4a0cae6048b554628fa67a4a9f6f1f70721cdd0" gracePeriod=2 Mar 18 20:11:26 crc kubenswrapper[4950]: I0318 20:11:26.699267 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nb7fx" Mar 18 20:11:26 crc kubenswrapper[4950]: I0318 20:11:26.747378 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nb7fx" Mar 18 20:11:27 crc kubenswrapper[4950]: I0318 20:11:27.337241 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fnpx" event={"ID":"d3ae2e86-be5d-4546-97a0-88301381868b","Type":"ContainerStarted","Data":"3752f560059d7127c4b227d082fedc6224ab2e2c7a96fab6dd6f9ab0b1f652d1"} Mar 18 20:11:27 crc kubenswrapper[4950]: I0318 20:11:27.340744 4950 generic.go:334] "Generic (PLEG): container finished" podID="82c68303-8af4-41a0-af35-312cc8b0bd7e" containerID="6526d85d276087eb39ac4ab3d4a0cae6048b554628fa67a4a9f6f1f70721cdd0" exitCode=0 Mar 18 20:11:27 crc kubenswrapper[4950]: I0318 20:11:27.341611 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwzv9" event={"ID":"82c68303-8af4-41a0-af35-312cc8b0bd7e","Type":"ContainerDied","Data":"6526d85d276087eb39ac4ab3d4a0cae6048b554628fa67a4a9f6f1f70721cdd0"} Mar 18 20:11:27 crc kubenswrapper[4950]: I0318 20:11:27.742667 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rwzv9" Mar 18 20:11:27 crc kubenswrapper[4950]: I0318 20:11:27.743386 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lhhrh"] Mar 18 20:11:27 crc kubenswrapper[4950]: I0318 20:11:27.743696 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lhhrh" podUID="8fe40723-62a6-4fe7-989c-b6f968467b4c" containerName="registry-server" containerID="cri-o://f8dcebf5f3960e992a7a9abf7429c75530bc03f926813c5fd243b4f6f03df668" gracePeriod=2 Mar 18 20:11:27 crc kubenswrapper[4950]: I0318 20:11:27.821067 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swzmx\" (UniqueName: \"kubernetes.io/projected/82c68303-8af4-41a0-af35-312cc8b0bd7e-kube-api-access-swzmx\") pod \"82c68303-8af4-41a0-af35-312cc8b0bd7e\" (UID: \"82c68303-8af4-41a0-af35-312cc8b0bd7e\") " Mar 18 20:11:27 crc kubenswrapper[4950]: I0318 20:11:27.821113 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82c68303-8af4-41a0-af35-312cc8b0bd7e-utilities\") pod \"82c68303-8af4-41a0-af35-312cc8b0bd7e\" (UID: \"82c68303-8af4-41a0-af35-312cc8b0bd7e\") " Mar 18 20:11:27 crc kubenswrapper[4950]: I0318 20:11:27.821234 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82c68303-8af4-41a0-af35-312cc8b0bd7e-catalog-content\") pod \"82c68303-8af4-41a0-af35-312cc8b0bd7e\" (UID: \"82c68303-8af4-41a0-af35-312cc8b0bd7e\") " Mar 18 20:11:27 crc kubenswrapper[4950]: I0318 20:11:27.822717 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82c68303-8af4-41a0-af35-312cc8b0bd7e-utilities" (OuterVolumeSpecName: "utilities") pod "82c68303-8af4-41a0-af35-312cc8b0bd7e" (UID: "82c68303-8af4-41a0-af35-312cc8b0bd7e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:11:27 crc kubenswrapper[4950]: I0318 20:11:27.829563 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82c68303-8af4-41a0-af35-312cc8b0bd7e-kube-api-access-swzmx" (OuterVolumeSpecName: "kube-api-access-swzmx") pod "82c68303-8af4-41a0-af35-312cc8b0bd7e" (UID: "82c68303-8af4-41a0-af35-312cc8b0bd7e"). InnerVolumeSpecName "kube-api-access-swzmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:11:27 crc kubenswrapper[4950]: I0318 20:11:27.875075 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82c68303-8af4-41a0-af35-312cc8b0bd7e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "82c68303-8af4-41a0-af35-312cc8b0bd7e" (UID: "82c68303-8af4-41a0-af35-312cc8b0bd7e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:11:27 crc kubenswrapper[4950]: I0318 20:11:27.922870 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82c68303-8af4-41a0-af35-312cc8b0bd7e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:27 crc kubenswrapper[4950]: I0318 20:11:27.922916 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swzmx\" (UniqueName: \"kubernetes.io/projected/82c68303-8af4-41a0-af35-312cc8b0bd7e-kube-api-access-swzmx\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:27 crc kubenswrapper[4950]: I0318 20:11:27.922928 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82c68303-8af4-41a0-af35-312cc8b0bd7e-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:28 crc kubenswrapper[4950]: I0318 20:11:28.349543 4950 generic.go:334] "Generic (PLEG): container finished" podID="4a386d81-3e2c-4df7-988d-5fb1758be616" containerID="b2559e3af7e229f9289ed9781209cece46a38140f4695a205450c808f8116617" exitCode=0 Mar 18 20:11:28 crc kubenswrapper[4950]: I0318 20:11:28.349616 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8kwk7" event={"ID":"4a386d81-3e2c-4df7-988d-5fb1758be616","Type":"ContainerDied","Data":"b2559e3af7e229f9289ed9781209cece46a38140f4695a205450c808f8116617"} Mar 18 20:11:28 crc kubenswrapper[4950]: I0318 20:11:28.360342 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rwzv9" Mar 18 20:11:28 crc kubenswrapper[4950]: I0318 20:11:28.360992 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwzv9" event={"ID":"82c68303-8af4-41a0-af35-312cc8b0bd7e","Type":"ContainerDied","Data":"2adb8218254954fff85e83b554f2aeff06ce47ac7fa13d106beaaf6aa8dc1682"} Mar 18 20:11:28 crc kubenswrapper[4950]: I0318 20:11:28.361043 4950 scope.go:117] "RemoveContainer" containerID="6526d85d276087eb39ac4ab3d4a0cae6048b554628fa67a4a9f6f1f70721cdd0" Mar 18 20:11:28 crc kubenswrapper[4950]: I0318 20:11:28.399844 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5fnpx" podStartSLOduration=4.44764546 podStartE2EDuration="1m12.399822764s" podCreationTimestamp="2026-03-18 20:10:16 +0000 UTC" firstStartedPulling="2026-03-18 20:10:18.867617497 +0000 UTC m=+232.108459365" lastFinishedPulling="2026-03-18 20:11:26.819794791 +0000 UTC m=+300.060636669" observedRunningTime="2026-03-18 20:11:28.396460473 +0000 UTC m=+301.637302341" watchObservedRunningTime="2026-03-18 20:11:28.399822764 +0000 UTC m=+301.640664652" Mar 18 20:11:28 crc kubenswrapper[4950]: I0318 20:11:28.420383 4950 scope.go:117] "RemoveContainer" containerID="38907306b830da3b6e55507900c1a6932b547eca7618ff03aef91df92c2300b3" Mar 18 20:11:28 crc kubenswrapper[4950]: I0318 20:11:28.431589 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rwzv9"] Mar 18 20:11:28 crc kubenswrapper[4950]: I0318 20:11:28.434690 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rwzv9"] Mar 18 20:11:28 crc kubenswrapper[4950]: I0318 20:11:28.445503 4950 scope.go:117] "RemoveContainer" containerID="b95e7bb85d3a41c87de48a23686cc15d691b7ba08e6e8739a1d514ef1506a4a4" Mar 18 20:11:28 crc kubenswrapper[4950]: I0318 20:11:28.620080 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" podUID="a0e597f4-b43a-47db-9f22-a38272060004" containerName="oauth-openshift" containerID="cri-o://9319cf28b6a29f82fc9c9f97d9bd49eb6cea689d474de0971b094d11aab3c153" gracePeriod=15 Mar 18 20:11:28 crc kubenswrapper[4950]: I0318 20:11:28.917135 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lhhrh" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.032333 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.043420 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvtkc\" (UniqueName: \"kubernetes.io/projected/8fe40723-62a6-4fe7-989c-b6f968467b4c-kube-api-access-mvtkc\") pod \"8fe40723-62a6-4fe7-989c-b6f968467b4c\" (UID: \"8fe40723-62a6-4fe7-989c-b6f968467b4c\") " Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.043517 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fe40723-62a6-4fe7-989c-b6f968467b4c-catalog-content\") pod \"8fe40723-62a6-4fe7-989c-b6f968467b4c\" (UID: \"8fe40723-62a6-4fe7-989c-b6f968467b4c\") " Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.043542 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fe40723-62a6-4fe7-989c-b6f968467b4c-utilities\") pod \"8fe40723-62a6-4fe7-989c-b6f968467b4c\" (UID: \"8fe40723-62a6-4fe7-989c-b6f968467b4c\") " Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.044476 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fe40723-62a6-4fe7-989c-b6f968467b4c-utilities" (OuterVolumeSpecName: "utilities") pod "8fe40723-62a6-4fe7-989c-b6f968467b4c" (UID: "8fe40723-62a6-4fe7-989c-b6f968467b4c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.049776 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fe40723-62a6-4fe7-989c-b6f968467b4c-kube-api-access-mvtkc" (OuterVolumeSpecName: "kube-api-access-mvtkc") pod "8fe40723-62a6-4fe7-989c-b6f968467b4c" (UID: "8fe40723-62a6-4fe7-989c-b6f968467b4c"). InnerVolumeSpecName "kube-api-access-mvtkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.078624 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fe40723-62a6-4fe7-989c-b6f968467b4c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8fe40723-62a6-4fe7-989c-b6f968467b4c" (UID: "8fe40723-62a6-4fe7-989c-b6f968467b4c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.144046 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-ocp-branding-template\") pod \"a0e597f4-b43a-47db-9f22-a38272060004\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.144102 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-template-provider-selection\") pod \"a0e597f4-b43a-47db-9f22-a38272060004\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.144137 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-serving-cert\") pod \"a0e597f4-b43a-47db-9f22-a38272060004\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.144162 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-idp-0-file-data\") pod \"a0e597f4-b43a-47db-9f22-a38272060004\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.144181 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-template-error\") pod \"a0e597f4-b43a-47db-9f22-a38272060004\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.144196 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-template-login\") pod \"a0e597f4-b43a-47db-9f22-a38272060004\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.144220 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-cliconfig\") pod \"a0e597f4-b43a-47db-9f22-a38272060004\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.144241 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-session\") pod \"a0e597f4-b43a-47db-9f22-a38272060004\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.144261 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5qm4\" (UniqueName: \"kubernetes.io/projected/a0e597f4-b43a-47db-9f22-a38272060004-kube-api-access-s5qm4\") pod \"a0e597f4-b43a-47db-9f22-a38272060004\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.144276 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-router-certs\") pod \"a0e597f4-b43a-47db-9f22-a38272060004\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.144293 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-audit-policies\") pod \"a0e597f4-b43a-47db-9f22-a38272060004\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.144320 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-trusted-ca-bundle\") pod \"a0e597f4-b43a-47db-9f22-a38272060004\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.144355 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-service-ca\") pod \"a0e597f4-b43a-47db-9f22-a38272060004\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.144373 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a0e597f4-b43a-47db-9f22-a38272060004-audit-dir\") pod \"a0e597f4-b43a-47db-9f22-a38272060004\" (UID: \"a0e597f4-b43a-47db-9f22-a38272060004\") " Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.144618 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvtkc\" (UniqueName: \"kubernetes.io/projected/8fe40723-62a6-4fe7-989c-b6f968467b4c-kube-api-access-mvtkc\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.144631 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fe40723-62a6-4fe7-989c-b6f968467b4c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.144640 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fe40723-62a6-4fe7-989c-b6f968467b4c-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.144681 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a0e597f4-b43a-47db-9f22-a38272060004-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "a0e597f4-b43a-47db-9f22-a38272060004" (UID: "a0e597f4-b43a-47db-9f22-a38272060004"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.145088 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "a0e597f4-b43a-47db-9f22-a38272060004" (UID: "a0e597f4-b43a-47db-9f22-a38272060004"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.145232 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "a0e597f4-b43a-47db-9f22-a38272060004" (UID: "a0e597f4-b43a-47db-9f22-a38272060004"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.145512 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "a0e597f4-b43a-47db-9f22-a38272060004" (UID: "a0e597f4-b43a-47db-9f22-a38272060004"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.145577 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "a0e597f4-b43a-47db-9f22-a38272060004" (UID: "a0e597f4-b43a-47db-9f22-a38272060004"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.149953 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "a0e597f4-b43a-47db-9f22-a38272060004" (UID: "a0e597f4-b43a-47db-9f22-a38272060004"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.150062 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0e597f4-b43a-47db-9f22-a38272060004-kube-api-access-s5qm4" (OuterVolumeSpecName: "kube-api-access-s5qm4") pod "a0e597f4-b43a-47db-9f22-a38272060004" (UID: "a0e597f4-b43a-47db-9f22-a38272060004"). InnerVolumeSpecName "kube-api-access-s5qm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.150290 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "a0e597f4-b43a-47db-9f22-a38272060004" (UID: "a0e597f4-b43a-47db-9f22-a38272060004"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.150569 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "a0e597f4-b43a-47db-9f22-a38272060004" (UID: "a0e597f4-b43a-47db-9f22-a38272060004"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.151207 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "a0e597f4-b43a-47db-9f22-a38272060004" (UID: "a0e597f4-b43a-47db-9f22-a38272060004"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.151581 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "a0e597f4-b43a-47db-9f22-a38272060004" (UID: "a0e597f4-b43a-47db-9f22-a38272060004"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.153582 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "a0e597f4-b43a-47db-9f22-a38272060004" (UID: "a0e597f4-b43a-47db-9f22-a38272060004"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.155058 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "a0e597f4-b43a-47db-9f22-a38272060004" (UID: "a0e597f4-b43a-47db-9f22-a38272060004"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.157907 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "a0e597f4-b43a-47db-9f22-a38272060004" (UID: "a0e597f4-b43a-47db-9f22-a38272060004"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.245546 4950 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.245583 4950 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.245592 4950 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.245601 4950 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.245610 4950 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.245619 4950 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.245628 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5qm4\" (UniqueName: \"kubernetes.io/projected/a0e597f4-b43a-47db-9f22-a38272060004-kube-api-access-s5qm4\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.245638 4950 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.245650 4950 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.245661 4950 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.245671 4950 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a0e597f4-b43a-47db-9f22-a38272060004-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.245681 4950 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.245689 4950 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.245699 4950 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a0e597f4-b43a-47db-9f22-a38272060004-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.366454 4950 generic.go:334] "Generic (PLEG): container finished" podID="a0e597f4-b43a-47db-9f22-a38272060004" containerID="9319cf28b6a29f82fc9c9f97d9bd49eb6cea689d474de0971b094d11aab3c153" exitCode=0 Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.366507 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.366531 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" event={"ID":"a0e597f4-b43a-47db-9f22-a38272060004","Type":"ContainerDied","Data":"9319cf28b6a29f82fc9c9f97d9bd49eb6cea689d474de0971b094d11aab3c153"} Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.366563 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wv7kr" event={"ID":"a0e597f4-b43a-47db-9f22-a38272060004","Type":"ContainerDied","Data":"0cf19fb082ef62fb43e346aa04417a8b6176e8a669c3039920e8a2f2d58d6852"} Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.366582 4950 scope.go:117] "RemoveContainer" containerID="9319cf28b6a29f82fc9c9f97d9bd49eb6cea689d474de0971b094d11aab3c153" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.368628 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8kwk7" event={"ID":"4a386d81-3e2c-4df7-988d-5fb1758be616","Type":"ContainerStarted","Data":"81f4ca959b3b2d24b3c009f7e373b147ee24bca5aaaf63ab6988946a341b5cc6"} Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.370573 4950 generic.go:334] "Generic (PLEG): container finished" podID="8fe40723-62a6-4fe7-989c-b6f968467b4c" containerID="f8dcebf5f3960e992a7a9abf7429c75530bc03f926813c5fd243b4f6f03df668" exitCode=0 Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.370634 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lhhrh" event={"ID":"8fe40723-62a6-4fe7-989c-b6f968467b4c","Type":"ContainerDied","Data":"f8dcebf5f3960e992a7a9abf7429c75530bc03f926813c5fd243b4f6f03df668"} Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.370669 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lhhrh" event={"ID":"8fe40723-62a6-4fe7-989c-b6f968467b4c","Type":"ContainerDied","Data":"0d2d3868583c0aabbdcdc6edea1bf63a62c608cfce81db97a2e4d1849b25beb7"} Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.370833 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lhhrh" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.384991 4950 scope.go:117] "RemoveContainer" containerID="9319cf28b6a29f82fc9c9f97d9bd49eb6cea689d474de0971b094d11aab3c153" Mar 18 20:11:29 crc kubenswrapper[4950]: E0318 20:11:29.385370 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9319cf28b6a29f82fc9c9f97d9bd49eb6cea689d474de0971b094d11aab3c153\": container with ID starting with 9319cf28b6a29f82fc9c9f97d9bd49eb6cea689d474de0971b094d11aab3c153 not found: ID does not exist" containerID="9319cf28b6a29f82fc9c9f97d9bd49eb6cea689d474de0971b094d11aab3c153" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.385429 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9319cf28b6a29f82fc9c9f97d9bd49eb6cea689d474de0971b094d11aab3c153"} err="failed to get container status \"9319cf28b6a29f82fc9c9f97d9bd49eb6cea689d474de0971b094d11aab3c153\": rpc error: code = NotFound desc = could not find container \"9319cf28b6a29f82fc9c9f97d9bd49eb6cea689d474de0971b094d11aab3c153\": container with ID starting with 9319cf28b6a29f82fc9c9f97d9bd49eb6cea689d474de0971b094d11aab3c153 not found: ID does not exist" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.385454 4950 scope.go:117] "RemoveContainer" containerID="f8dcebf5f3960e992a7a9abf7429c75530bc03f926813c5fd243b4f6f03df668" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.394105 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8kwk7" podStartSLOduration=2.992899049 podStartE2EDuration="1m16.394088097s" podCreationTimestamp="2026-03-18 20:10:13 +0000 UTC" firstStartedPulling="2026-03-18 20:10:15.374871149 +0000 UTC m=+228.615713017" lastFinishedPulling="2026-03-18 20:11:28.776060197 +0000 UTC m=+302.016902065" observedRunningTime="2026-03-18 20:11:29.391040134 +0000 UTC m=+302.631882022" watchObservedRunningTime="2026-03-18 20:11:29.394088097 +0000 UTC m=+302.634929965" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.403687 4950 scope.go:117] "RemoveContainer" containerID="4858a2c5928016a18fcb073654870f8c2593355ffb79728629abb9f88912a4c0" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.407477 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lhhrh"] Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.409800 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lhhrh"] Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.421536 4950 scope.go:117] "RemoveContainer" containerID="066cc63479a5ddbb896f736060d5c98157b3f0cc00f596abd47b8bf80f182383" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.421646 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wv7kr"] Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.430782 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wv7kr"] Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.442239 4950 scope.go:117] "RemoveContainer" containerID="f8dcebf5f3960e992a7a9abf7429c75530bc03f926813c5fd243b4f6f03df668" Mar 18 20:11:29 crc kubenswrapper[4950]: E0318 20:11:29.442760 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8dcebf5f3960e992a7a9abf7429c75530bc03f926813c5fd243b4f6f03df668\": container with ID starting with f8dcebf5f3960e992a7a9abf7429c75530bc03f926813c5fd243b4f6f03df668 not found: ID does not exist" containerID="f8dcebf5f3960e992a7a9abf7429c75530bc03f926813c5fd243b4f6f03df668" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.442788 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8dcebf5f3960e992a7a9abf7429c75530bc03f926813c5fd243b4f6f03df668"} err="failed to get container status \"f8dcebf5f3960e992a7a9abf7429c75530bc03f926813c5fd243b4f6f03df668\": rpc error: code = NotFound desc = could not find container \"f8dcebf5f3960e992a7a9abf7429c75530bc03f926813c5fd243b4f6f03df668\": container with ID starting with f8dcebf5f3960e992a7a9abf7429c75530bc03f926813c5fd243b4f6f03df668 not found: ID does not exist" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.442809 4950 scope.go:117] "RemoveContainer" containerID="4858a2c5928016a18fcb073654870f8c2593355ffb79728629abb9f88912a4c0" Mar 18 20:11:29 crc kubenswrapper[4950]: E0318 20:11:29.443103 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4858a2c5928016a18fcb073654870f8c2593355ffb79728629abb9f88912a4c0\": container with ID starting with 4858a2c5928016a18fcb073654870f8c2593355ffb79728629abb9f88912a4c0 not found: ID does not exist" containerID="4858a2c5928016a18fcb073654870f8c2593355ffb79728629abb9f88912a4c0" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.443143 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4858a2c5928016a18fcb073654870f8c2593355ffb79728629abb9f88912a4c0"} err="failed to get container status \"4858a2c5928016a18fcb073654870f8c2593355ffb79728629abb9f88912a4c0\": rpc error: code = NotFound desc = could not find container \"4858a2c5928016a18fcb073654870f8c2593355ffb79728629abb9f88912a4c0\": container with ID starting with 4858a2c5928016a18fcb073654870f8c2593355ffb79728629abb9f88912a4c0 not found: ID does not exist" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.443169 4950 scope.go:117] "RemoveContainer" containerID="066cc63479a5ddbb896f736060d5c98157b3f0cc00f596abd47b8bf80f182383" Mar 18 20:11:29 crc kubenswrapper[4950]: E0318 20:11:29.443497 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"066cc63479a5ddbb896f736060d5c98157b3f0cc00f596abd47b8bf80f182383\": container with ID starting with 066cc63479a5ddbb896f736060d5c98157b3f0cc00f596abd47b8bf80f182383 not found: ID does not exist" containerID="066cc63479a5ddbb896f736060d5c98157b3f0cc00f596abd47b8bf80f182383" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.443518 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"066cc63479a5ddbb896f736060d5c98157b3f0cc00f596abd47b8bf80f182383"} err="failed to get container status \"066cc63479a5ddbb896f736060d5c98157b3f0cc00f596abd47b8bf80f182383\": rpc error: code = NotFound desc = could not find container \"066cc63479a5ddbb896f736060d5c98157b3f0cc00f596abd47b8bf80f182383\": container with ID starting with 066cc63479a5ddbb896f736060d5c98157b3f0cc00f596abd47b8bf80f182383 not found: ID does not exist" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.489769 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82c68303-8af4-41a0-af35-312cc8b0bd7e" path="/var/lib/kubelet/pods/82c68303-8af4-41a0-af35-312cc8b0bd7e/volumes" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.491031 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fe40723-62a6-4fe7-989c-b6f968467b4c" path="/var/lib/kubelet/pods/8fe40723-62a6-4fe7-989c-b6f968467b4c/volumes" Mar 18 20:11:29 crc kubenswrapper[4950]: I0318 20:11:29.492316 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0e597f4-b43a-47db-9f22-a38272060004" path="/var/lib/kubelet/pods/a0e597f4-b43a-47db-9f22-a38272060004/volumes" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.017689 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-664567ccff-7rdnp"] Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.019282 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" podUID="26289d96-c077-4b10-9139-09231241e9d0" containerName="controller-manager" containerID="cri-o://0c33b6169085d2c06985cd0bc6eff75cbf956e5cf0a877495954750461b8ef53" gracePeriod=30 Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.111011 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr"] Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.111207 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" podUID="fa7ec310-f833-410f-9184-0c9b910fe20f" containerName="route-controller-manager" containerID="cri-o://c1e84b4e561754ff798344621025273750eb6771b74ff9fd4ba00dadd87421bf" gracePeriod=30 Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.400737 4950 generic.go:334] "Generic (PLEG): container finished" podID="26289d96-c077-4b10-9139-09231241e9d0" containerID="0c33b6169085d2c06985cd0bc6eff75cbf956e5cf0a877495954750461b8ef53" exitCode=0 Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.400807 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" event={"ID":"26289d96-c077-4b10-9139-09231241e9d0","Type":"ContainerDied","Data":"0c33b6169085d2c06985cd0bc6eff75cbf956e5cf0a877495954750461b8ef53"} Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.403036 4950 generic.go:334] "Generic (PLEG): container finished" podID="fa7ec310-f833-410f-9184-0c9b910fe20f" containerID="c1e84b4e561754ff798344621025273750eb6771b74ff9fd4ba00dadd87421bf" exitCode=0 Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.403077 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" event={"ID":"fa7ec310-f833-410f-9184-0c9b910fe20f","Type":"ContainerDied","Data":"c1e84b4e561754ff798344621025273750eb6771b74ff9fd4ba00dadd87421bf"} Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.526265 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8kwk7" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.526312 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8kwk7" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.595315 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8kwk7" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.629776 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.636814 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.703643 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h42ds\" (UniqueName: \"kubernetes.io/projected/26289d96-c077-4b10-9139-09231241e9d0-kube-api-access-h42ds\") pod \"26289d96-c077-4b10-9139-09231241e9d0\" (UID: \"26289d96-c077-4b10-9139-09231241e9d0\") " Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.704175 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwsq2\" (UniqueName: \"kubernetes.io/projected/fa7ec310-f833-410f-9184-0c9b910fe20f-kube-api-access-cwsq2\") pod \"fa7ec310-f833-410f-9184-0c9b910fe20f\" (UID: \"fa7ec310-f833-410f-9184-0c9b910fe20f\") " Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.704804 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa7ec310-f833-410f-9184-0c9b910fe20f-serving-cert\") pod \"fa7ec310-f833-410f-9184-0c9b910fe20f\" (UID: \"fa7ec310-f833-410f-9184-0c9b910fe20f\") " Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.704940 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa7ec310-f833-410f-9184-0c9b910fe20f-client-ca\") pod \"fa7ec310-f833-410f-9184-0c9b910fe20f\" (UID: \"fa7ec310-f833-410f-9184-0c9b910fe20f\") " Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.705027 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/26289d96-c077-4b10-9139-09231241e9d0-proxy-ca-bundles\") pod \"26289d96-c077-4b10-9139-09231241e9d0\" (UID: \"26289d96-c077-4b10-9139-09231241e9d0\") " Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.705098 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26289d96-c077-4b10-9139-09231241e9d0-config\") pod \"26289d96-c077-4b10-9139-09231241e9d0\" (UID: \"26289d96-c077-4b10-9139-09231241e9d0\") " Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.705180 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26289d96-c077-4b10-9139-09231241e9d0-serving-cert\") pod \"26289d96-c077-4b10-9139-09231241e9d0\" (UID: \"26289d96-c077-4b10-9139-09231241e9d0\") " Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.705272 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa7ec310-f833-410f-9184-0c9b910fe20f-config\") pod \"fa7ec310-f833-410f-9184-0c9b910fe20f\" (UID: \"fa7ec310-f833-410f-9184-0c9b910fe20f\") " Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.705340 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/26289d96-c077-4b10-9139-09231241e9d0-client-ca\") pod \"26289d96-c077-4b10-9139-09231241e9d0\" (UID: \"26289d96-c077-4b10-9139-09231241e9d0\") " Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.706848 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26289d96-c077-4b10-9139-09231241e9d0-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "26289d96-c077-4b10-9139-09231241e9d0" (UID: "26289d96-c077-4b10-9139-09231241e9d0"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.707211 4950 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/26289d96-c077-4b10-9139-09231241e9d0-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.707607 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26289d96-c077-4b10-9139-09231241e9d0-client-ca" (OuterVolumeSpecName: "client-ca") pod "26289d96-c077-4b10-9139-09231241e9d0" (UID: "26289d96-c077-4b10-9139-09231241e9d0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.708274 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa7ec310-f833-410f-9184-0c9b910fe20f-config" (OuterVolumeSpecName: "config") pod "fa7ec310-f833-410f-9184-0c9b910fe20f" (UID: "fa7ec310-f833-410f-9184-0c9b910fe20f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.708905 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26289d96-c077-4b10-9139-09231241e9d0-config" (OuterVolumeSpecName: "config") pod "26289d96-c077-4b10-9139-09231241e9d0" (UID: "26289d96-c077-4b10-9139-09231241e9d0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.709037 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa7ec310-f833-410f-9184-0c9b910fe20f-client-ca" (OuterVolumeSpecName: "client-ca") pod "fa7ec310-f833-410f-9184-0c9b910fe20f" (UID: "fa7ec310-f833-410f-9184-0c9b910fe20f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.713547 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26289d96-c077-4b10-9139-09231241e9d0-kube-api-access-h42ds" (OuterVolumeSpecName: "kube-api-access-h42ds") pod "26289d96-c077-4b10-9139-09231241e9d0" (UID: "26289d96-c077-4b10-9139-09231241e9d0"). InnerVolumeSpecName "kube-api-access-h42ds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.713654 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa7ec310-f833-410f-9184-0c9b910fe20f-kube-api-access-cwsq2" (OuterVolumeSpecName: "kube-api-access-cwsq2") pod "fa7ec310-f833-410f-9184-0c9b910fe20f" (UID: "fa7ec310-f833-410f-9184-0c9b910fe20f"). InnerVolumeSpecName "kube-api-access-cwsq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.714179 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa7ec310-f833-410f-9184-0c9b910fe20f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "fa7ec310-f833-410f-9184-0c9b910fe20f" (UID: "fa7ec310-f833-410f-9184-0c9b910fe20f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.721855 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26289d96-c077-4b10-9139-09231241e9d0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "26289d96-c077-4b10-9139-09231241e9d0" (UID: "26289d96-c077-4b10-9139-09231241e9d0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.808953 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h42ds\" (UniqueName: \"kubernetes.io/projected/26289d96-c077-4b10-9139-09231241e9d0-kube-api-access-h42ds\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.808999 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwsq2\" (UniqueName: \"kubernetes.io/projected/fa7ec310-f833-410f-9184-0c9b910fe20f-kube-api-access-cwsq2\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.809009 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa7ec310-f833-410f-9184-0c9b910fe20f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.809021 4950 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa7ec310-f833-410f-9184-0c9b910fe20f-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.809030 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26289d96-c077-4b10-9139-09231241e9d0-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.809038 4950 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26289d96-c077-4b10-9139-09231241e9d0-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.809047 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa7ec310-f833-410f-9184-0c9b910fe20f-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.809055 4950 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/26289d96-c077-4b10-9139-09231241e9d0-client-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.836039 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.836099 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.836150 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.836750 4950 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217"} pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.836805 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" containerID="cri-o://a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217" gracePeriod=600 Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.980157 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gzj5q" Mar 18 20:11:33 crc kubenswrapper[4950]: I0318 20:11:33.980781 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gzj5q" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.049812 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gzj5q" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.409355 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" event={"ID":"fa7ec310-f833-410f-9184-0c9b910fe20f","Type":"ContainerDied","Data":"d389c536f83e1c5012c761678d56f6b15a647db002901c23716dc059707e465a"} Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.409442 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.409709 4950 scope.go:117] "RemoveContainer" containerID="c1e84b4e561754ff798344621025273750eb6771b74ff9fd4ba00dadd87421bf" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.411546 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" event={"ID":"26289d96-c077-4b10-9139-09231241e9d0","Type":"ContainerDied","Data":"8d42d04b351183ca696848df231c2926996d7e1cafda3b5c825d66fdfb915055"} Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.411561 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-664567ccff-7rdnp" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.420125 4950 generic.go:334] "Generic (PLEG): container finished" podID="4048b439-3266-46e7-9de0-22377efacc46" containerID="a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217" exitCode=0 Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.420763 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerDied","Data":"a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217"} Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.420794 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerStarted","Data":"01edc7246d617f13699dceaa220ac50fb8349729df73606eb00a73d4e3ae8f6a"} Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.438188 4950 scope.go:117] "RemoveContainer" containerID="0c33b6169085d2c06985cd0bc6eff75cbf956e5cf0a877495954750461b8ef53" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.465324 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-664567ccff-7rdnp"] Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.474649 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-664567ccff-7rdnp"] Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.474747 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8kwk7" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.482137 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr"] Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.485402 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cf7cfb578-tpqwr"] Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.486101 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gzj5q" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.686621 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5dbc79446c-qbg4n"] Mar 18 20:11:34 crc kubenswrapper[4950]: E0318 20:11:34.686821 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f54c83ec-47ad-4ced-93c3-d321460f40f3" containerName="extract-utilities" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.686831 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="f54c83ec-47ad-4ced-93c3-d321460f40f3" containerName="extract-utilities" Mar 18 20:11:34 crc kubenswrapper[4950]: E0318 20:11:34.686843 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fe40723-62a6-4fe7-989c-b6f968467b4c" containerName="extract-content" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.686849 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fe40723-62a6-4fe7-989c-b6f968467b4c" containerName="extract-content" Mar 18 20:11:34 crc kubenswrapper[4950]: E0318 20:11:34.686858 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fe40723-62a6-4fe7-989c-b6f968467b4c" containerName="registry-server" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.686865 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fe40723-62a6-4fe7-989c-b6f968467b4c" containerName="registry-server" Mar 18 20:11:34 crc kubenswrapper[4950]: E0318 20:11:34.686875 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa7ec310-f833-410f-9184-0c9b910fe20f" containerName="route-controller-manager" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.686880 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa7ec310-f833-410f-9184-0c9b910fe20f" containerName="route-controller-manager" Mar 18 20:11:34 crc kubenswrapper[4950]: E0318 20:11:34.686889 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fe40723-62a6-4fe7-989c-b6f968467b4c" containerName="extract-utilities" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.686895 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fe40723-62a6-4fe7-989c-b6f968467b4c" containerName="extract-utilities" Mar 18 20:11:34 crc kubenswrapper[4950]: E0318 20:11:34.686904 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c68303-8af4-41a0-af35-312cc8b0bd7e" containerName="extract-utilities" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.686910 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c68303-8af4-41a0-af35-312cc8b0bd7e" containerName="extract-utilities" Mar 18 20:11:34 crc kubenswrapper[4950]: E0318 20:11:34.686918 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f54c83ec-47ad-4ced-93c3-d321460f40f3" containerName="extract-content" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.686923 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="f54c83ec-47ad-4ced-93c3-d321460f40f3" containerName="extract-content" Mar 18 20:11:34 crc kubenswrapper[4950]: E0318 20:11:34.686932 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f54c83ec-47ad-4ced-93c3-d321460f40f3" containerName="registry-server" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.686937 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="f54c83ec-47ad-4ced-93c3-d321460f40f3" containerName="registry-server" Mar 18 20:11:34 crc kubenswrapper[4950]: E0318 20:11:34.686945 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26289d96-c077-4b10-9139-09231241e9d0" containerName="controller-manager" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.686951 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="26289d96-c077-4b10-9139-09231241e9d0" containerName="controller-manager" Mar 18 20:11:34 crc kubenswrapper[4950]: E0318 20:11:34.686961 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0e597f4-b43a-47db-9f22-a38272060004" containerName="oauth-openshift" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.686967 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0e597f4-b43a-47db-9f22-a38272060004" containerName="oauth-openshift" Mar 18 20:11:34 crc kubenswrapper[4950]: E0318 20:11:34.686976 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c68303-8af4-41a0-af35-312cc8b0bd7e" containerName="extract-content" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.686981 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c68303-8af4-41a0-af35-312cc8b0bd7e" containerName="extract-content" Mar 18 20:11:34 crc kubenswrapper[4950]: E0318 20:11:34.686992 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c68303-8af4-41a0-af35-312cc8b0bd7e" containerName="registry-server" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.686997 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c68303-8af4-41a0-af35-312cc8b0bd7e" containerName="registry-server" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.687075 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0e597f4-b43a-47db-9f22-a38272060004" containerName="oauth-openshift" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.687084 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="82c68303-8af4-41a0-af35-312cc8b0bd7e" containerName="registry-server" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.687092 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="f54c83ec-47ad-4ced-93c3-d321460f40f3" containerName="registry-server" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.687103 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fe40723-62a6-4fe7-989c-b6f968467b4c" containerName="registry-server" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.687113 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="26289d96-c077-4b10-9139-09231241e9d0" containerName="controller-manager" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.687120 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa7ec310-f833-410f-9184-0c9b910fe20f" containerName="route-controller-manager" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.687377 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7"] Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.687770 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5dbc79446c-qbg4n" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.687798 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.692161 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.692294 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.692698 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.692715 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.692816 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.697484 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.698072 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.698404 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.698677 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.698939 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.699460 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.699608 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.707764 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.711137 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7"] Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.714510 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5dbc79446c-qbg4n"] Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.819831 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f304e988-7902-4af8-bca7-ce8171bba76d-proxy-ca-bundles\") pod \"controller-manager-6f8b8c5c49-rskv7\" (UID: \"f304e988-7902-4af8-bca7-ce8171bba76d\") " pod="openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.819880 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f304e988-7902-4af8-bca7-ce8171bba76d-serving-cert\") pod \"controller-manager-6f8b8c5c49-rskv7\" (UID: \"f304e988-7902-4af8-bca7-ce8171bba76d\") " pod="openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.819926 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/475f7871-6ae8-436a-b5df-d32b1a2e1404-config\") pod \"route-controller-manager-5dbc79446c-qbg4n\" (UID: \"475f7871-6ae8-436a-b5df-d32b1a2e1404\") " pod="openshift-route-controller-manager/route-controller-manager-5dbc79446c-qbg4n" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.819959 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w4w7\" (UniqueName: \"kubernetes.io/projected/475f7871-6ae8-436a-b5df-d32b1a2e1404-kube-api-access-6w4w7\") pod \"route-controller-manager-5dbc79446c-qbg4n\" (UID: \"475f7871-6ae8-436a-b5df-d32b1a2e1404\") " pod="openshift-route-controller-manager/route-controller-manager-5dbc79446c-qbg4n" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.819991 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zphls\" (UniqueName: \"kubernetes.io/projected/f304e988-7902-4af8-bca7-ce8171bba76d-kube-api-access-zphls\") pod \"controller-manager-6f8b8c5c49-rskv7\" (UID: \"f304e988-7902-4af8-bca7-ce8171bba76d\") " pod="openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.820019 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/475f7871-6ae8-436a-b5df-d32b1a2e1404-serving-cert\") pod \"route-controller-manager-5dbc79446c-qbg4n\" (UID: \"475f7871-6ae8-436a-b5df-d32b1a2e1404\") " pod="openshift-route-controller-manager/route-controller-manager-5dbc79446c-qbg4n" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.820051 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/475f7871-6ae8-436a-b5df-d32b1a2e1404-client-ca\") pod \"route-controller-manager-5dbc79446c-qbg4n\" (UID: \"475f7871-6ae8-436a-b5df-d32b1a2e1404\") " pod="openshift-route-controller-manager/route-controller-manager-5dbc79446c-qbg4n" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.820086 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f304e988-7902-4af8-bca7-ce8171bba76d-config\") pod \"controller-manager-6f8b8c5c49-rskv7\" (UID: \"f304e988-7902-4af8-bca7-ce8171bba76d\") " pod="openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.820113 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f304e988-7902-4af8-bca7-ce8171bba76d-client-ca\") pod \"controller-manager-6f8b8c5c49-rskv7\" (UID: \"f304e988-7902-4af8-bca7-ce8171bba76d\") " pod="openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.921475 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/475f7871-6ae8-436a-b5df-d32b1a2e1404-client-ca\") pod \"route-controller-manager-5dbc79446c-qbg4n\" (UID: \"475f7871-6ae8-436a-b5df-d32b1a2e1404\") " pod="openshift-route-controller-manager/route-controller-manager-5dbc79446c-qbg4n" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.921583 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f304e988-7902-4af8-bca7-ce8171bba76d-config\") pod \"controller-manager-6f8b8c5c49-rskv7\" (UID: \"f304e988-7902-4af8-bca7-ce8171bba76d\") " pod="openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.921632 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f304e988-7902-4af8-bca7-ce8171bba76d-client-ca\") pod \"controller-manager-6f8b8c5c49-rskv7\" (UID: \"f304e988-7902-4af8-bca7-ce8171bba76d\") " pod="openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.921667 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f304e988-7902-4af8-bca7-ce8171bba76d-proxy-ca-bundles\") pod \"controller-manager-6f8b8c5c49-rskv7\" (UID: \"f304e988-7902-4af8-bca7-ce8171bba76d\") " pod="openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.921707 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f304e988-7902-4af8-bca7-ce8171bba76d-serving-cert\") pod \"controller-manager-6f8b8c5c49-rskv7\" (UID: \"f304e988-7902-4af8-bca7-ce8171bba76d\") " pod="openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.921766 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/475f7871-6ae8-436a-b5df-d32b1a2e1404-config\") pod \"route-controller-manager-5dbc79446c-qbg4n\" (UID: \"475f7871-6ae8-436a-b5df-d32b1a2e1404\") " pod="openshift-route-controller-manager/route-controller-manager-5dbc79446c-qbg4n" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.921814 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w4w7\" (UniqueName: \"kubernetes.io/projected/475f7871-6ae8-436a-b5df-d32b1a2e1404-kube-api-access-6w4w7\") pod \"route-controller-manager-5dbc79446c-qbg4n\" (UID: \"475f7871-6ae8-436a-b5df-d32b1a2e1404\") " pod="openshift-route-controller-manager/route-controller-manager-5dbc79446c-qbg4n" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.921864 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zphls\" (UniqueName: \"kubernetes.io/projected/f304e988-7902-4af8-bca7-ce8171bba76d-kube-api-access-zphls\") pod \"controller-manager-6f8b8c5c49-rskv7\" (UID: \"f304e988-7902-4af8-bca7-ce8171bba76d\") " pod="openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.921907 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/475f7871-6ae8-436a-b5df-d32b1a2e1404-serving-cert\") pod \"route-controller-manager-5dbc79446c-qbg4n\" (UID: \"475f7871-6ae8-436a-b5df-d32b1a2e1404\") " pod="openshift-route-controller-manager/route-controller-manager-5dbc79446c-qbg4n" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.923185 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f304e988-7902-4af8-bca7-ce8171bba76d-client-ca\") pod \"controller-manager-6f8b8c5c49-rskv7\" (UID: \"f304e988-7902-4af8-bca7-ce8171bba76d\") " pod="openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.923281 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f304e988-7902-4af8-bca7-ce8171bba76d-config\") pod \"controller-manager-6f8b8c5c49-rskv7\" (UID: \"f304e988-7902-4af8-bca7-ce8171bba76d\") " pod="openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.923192 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/475f7871-6ae8-436a-b5df-d32b1a2e1404-client-ca\") pod \"route-controller-manager-5dbc79446c-qbg4n\" (UID: \"475f7871-6ae8-436a-b5df-d32b1a2e1404\") " pod="openshift-route-controller-manager/route-controller-manager-5dbc79446c-qbg4n" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.924668 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f304e988-7902-4af8-bca7-ce8171bba76d-proxy-ca-bundles\") pod \"controller-manager-6f8b8c5c49-rskv7\" (UID: \"f304e988-7902-4af8-bca7-ce8171bba76d\") " pod="openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.925725 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/475f7871-6ae8-436a-b5df-d32b1a2e1404-config\") pod \"route-controller-manager-5dbc79446c-qbg4n\" (UID: \"475f7871-6ae8-436a-b5df-d32b1a2e1404\") " pod="openshift-route-controller-manager/route-controller-manager-5dbc79446c-qbg4n" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.934372 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/475f7871-6ae8-436a-b5df-d32b1a2e1404-serving-cert\") pod \"route-controller-manager-5dbc79446c-qbg4n\" (UID: \"475f7871-6ae8-436a-b5df-d32b1a2e1404\") " pod="openshift-route-controller-manager/route-controller-manager-5dbc79446c-qbg4n" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.935453 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f304e988-7902-4af8-bca7-ce8171bba76d-serving-cert\") pod \"controller-manager-6f8b8c5c49-rskv7\" (UID: \"f304e988-7902-4af8-bca7-ce8171bba76d\") " pod="openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.949042 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zphls\" (UniqueName: \"kubernetes.io/projected/f304e988-7902-4af8-bca7-ce8171bba76d-kube-api-access-zphls\") pod \"controller-manager-6f8b8c5c49-rskv7\" (UID: \"f304e988-7902-4af8-bca7-ce8171bba76d\") " pod="openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7" Mar 18 20:11:34 crc kubenswrapper[4950]: I0318 20:11:34.953749 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w4w7\" (UniqueName: \"kubernetes.io/projected/475f7871-6ae8-436a-b5df-d32b1a2e1404-kube-api-access-6w4w7\") pod \"route-controller-manager-5dbc79446c-qbg4n\" (UID: \"475f7871-6ae8-436a-b5df-d32b1a2e1404\") " pod="openshift-route-controller-manager/route-controller-manager-5dbc79446c-qbg4n" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.010873 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.021613 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5dbc79446c-qbg4n" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.335193 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5dbc79446c-qbg4n"] Mar 18 20:11:35 crc kubenswrapper[4950]: W0318 20:11:35.345546 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod475f7871_6ae8_436a_b5df_d32b1a2e1404.slice/crio-a72253a04f6669083edc756f9cb9f571ad587891a2e6517e712728df1698d800 WatchSource:0}: Error finding container a72253a04f6669083edc756f9cb9f571ad587891a2e6517e712728df1698d800: Status 404 returned error can't find the container with id a72253a04f6669083edc756f9cb9f571ad587891a2e6517e712728df1698d800 Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.444559 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5dbc79446c-qbg4n" event={"ID":"475f7871-6ae8-436a-b5df-d32b1a2e1404","Type":"ContainerStarted","Data":"a72253a04f6669083edc756f9cb9f571ad587891a2e6517e712728df1698d800"} Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.512508 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26289d96-c077-4b10-9139-09231241e9d0" path="/var/lib/kubelet/pods/26289d96-c077-4b10-9139-09231241e9d0/volumes" Mar 18 20:11:35 crc kubenswrapper[4950]: W0318 20:11:35.517861 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf304e988_7902_4af8_bca7_ce8171bba76d.slice/crio-3d20c331c2b272408873faea41768eba338017bf2236cedf2d3ae4245029e025 WatchSource:0}: Error finding container 3d20c331c2b272408873faea41768eba338017bf2236cedf2d3ae4245029e025: Status 404 returned error can't find the container with id 3d20c331c2b272408873faea41768eba338017bf2236cedf2d3ae4245029e025 Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.518352 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa7ec310-f833-410f-9184-0c9b910fe20f" path="/var/lib/kubelet/pods/fa7ec310-f833-410f-9184-0c9b910fe20f/volumes" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.520746 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7"] Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.682195 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-6b5f774455-xd699"] Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.682870 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: W0318 20:11:35.685336 4950 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-cliconfig": failed to list *v1.ConfigMap: configmaps "v4-0-config-system-cliconfig" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 18 20:11:35 crc kubenswrapper[4950]: E0318 20:11:35.685479 4950 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-cliconfig\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"v4-0-config-system-cliconfig\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 18 20:11:35 crc kubenswrapper[4950]: W0318 20:11:35.686484 4950 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle": failed to list *v1.ConfigMap: configmaps "v4-0-config-system-trusted-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 18 20:11:35 crc kubenswrapper[4950]: E0318 20:11:35.686590 4950 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-trusted-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"v4-0-config-system-trusted-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 18 20:11:35 crc kubenswrapper[4950]: W0318 20:11:35.689035 4950 reflector.go:561] object-"openshift-authentication"/"v4-0-config-user-template-login": failed to list *v1.Secret: secrets "v4-0-config-user-template-login" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 18 20:11:35 crc kubenswrapper[4950]: W0318 20:11:35.692042 4950 reflector.go:561] object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data": failed to list *v1.Secret: secrets "v4-0-config-user-idp-0-file-data" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 18 20:11:35 crc kubenswrapper[4950]: E0318 20:11:35.692081 4950 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-user-idp-0-file-data\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-user-idp-0-file-data\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 18 20:11:35 crc kubenswrapper[4950]: E0318 20:11:35.692045 4950 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-user-template-login\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-user-template-login\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 18 20:11:35 crc kubenswrapper[4950]: W0318 20:11:35.692115 4950 reflector.go:561] object-"openshift-authentication"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 18 20:11:35 crc kubenswrapper[4950]: E0318 20:11:35.692171 4950 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 18 20:11:35 crc kubenswrapper[4950]: W0318 20:11:35.691806 4950 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-serving-cert": failed to list *v1.Secret: secrets "v4-0-config-system-serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 18 20:11:35 crc kubenswrapper[4950]: E0318 20:11:35.692507 4950 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-system-serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 18 20:11:35 crc kubenswrapper[4950]: W0318 20:11:35.691826 4950 reflector.go:561] object-"openshift-authentication"/"audit": failed to list *v1.ConfigMap: configmaps "audit" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 18 20:11:35 crc kubenswrapper[4950]: E0318 20:11:35.692724 4950 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"audit\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"audit\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 18 20:11:35 crc kubenswrapper[4950]: W0318 20:11:35.691861 4950 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-router-certs": failed to list *v1.Secret: secrets "v4-0-config-system-router-certs" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 18 20:11:35 crc kubenswrapper[4950]: W0318 20:11:35.691895 4950 reflector.go:561] object-"openshift-authentication"/"v4-0-config-user-template-error": failed to list *v1.Secret: secrets "v4-0-config-user-template-error" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 18 20:11:35 crc kubenswrapper[4950]: E0318 20:11:35.693007 4950 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-user-template-error\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-user-template-error\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 18 20:11:35 crc kubenswrapper[4950]: W0318 20:11:35.691920 4950 reflector.go:561] object-"openshift-authentication"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 18 20:11:35 crc kubenswrapper[4950]: E0318 20:11:35.693035 4950 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 18 20:11:35 crc kubenswrapper[4950]: W0318 20:11:35.691957 4950 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-service-ca": failed to list *v1.ConfigMap: configmaps "v4-0-config-system-service-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 18 20:11:35 crc kubenswrapper[4950]: E0318 20:11:35.693059 4950 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-service-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"v4-0-config-system-service-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 18 20:11:35 crc kubenswrapper[4950]: W0318 20:11:35.691970 4950 reflector.go:561] object-"openshift-authentication"/"v4-0-config-user-template-provider-selection": failed to list *v1.Secret: secrets "v4-0-config-user-template-provider-selection" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 18 20:11:35 crc kubenswrapper[4950]: E0318 20:11:35.693077 4950 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-user-template-provider-selection\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-user-template-provider-selection\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 18 20:11:35 crc kubenswrapper[4950]: W0318 20:11:35.692000 4950 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-session": failed to list *v1.Secret: secrets "v4-0-config-system-session" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 18 20:11:35 crc kubenswrapper[4950]: E0318 20:11:35.693096 4950 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-session\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-system-session\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 18 20:11:35 crc kubenswrapper[4950]: W0318 20:11:35.692001 4950 reflector.go:561] object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc": failed to list *v1.Secret: secrets "oauth-openshift-dockercfg-znhcc" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 18 20:11:35 crc kubenswrapper[4950]: E0318 20:11:35.693113 4950 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"oauth-openshift-dockercfg-znhcc\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"oauth-openshift-dockercfg-znhcc\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 18 20:11:35 crc kubenswrapper[4950]: E0318 20:11:35.693369 4950 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-router-certs\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-system-router-certs\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.696206 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.700343 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6b5f774455-xd699"] Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.836996 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.837512 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-system-service-ca\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.837633 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-system-router-certs\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.837731 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-user-template-login\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.837822 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.837922 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-system-session\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.838019 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxtts\" (UniqueName: \"kubernetes.io/projected/63b375d7-9a06-47aa-8c26-754e61bf2b91-kube-api-access-lxtts\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.838119 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.839015 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/63b375d7-9a06-47aa-8c26-754e61bf2b91-audit-dir\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.839155 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.839246 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.839336 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-user-template-error\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.839444 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/63b375d7-9a06-47aa-8c26-754e61bf2b91-audit-policies\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.839532 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.940715 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.941035 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.941239 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-user-template-error\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.941433 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/63b375d7-9a06-47aa-8c26-754e61bf2b91-audit-policies\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.941579 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.941736 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.941878 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-system-service-ca\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.942030 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-system-router-certs\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.942165 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-user-template-login\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.942305 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.942454 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-system-session\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.942601 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxtts\" (UniqueName: \"kubernetes.io/projected/63b375d7-9a06-47aa-8c26-754e61bf2b91-kube-api-access-lxtts\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.942953 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.943812 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/63b375d7-9a06-47aa-8c26-754e61bf2b91-audit-dir\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.943920 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/63b375d7-9a06-47aa-8c26-754e61bf2b91-audit-dir\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:35 crc kubenswrapper[4950]: I0318 20:11:35.952349 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.450849 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7" event={"ID":"f304e988-7902-4af8-bca7-ce8171bba76d","Type":"ContainerStarted","Data":"a7537e5055e63cc848593d5f8ac7ed23abb28101a2c8021cd0cbaadb2bdebb87"} Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.451529 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7" event={"ID":"f304e988-7902-4af8-bca7-ce8171bba76d","Type":"ContainerStarted","Data":"3d20c331c2b272408873faea41768eba338017bf2236cedf2d3ae4245029e025"} Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.452522 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.453940 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5dbc79446c-qbg4n" event={"ID":"475f7871-6ae8-436a-b5df-d32b1a2e1404","Type":"ContainerStarted","Data":"dcbcc237ff8806b236e3631fab21e7e116453ce5849cad3def323026d9a9985d"} Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.454270 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5dbc79446c-qbg4n" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.459756 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.460048 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5dbc79446c-qbg4n" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.471378 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6f8b8c5c49-rskv7" podStartSLOduration=3.471362881 podStartE2EDuration="3.471362881s" podCreationTimestamp="2026-03-18 20:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:11:36.470805956 +0000 UTC m=+309.711647864" watchObservedRunningTime="2026-03-18 20:11:36.471362881 +0000 UTC m=+309.712204759" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.502202 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5dbc79446c-qbg4n" podStartSLOduration=3.502179553 podStartE2EDuration="3.502179553s" podCreationTimestamp="2026-03-18 20:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:11:36.501990058 +0000 UTC m=+309.742831936" watchObservedRunningTime="2026-03-18 20:11:36.502179553 +0000 UTC m=+309.743021421" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.510004 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.536884 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.543434 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-system-session\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.546930 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.553638 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.563078 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-system-router-certs\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.564841 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.630244 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.632984 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-system-service-ca\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.734811 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.742181 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/63b375d7-9a06-47aa-8c26-754e61bf2b91-audit-policies\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.827307 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.843288 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.867133 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.889715 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.894808 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.933675 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 18 20:11:36 crc kubenswrapper[4950]: E0318 20:11:36.941580 4950 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-error: failed to sync secret cache: timed out waiting for the condition Mar 18 20:11:36 crc kubenswrapper[4950]: E0318 20:11:36.941665 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-user-template-error podName:63b375d7-9a06-47aa-8c26-754e61bf2b91 nodeName:}" failed. No retries permitted until 2026-03-18 20:11:37.441644784 +0000 UTC m=+310.682486642 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-error" (UniqueName: "kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-user-template-error") pod "oauth-openshift-6b5f774455-xd699" (UID: "63b375d7-9a06-47aa-8c26-754e61bf2b91") : failed to sync secret cache: timed out waiting for the condition Mar 18 20:11:36 crc kubenswrapper[4950]: E0318 20:11:36.941592 4950 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-provider-selection: failed to sync secret cache: timed out waiting for the condition Mar 18 20:11:36 crc kubenswrapper[4950]: E0318 20:11:36.941740 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-user-template-provider-selection podName:63b375d7-9a06-47aa-8c26-754e61bf2b91 nodeName:}" failed. No retries permitted until 2026-03-18 20:11:37.441721426 +0000 UTC m=+310.682563284 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-provider-selection" (UniqueName: "kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-user-template-provider-selection") pod "oauth-openshift-6b5f774455-xd699" (UID: "63b375d7-9a06-47aa-8c26-754e61bf2b91") : failed to sync secret cache: timed out waiting for the condition Mar 18 20:11:36 crc kubenswrapper[4950]: E0318 20:11:36.942729 4950 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-login: failed to sync secret cache: timed out waiting for the condition Mar 18 20:11:36 crc kubenswrapper[4950]: E0318 20:11:36.942769 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-user-template-login podName:63b375d7-9a06-47aa-8c26-754e61bf2b91 nodeName:}" failed. No retries permitted until 2026-03-18 20:11:37.442761184 +0000 UTC m=+310.683603062 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-login" (UniqueName: "kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-user-template-login") pod "oauth-openshift-6b5f774455-xd699" (UID: "63b375d7-9a06-47aa-8c26-754e61bf2b91") : failed to sync secret cache: timed out waiting for the condition Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.943882 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:36 crc kubenswrapper[4950]: I0318 20:11:36.999272 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 18 20:11:37 crc kubenswrapper[4950]: I0318 20:11:37.008948 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxtts\" (UniqueName: \"kubernetes.io/projected/63b375d7-9a06-47aa-8c26-754e61bf2b91-kube-api-access-lxtts\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:37 crc kubenswrapper[4950]: I0318 20:11:37.079934 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 18 20:11:37 crc kubenswrapper[4950]: I0318 20:11:37.086657 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5fnpx" Mar 18 20:11:37 crc kubenswrapper[4950]: I0318 20:11:37.087519 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5fnpx" Mar 18 20:11:37 crc kubenswrapper[4950]: I0318 20:11:37.142082 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5fnpx" Mar 18 20:11:37 crc kubenswrapper[4950]: I0318 20:11:37.150446 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 18 20:11:37 crc kubenswrapper[4950]: I0318 20:11:37.173953 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 18 20:11:37 crc kubenswrapper[4950]: I0318 20:11:37.460075 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-user-template-login\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:37 crc kubenswrapper[4950]: I0318 20:11:37.460157 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:37 crc kubenswrapper[4950]: I0318 20:11:37.460182 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-user-template-error\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:37 crc kubenswrapper[4950]: I0318 20:11:37.463791 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-user-template-error\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:37 crc kubenswrapper[4950]: I0318 20:11:37.464455 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-user-template-login\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:37 crc kubenswrapper[4950]: I0318 20:11:37.466627 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/63b375d7-9a06-47aa-8c26-754e61bf2b91-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6b5f774455-xd699\" (UID: \"63b375d7-9a06-47aa-8c26-754e61bf2b91\") " pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:37 crc kubenswrapper[4950]: I0318 20:11:37.523403 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5fnpx" Mar 18 20:11:37 crc kubenswrapper[4950]: I0318 20:11:37.542750 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:37 crc kubenswrapper[4950]: I0318 20:11:37.771056 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6b5f774455-xd699"] Mar 18 20:11:37 crc kubenswrapper[4950]: W0318 20:11:37.781522 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63b375d7_9a06_47aa_8c26_754e61bf2b91.slice/crio-7840b42774dd63a5641d5e004f01a228f90c6f088b7400ec0fb707f8f63846a0 WatchSource:0}: Error finding container 7840b42774dd63a5641d5e004f01a228f90c6f088b7400ec0fb707f8f63846a0: Status 404 returned error can't find the container with id 7840b42774dd63a5641d5e004f01a228f90c6f088b7400ec0fb707f8f63846a0 Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.237729 4950 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.238553 4950 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.238690 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.238865 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc" gracePeriod=15 Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.238942 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3" gracePeriod=15 Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.238939 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a" gracePeriod=15 Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.238972 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db" gracePeriod=15 Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.238994 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4" gracePeriod=15 Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.239366 4950 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 18 20:11:38 crc kubenswrapper[4950]: E0318 20:11:38.239526 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.239541 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 18 20:11:38 crc kubenswrapper[4950]: E0318 20:11:38.239549 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.239557 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 18 20:11:38 crc kubenswrapper[4950]: E0318 20:11:38.239568 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.239575 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 20:11:38 crc kubenswrapper[4950]: E0318 20:11:38.239588 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.239595 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 18 20:11:38 crc kubenswrapper[4950]: E0318 20:11:38.239605 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.239613 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 20:11:38 crc kubenswrapper[4950]: E0318 20:11:38.239620 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.239625 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 18 20:11:38 crc kubenswrapper[4950]: E0318 20:11:38.239634 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.239640 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 20:11:38 crc kubenswrapper[4950]: E0318 20:11:38.239649 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.239655 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 20:11:38 crc kubenswrapper[4950]: E0318 20:11:38.239666 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.239673 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.239800 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.239813 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.239821 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.239828 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.239836 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.239845 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.239855 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 18 20:11:38 crc kubenswrapper[4950]: E0318 20:11:38.239985 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.239996 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.240128 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.240359 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.373792 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.373834 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.373888 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.373997 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.374099 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.374122 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.374147 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.374199 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.467659 4950 generic.go:334] "Generic (PLEG): container finished" podID="e7e1cabf-6954-4269-a638-16fb70a90850" containerID="1b098843c103958ad8366dd7296e3874a197d70b5b5e7762b091b24705d5ca80" exitCode=0 Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.467735 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e7e1cabf-6954-4269-a638-16fb70a90850","Type":"ContainerDied","Data":"1b098843c103958ad8366dd7296e3874a197d70b5b5e7762b091b24705d5ca80"} Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.468579 4950 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.469191 4950 status_manager.go:851] "Failed to get status for pod" podUID="e7e1cabf-6954-4269-a638-16fb70a90850" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.469717 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.471120 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.471800 4950 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a" exitCode=0 Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.471826 4950 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3" exitCode=0 Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.471837 4950 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db" exitCode=0 Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.471849 4950 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4" exitCode=2 Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.471855 4950 scope.go:117] "RemoveContainer" containerID="7bd5f0309a5038205f9838ad9cccad9348c05324de5bb55f22d90ca6dfa5a3eb" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.473261 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-6b5f774455-xd699_63b375d7-9a06-47aa-8c26-754e61bf2b91/oauth-openshift/0.log" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.473303 4950 generic.go:334] "Generic (PLEG): container finished" podID="63b375d7-9a06-47aa-8c26-754e61bf2b91" containerID="6791cc239bf72607bb5e4e5a1f5901c62573d0f5e50d0078da323b42753244be" exitCode=255 Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.473383 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" event={"ID":"63b375d7-9a06-47aa-8c26-754e61bf2b91","Type":"ContainerDied","Data":"6791cc239bf72607bb5e4e5a1f5901c62573d0f5e50d0078da323b42753244be"} Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.473441 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" event={"ID":"63b375d7-9a06-47aa-8c26-754e61bf2b91","Type":"ContainerStarted","Data":"7840b42774dd63a5641d5e004f01a228f90c6f088b7400ec0fb707f8f63846a0"} Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.473853 4950 scope.go:117] "RemoveContainer" containerID="6791cc239bf72607bb5e4e5a1f5901c62573d0f5e50d0078da323b42753244be" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.474123 4950 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.474470 4950 status_manager.go:851] "Failed to get status for pod" podUID="63b375d7-9a06-47aa-8c26-754e61bf2b91" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-6b5f774455-xd699\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.474738 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.474768 4950 status_manager.go:851] "Failed to get status for pod" podUID="e7e1cabf-6954-4269-a638-16fb70a90850" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.474792 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.474824 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.474827 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.474845 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.474874 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.474879 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.474908 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.474935 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.474978 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.474998 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.475025 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.475044 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.475048 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.475010 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.475083 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 20:11:38 crc kubenswrapper[4950]: E0318 20:11:38.478109 4950 event.go:368] "Unable to write event (may retry after sleeping)" err="Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/events/oauth-openshift-6b5f774455-xd699.189e088b5717aa7c\": dial tcp 38.102.83.163:6443: connect: connection refused" event="&Event{ObjectMeta:{oauth-openshift-6b5f774455-xd699.189e088b5717aa7c openshift-authentication 30025 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-authentication,Name:oauth-openshift-6b5f774455-xd699,UID:63b375d7-9a06-47aa-8c26-754e61bf2b91,APIVersion:v1,ResourceVersion:29991,FieldPath:spec.containers{oauth-openshift},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:11:37 +0000 UTC,LastTimestamp:2026-03-18 20:11:38.47752314 +0000 UTC m=+311.718365028,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.690448 4950 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.690730 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.754900 4950 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Mar 18 20:11:38 crc kubenswrapper[4950]: I0318 20:11:38.754953 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" Mar 18 20:11:39 crc kubenswrapper[4950]: I0318 20:11:39.483543 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-6b5f774455-xd699_63b375d7-9a06-47aa-8c26-754e61bf2b91/oauth-openshift/1.log" Mar 18 20:11:39 crc kubenswrapper[4950]: I0318 20:11:39.484377 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-6b5f774455-xd699_63b375d7-9a06-47aa-8c26-754e61bf2b91/oauth-openshift/0.log" Mar 18 20:11:39 crc kubenswrapper[4950]: I0318 20:11:39.484491 4950 generic.go:334] "Generic (PLEG): container finished" podID="63b375d7-9a06-47aa-8c26-754e61bf2b91" containerID="b61733f6eb8ab8c2ed359bc0e5745dccc9915c0cf005ef9fcb37187c46018ee0" exitCode=255 Mar 18 20:11:39 crc kubenswrapper[4950]: I0318 20:11:39.485298 4950 scope.go:117] "RemoveContainer" containerID="b61733f6eb8ab8c2ed359bc0e5745dccc9915c0cf005ef9fcb37187c46018ee0" Mar 18 20:11:39 crc kubenswrapper[4950]: I0318 20:11:39.485449 4950 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:39 crc kubenswrapper[4950]: E0318 20:11:39.485786 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-6b5f774455-xd699_openshift-authentication(63b375d7-9a06-47aa-8c26-754e61bf2b91)\"" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" podUID="63b375d7-9a06-47aa-8c26-754e61bf2b91" Mar 18 20:11:39 crc kubenswrapper[4950]: I0318 20:11:39.486050 4950 status_manager.go:851] "Failed to get status for pod" podUID="63b375d7-9a06-47aa-8c26-754e61bf2b91" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-6b5f774455-xd699\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:39 crc kubenswrapper[4950]: I0318 20:11:39.486649 4950 status_manager.go:851] "Failed to get status for pod" podUID="e7e1cabf-6954-4269-a638-16fb70a90850" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:39 crc kubenswrapper[4950]: I0318 20:11:39.490312 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 18 20:11:39 crc kubenswrapper[4950]: I0318 20:11:39.493895 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" event={"ID":"63b375d7-9a06-47aa-8c26-754e61bf2b91","Type":"ContainerDied","Data":"b61733f6eb8ab8c2ed359bc0e5745dccc9915c0cf005ef9fcb37187c46018ee0"} Mar 18 20:11:39 crc kubenswrapper[4950]: I0318 20:11:39.493941 4950 scope.go:117] "RemoveContainer" containerID="6791cc239bf72607bb5e4e5a1f5901c62573d0f5e50d0078da323b42753244be" Mar 18 20:11:39 crc kubenswrapper[4950]: E0318 20:11:39.613983 4950 event.go:368] "Unable to write event (may retry after sleeping)" err="Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/events/oauth-openshift-6b5f774455-xd699.189e088b5717aa7c\": dial tcp 38.102.83.163:6443: connect: connection refused" event="&Event{ObjectMeta:{oauth-openshift-6b5f774455-xd699.189e088b5717aa7c openshift-authentication 30025 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-authentication,Name:oauth-openshift-6b5f774455-xd699,UID:63b375d7-9a06-47aa-8c26-754e61bf2b91,APIVersion:v1,ResourceVersion:29991,FieldPath:spec.containers{oauth-openshift},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:11:37 +0000 UTC,LastTimestamp:2026-03-18 20:11:38.47752314 +0000 UTC m=+311.718365028,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:11:39 crc kubenswrapper[4950]: I0318 20:11:39.910188 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 18 20:11:39 crc kubenswrapper[4950]: I0318 20:11:39.910711 4950 status_manager.go:851] "Failed to get status for pod" podUID="63b375d7-9a06-47aa-8c26-754e61bf2b91" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-6b5f774455-xd699\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:39 crc kubenswrapper[4950]: I0318 20:11:39.911088 4950 status_manager.go:851] "Failed to get status for pod" podUID="e7e1cabf-6954-4269-a638-16fb70a90850" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:40 crc kubenswrapper[4950]: I0318 20:11:40.095318 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e7e1cabf-6954-4269-a638-16fb70a90850-kubelet-dir\") pod \"e7e1cabf-6954-4269-a638-16fb70a90850\" (UID: \"e7e1cabf-6954-4269-a638-16fb70a90850\") " Mar 18 20:11:40 crc kubenswrapper[4950]: I0318 20:11:40.095395 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e7e1cabf-6954-4269-a638-16fb70a90850-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e7e1cabf-6954-4269-a638-16fb70a90850" (UID: "e7e1cabf-6954-4269-a638-16fb70a90850"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:11:40 crc kubenswrapper[4950]: I0318 20:11:40.095761 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e1cabf-6954-4269-a638-16fb70a90850-kube-api-access\") pod \"e7e1cabf-6954-4269-a638-16fb70a90850\" (UID: \"e7e1cabf-6954-4269-a638-16fb70a90850\") " Mar 18 20:11:40 crc kubenswrapper[4950]: I0318 20:11:40.095841 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e7e1cabf-6954-4269-a638-16fb70a90850-var-lock\") pod \"e7e1cabf-6954-4269-a638-16fb70a90850\" (UID: \"e7e1cabf-6954-4269-a638-16fb70a90850\") " Mar 18 20:11:40 crc kubenswrapper[4950]: I0318 20:11:40.095938 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e7e1cabf-6954-4269-a638-16fb70a90850-var-lock" (OuterVolumeSpecName: "var-lock") pod "e7e1cabf-6954-4269-a638-16fb70a90850" (UID: "e7e1cabf-6954-4269-a638-16fb70a90850"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:11:40 crc kubenswrapper[4950]: I0318 20:11:40.096335 4950 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e7e1cabf-6954-4269-a638-16fb70a90850-var-lock\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:40 crc kubenswrapper[4950]: I0318 20:11:40.096369 4950 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e7e1cabf-6954-4269-a638-16fb70a90850-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:40 crc kubenswrapper[4950]: I0318 20:11:40.110161 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e1cabf-6954-4269-a638-16fb70a90850-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e1cabf-6954-4269-a638-16fb70a90850" (UID: "e7e1cabf-6954-4269-a638-16fb70a90850"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:11:40 crc kubenswrapper[4950]: I0318 20:11:40.197445 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e1cabf-6954-4269-a638-16fb70a90850-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:40 crc kubenswrapper[4950]: I0318 20:11:40.508138 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-6b5f774455-xd699_63b375d7-9a06-47aa-8c26-754e61bf2b91/oauth-openshift/1.log" Mar 18 20:11:40 crc kubenswrapper[4950]: I0318 20:11:40.508990 4950 scope.go:117] "RemoveContainer" containerID="b61733f6eb8ab8c2ed359bc0e5745dccc9915c0cf005ef9fcb37187c46018ee0" Mar 18 20:11:40 crc kubenswrapper[4950]: E0318 20:11:40.509162 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-6b5f774455-xd699_openshift-authentication(63b375d7-9a06-47aa-8c26-754e61bf2b91)\"" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" podUID="63b375d7-9a06-47aa-8c26-754e61bf2b91" Mar 18 20:11:40 crc kubenswrapper[4950]: I0318 20:11:40.509286 4950 status_manager.go:851] "Failed to get status for pod" podUID="63b375d7-9a06-47aa-8c26-754e61bf2b91" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-6b5f774455-xd699\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:40 crc kubenswrapper[4950]: I0318 20:11:40.509586 4950 status_manager.go:851] "Failed to get status for pod" podUID="e7e1cabf-6954-4269-a638-16fb70a90850" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:40 crc kubenswrapper[4950]: I0318 20:11:40.511265 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e7e1cabf-6954-4269-a638-16fb70a90850","Type":"ContainerDied","Data":"2cf6261e65c92df022363e1bc50e9bf0c8424f4c5df276f7ca9365c4b90d2f79"} Mar 18 20:11:40 crc kubenswrapper[4950]: I0318 20:11:40.511299 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2cf6261e65c92df022363e1bc50e9bf0c8424f4c5df276f7ca9365c4b90d2f79" Mar 18 20:11:40 crc kubenswrapper[4950]: I0318 20:11:40.511370 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 18 20:11:40 crc kubenswrapper[4950]: I0318 20:11:40.523295 4950 status_manager.go:851] "Failed to get status for pod" podUID="63b375d7-9a06-47aa-8c26-754e61bf2b91" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-6b5f774455-xd699\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:40 crc kubenswrapper[4950]: I0318 20:11:40.523739 4950 status_manager.go:851] "Failed to get status for pod" podUID="e7e1cabf-6954-4269-a638-16fb70a90850" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.118761 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.121845 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.122772 4950 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.123219 4950 status_manager.go:851] "Failed to get status for pod" podUID="63b375d7-9a06-47aa-8c26-754e61bf2b91" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-6b5f774455-xd699\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.123923 4950 status_manager.go:851] "Failed to get status for pod" podUID="e7e1cabf-6954-4269-a638-16fb70a90850" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.223594 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.223764 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.223939 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.223694 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.224294 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.226540 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.324915 4950 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.324985 4950 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.325003 4950 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.486748 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.518821 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.519423 4950 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc" exitCode=0 Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.519474 4950 scope.go:117] "RemoveContainer" containerID="926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.519520 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.519889 4950 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.520028 4950 status_manager.go:851] "Failed to get status for pod" podUID="63b375d7-9a06-47aa-8c26-754e61bf2b91" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-6b5f774455-xd699\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.520330 4950 status_manager.go:851] "Failed to get status for pod" podUID="e7e1cabf-6954-4269-a638-16fb70a90850" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.523943 4950 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.524335 4950 status_manager.go:851] "Failed to get status for pod" podUID="63b375d7-9a06-47aa-8c26-754e61bf2b91" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-6b5f774455-xd699\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.524860 4950 status_manager.go:851] "Failed to get status for pod" podUID="e7e1cabf-6954-4269-a638-16fb70a90850" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.535620 4950 scope.go:117] "RemoveContainer" containerID="301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.551075 4950 scope.go:117] "RemoveContainer" containerID="ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.566147 4950 scope.go:117] "RemoveContainer" containerID="c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.581600 4950 scope.go:117] "RemoveContainer" containerID="01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.610239 4950 scope.go:117] "RemoveContainer" containerID="576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.630453 4950 scope.go:117] "RemoveContainer" containerID="926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a" Mar 18 20:11:41 crc kubenswrapper[4950]: E0318 20:11:41.632266 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a\": container with ID starting with 926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a not found: ID does not exist" containerID="926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.632316 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a"} err="failed to get container status \"926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a\": rpc error: code = NotFound desc = could not find container \"926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a\": container with ID starting with 926b76b2b7fea88a07a7d5f6f99aa07ee969f8f385c1325de4a05af1ec8c4d9a not found: ID does not exist" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.632346 4950 scope.go:117] "RemoveContainer" containerID="301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3" Mar 18 20:11:41 crc kubenswrapper[4950]: E0318 20:11:41.632834 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3\": container with ID starting with 301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3 not found: ID does not exist" containerID="301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.632862 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3"} err="failed to get container status \"301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3\": rpc error: code = NotFound desc = could not find container \"301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3\": container with ID starting with 301af1b034f254843be360b752bf3064ecd42b533f663a8885bfa012e45c42a3 not found: ID does not exist" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.632898 4950 scope.go:117] "RemoveContainer" containerID="ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db" Mar 18 20:11:41 crc kubenswrapper[4950]: E0318 20:11:41.633246 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db\": container with ID starting with ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db not found: ID does not exist" containerID="ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.633276 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db"} err="failed to get container status \"ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db\": rpc error: code = NotFound desc = could not find container \"ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db\": container with ID starting with ff9663c6ee2f6517a2e7e4b403793eab9fc6424bbedb0cf65ce4b99e6d7a88db not found: ID does not exist" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.633297 4950 scope.go:117] "RemoveContainer" containerID="c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4" Mar 18 20:11:41 crc kubenswrapper[4950]: E0318 20:11:41.634665 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4\": container with ID starting with c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4 not found: ID does not exist" containerID="c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.634695 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4"} err="failed to get container status \"c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4\": rpc error: code = NotFound desc = could not find container \"c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4\": container with ID starting with c913f51d4080f9ddb46a886638806067b83d07cdc848ff8747d3a4ad5bb2afe4 not found: ID does not exist" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.634709 4950 scope.go:117] "RemoveContainer" containerID="01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc" Mar 18 20:11:41 crc kubenswrapper[4950]: E0318 20:11:41.635019 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc\": container with ID starting with 01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc not found: ID does not exist" containerID="01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.635051 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc"} err="failed to get container status \"01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc\": rpc error: code = NotFound desc = could not find container \"01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc\": container with ID starting with 01c75950388c1aa2e70dbcbe9222ee7331d119b6efefe820c61136dc4f81cacc not found: ID does not exist" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.635066 4950 scope.go:117] "RemoveContainer" containerID="576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb" Mar 18 20:11:41 crc kubenswrapper[4950]: E0318 20:11:41.635506 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\": container with ID starting with 576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb not found: ID does not exist" containerID="576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb" Mar 18 20:11:41 crc kubenswrapper[4950]: I0318 20:11:41.635542 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb"} err="failed to get container status \"576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\": rpc error: code = NotFound desc = could not find container \"576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb\": container with ID starting with 576059bf28fad6ad51f784495f5e07b5501691c95b9a16478e4032b2e9754ecb not found: ID does not exist" Mar 18 20:11:43 crc kubenswrapper[4950]: E0318 20:11:43.269738 4950 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.163:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 20:11:43 crc kubenswrapper[4950]: I0318 20:11:43.270360 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 20:11:43 crc kubenswrapper[4950]: I0318 20:11:43.531829 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b447b83b650669d3caeee0b37fd0c40737beaafdfb9e697aa82ad08c9e63b781"} Mar 18 20:11:43 crc kubenswrapper[4950]: E0318 20:11:43.562004 4950 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:43 crc kubenswrapper[4950]: E0318 20:11:43.562263 4950 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:43 crc kubenswrapper[4950]: E0318 20:11:43.562509 4950 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:43 crc kubenswrapper[4950]: E0318 20:11:43.562738 4950 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:43 crc kubenswrapper[4950]: E0318 20:11:43.562995 4950 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:43 crc kubenswrapper[4950]: I0318 20:11:43.563021 4950 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 18 20:11:43 crc kubenswrapper[4950]: E0318 20:11:43.563248 4950 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.163:6443: connect: connection refused" interval="200ms" Mar 18 20:11:43 crc kubenswrapper[4950]: E0318 20:11:43.764653 4950 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.163:6443: connect: connection refused" interval="400ms" Mar 18 20:11:44 crc kubenswrapper[4950]: E0318 20:11:44.165557 4950 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.163:6443: connect: connection refused" interval="800ms" Mar 18 20:11:44 crc kubenswrapper[4950]: I0318 20:11:44.539017 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"86227fc700ceea5952784deaf8f32fa67d24f9a70021929a4850d6c3f0bc5806"} Mar 18 20:11:44 crc kubenswrapper[4950]: I0318 20:11:44.539956 4950 status_manager.go:851] "Failed to get status for pod" podUID="63b375d7-9a06-47aa-8c26-754e61bf2b91" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-6b5f774455-xd699\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:44 crc kubenswrapper[4950]: I0318 20:11:44.540264 4950 status_manager.go:851] "Failed to get status for pod" podUID="e7e1cabf-6954-4269-a638-16fb70a90850" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:44 crc kubenswrapper[4950]: E0318 20:11:44.541179 4950 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.163:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 20:11:44 crc kubenswrapper[4950]: E0318 20:11:44.966323 4950 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.163:6443: connect: connection refused" interval="1.6s" Mar 18 20:11:45 crc kubenswrapper[4950]: E0318 20:11:45.556061 4950 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.163:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 20:11:46 crc kubenswrapper[4950]: E0318 20:11:46.568377 4950 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.163:6443: connect: connection refused" interval="3.2s" Mar 18 20:11:47 crc kubenswrapper[4950]: I0318 20:11:47.503073 4950 status_manager.go:851] "Failed to get status for pod" podUID="63b375d7-9a06-47aa-8c26-754e61bf2b91" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-6b5f774455-xd699\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:47 crc kubenswrapper[4950]: I0318 20:11:47.507531 4950 status_manager.go:851] "Failed to get status for pod" podUID="e7e1cabf-6954-4269-a638-16fb70a90850" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:47 crc kubenswrapper[4950]: I0318 20:11:47.543854 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:47 crc kubenswrapper[4950]: I0318 20:11:47.543942 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:11:47 crc kubenswrapper[4950]: I0318 20:11:47.544834 4950 scope.go:117] "RemoveContainer" containerID="b61733f6eb8ab8c2ed359bc0e5745dccc9915c0cf005ef9fcb37187c46018ee0" Mar 18 20:11:47 crc kubenswrapper[4950]: E0318 20:11:47.545227 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-6b5f774455-xd699_openshift-authentication(63b375d7-9a06-47aa-8c26-754e61bf2b91)\"" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" podUID="63b375d7-9a06-47aa-8c26-754e61bf2b91" Mar 18 20:11:49 crc kubenswrapper[4950]: E0318 20:11:49.615684 4950 event.go:368] "Unable to write event (may retry after sleeping)" err="Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/events/oauth-openshift-6b5f774455-xd699.189e088b5717aa7c\": dial tcp 38.102.83.163:6443: connect: connection refused" event="&Event{ObjectMeta:{oauth-openshift-6b5f774455-xd699.189e088b5717aa7c openshift-authentication 30025 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-authentication,Name:oauth-openshift-6b5f774455-xd699,UID:63b375d7-9a06-47aa-8c26-754e61bf2b91,APIVersion:v1,ResourceVersion:29991,FieldPath:spec.containers{oauth-openshift},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-18 20:11:37 +0000 UTC,LastTimestamp:2026-03-18 20:11:38.47752314 +0000 UTC m=+311.718365028,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 18 20:11:49 crc kubenswrapper[4950]: E0318 20:11:49.769956 4950 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.163:6443: connect: connection refused" interval="6.4s" Mar 18 20:11:51 crc kubenswrapper[4950]: I0318 20:11:51.600234 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 18 20:11:51 crc kubenswrapper[4950]: I0318 20:11:51.601483 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 18 20:11:51 crc kubenswrapper[4950]: I0318 20:11:51.601542 4950 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="5e662b624a83fcb0b0eab1048193fd00ce4a744f1a4653d776e0584aec856c49" exitCode=1 Mar 18 20:11:51 crc kubenswrapper[4950]: I0318 20:11:51.601580 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"5e662b624a83fcb0b0eab1048193fd00ce4a744f1a4653d776e0584aec856c49"} Mar 18 20:11:51 crc kubenswrapper[4950]: I0318 20:11:51.602088 4950 scope.go:117] "RemoveContainer" containerID="5e662b624a83fcb0b0eab1048193fd00ce4a744f1a4653d776e0584aec856c49" Mar 18 20:11:51 crc kubenswrapper[4950]: I0318 20:11:51.603609 4950 status_manager.go:851] "Failed to get status for pod" podUID="63b375d7-9a06-47aa-8c26-754e61bf2b91" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-6b5f774455-xd699\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:51 crc kubenswrapper[4950]: I0318 20:11:51.604246 4950 status_manager.go:851] "Failed to get status for pod" podUID="e7e1cabf-6954-4269-a638-16fb70a90850" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:51 crc kubenswrapper[4950]: I0318 20:11:51.604905 4950 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:52 crc kubenswrapper[4950]: I0318 20:11:52.625054 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 18 20:11:52 crc kubenswrapper[4950]: I0318 20:11:52.626203 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 18 20:11:52 crc kubenswrapper[4950]: I0318 20:11:52.626261 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"09ea3abb3c96a15302297ba6f39c8f02316ce89c93d1415fda445ae61800eed6"} Mar 18 20:11:52 crc kubenswrapper[4950]: I0318 20:11:52.627632 4950 status_manager.go:851] "Failed to get status for pod" podUID="63b375d7-9a06-47aa-8c26-754e61bf2b91" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-6b5f774455-xd699\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:52 crc kubenswrapper[4950]: I0318 20:11:52.628154 4950 status_manager.go:851] "Failed to get status for pod" podUID="e7e1cabf-6954-4269-a638-16fb70a90850" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:52 crc kubenswrapper[4950]: I0318 20:11:52.628773 4950 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:53 crc kubenswrapper[4950]: I0318 20:11:53.479865 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:11:53 crc kubenswrapper[4950]: I0318 20:11:53.480927 4950 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:53 crc kubenswrapper[4950]: I0318 20:11:53.481875 4950 status_manager.go:851] "Failed to get status for pod" podUID="63b375d7-9a06-47aa-8c26-754e61bf2b91" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-6b5f774455-xd699\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:53 crc kubenswrapper[4950]: I0318 20:11:53.482743 4950 status_manager.go:851] "Failed to get status for pod" podUID="e7e1cabf-6954-4269-a638-16fb70a90850" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:53 crc kubenswrapper[4950]: I0318 20:11:53.500572 4950 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="831bf4ca-f84c-48d2-9054-410480eb986b" Mar 18 20:11:53 crc kubenswrapper[4950]: I0318 20:11:53.500621 4950 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="831bf4ca-f84c-48d2-9054-410480eb986b" Mar 18 20:11:53 crc kubenswrapper[4950]: E0318 20:11:53.501260 4950 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:11:53 crc kubenswrapper[4950]: I0318 20:11:53.501910 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:11:53 crc kubenswrapper[4950]: W0318 20:11:53.533443 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-a822e5680cb2fca805b8d793dc0a8021e501d885add176d491bbdb219ae568d4 WatchSource:0}: Error finding container a822e5680cb2fca805b8d793dc0a8021e501d885add176d491bbdb219ae568d4: Status 404 returned error can't find the container with id a822e5680cb2fca805b8d793dc0a8021e501d885add176d491bbdb219ae568d4 Mar 18 20:11:53 crc kubenswrapper[4950]: I0318 20:11:53.633387 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a822e5680cb2fca805b8d793dc0a8021e501d885add176d491bbdb219ae568d4"} Mar 18 20:11:54 crc kubenswrapper[4950]: I0318 20:11:54.643444 4950 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="9af2d194904c5d7adf3e7538daaa054f1e771943c6a7f9eb52a80c494505b7ab" exitCode=0 Mar 18 20:11:54 crc kubenswrapper[4950]: I0318 20:11:54.643528 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"9af2d194904c5d7adf3e7538daaa054f1e771943c6a7f9eb52a80c494505b7ab"} Mar 18 20:11:54 crc kubenswrapper[4950]: I0318 20:11:54.643942 4950 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="831bf4ca-f84c-48d2-9054-410480eb986b" Mar 18 20:11:54 crc kubenswrapper[4950]: I0318 20:11:54.643974 4950 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="831bf4ca-f84c-48d2-9054-410480eb986b" Mar 18 20:11:54 crc kubenswrapper[4950]: E0318 20:11:54.644486 4950 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:11:54 crc kubenswrapper[4950]: I0318 20:11:54.645587 4950 status_manager.go:851] "Failed to get status for pod" podUID="63b375d7-9a06-47aa-8c26-754e61bf2b91" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-6b5f774455-xd699\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:54 crc kubenswrapper[4950]: I0318 20:11:54.646352 4950 status_manager.go:851] "Failed to get status for pod" podUID="e7e1cabf-6954-4269-a638-16fb70a90850" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:54 crc kubenswrapper[4950]: I0318 20:11:54.647339 4950 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.163:6443: connect: connection refused" Mar 18 20:11:55 crc kubenswrapper[4950]: I0318 20:11:55.458818 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:11:55 crc kubenswrapper[4950]: I0318 20:11:55.650026 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"245c3bfa3652cc300d420d349f0e2ba536ec6d459c68e139138e3308bd2173fd"} Mar 18 20:11:55 crc kubenswrapper[4950]: I0318 20:11:55.651537 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"1f183635caf6dad60a347ac4ed58590e22ebaa1d028a43d773da5701d07b7c6e"} Mar 18 20:11:55 crc kubenswrapper[4950]: I0318 20:11:55.651610 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"147c836880b8cb6faa1530e1ecf408f9edf45466d5b8bb8bb5cd5ba17b7af0a7"} Mar 18 20:11:56 crc kubenswrapper[4950]: I0318 20:11:56.658426 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"648bbc8a5dd94fd78f94bf10024784102d137052495a5e6969cd82a6f1f96d37"} Mar 18 20:11:56 crc kubenswrapper[4950]: I0318 20:11:56.659860 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b52912b7425af329c48742a5429188bdfa930bab150240bcf8382b9bfabd0e6b"} Mar 18 20:11:56 crc kubenswrapper[4950]: I0318 20:11:56.660294 4950 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="831bf4ca-f84c-48d2-9054-410480eb986b" Mar 18 20:11:56 crc kubenswrapper[4950]: I0318 20:11:56.660385 4950 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="831bf4ca-f84c-48d2-9054-410480eb986b" Mar 18 20:11:56 crc kubenswrapper[4950]: I0318 20:11:56.660532 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:11:58 crc kubenswrapper[4950]: I0318 20:11:58.299700 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:11:58 crc kubenswrapper[4950]: I0318 20:11:58.299968 4950 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 18 20:11:58 crc kubenswrapper[4950]: I0318 20:11:58.300037 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 18 20:11:58 crc kubenswrapper[4950]: I0318 20:11:58.502772 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:11:58 crc kubenswrapper[4950]: I0318 20:11:58.502839 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:11:58 crc kubenswrapper[4950]: I0318 20:11:58.512357 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:12:01 crc kubenswrapper[4950]: I0318 20:12:01.480631 4950 scope.go:117] "RemoveContainer" containerID="b61733f6eb8ab8c2ed359bc0e5745dccc9915c0cf005ef9fcb37187c46018ee0" Mar 18 20:12:01 crc kubenswrapper[4950]: I0318 20:12:01.689881 4950 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:12:01 crc kubenswrapper[4950]: I0318 20:12:01.766597 4950 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="8c799952-528f-4f92-b3fd-0329bc4e03a6" Mar 18 20:12:02 crc kubenswrapper[4950]: I0318 20:12:02.699790 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-6b5f774455-xd699_63b375d7-9a06-47aa-8c26-754e61bf2b91/oauth-openshift/2.log" Mar 18 20:12:02 crc kubenswrapper[4950]: I0318 20:12:02.700310 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-6b5f774455-xd699_63b375d7-9a06-47aa-8c26-754e61bf2b91/oauth-openshift/1.log" Mar 18 20:12:02 crc kubenswrapper[4950]: I0318 20:12:02.700361 4950 generic.go:334] "Generic (PLEG): container finished" podID="63b375d7-9a06-47aa-8c26-754e61bf2b91" containerID="ff6841998f2a9e3b30df4ec89243e8597a2db6945b2c46cadec1fd5193c8adf8" exitCode=255 Mar 18 20:12:02 crc kubenswrapper[4950]: I0318 20:12:02.700669 4950 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="831bf4ca-f84c-48d2-9054-410480eb986b" Mar 18 20:12:02 crc kubenswrapper[4950]: I0318 20:12:02.700683 4950 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="831bf4ca-f84c-48d2-9054-410480eb986b" Mar 18 20:12:02 crc kubenswrapper[4950]: I0318 20:12:02.701008 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" event={"ID":"63b375d7-9a06-47aa-8c26-754e61bf2b91","Type":"ContainerDied","Data":"ff6841998f2a9e3b30df4ec89243e8597a2db6945b2c46cadec1fd5193c8adf8"} Mar 18 20:12:02 crc kubenswrapper[4950]: I0318 20:12:02.701077 4950 scope.go:117] "RemoveContainer" containerID="b61733f6eb8ab8c2ed359bc0e5745dccc9915c0cf005ef9fcb37187c46018ee0" Mar 18 20:12:02 crc kubenswrapper[4950]: I0318 20:12:02.702144 4950 scope.go:117] "RemoveContainer" containerID="ff6841998f2a9e3b30df4ec89243e8597a2db6945b2c46cadec1fd5193c8adf8" Mar 18 20:12:02 crc kubenswrapper[4950]: E0318 20:12:02.702390 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 20s restarting failed container=oauth-openshift pod=oauth-openshift-6b5f774455-xd699_openshift-authentication(63b375d7-9a06-47aa-8c26-754e61bf2b91)\"" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" podUID="63b375d7-9a06-47aa-8c26-754e61bf2b91" Mar 18 20:12:02 crc kubenswrapper[4950]: I0318 20:12:02.705092 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:12:02 crc kubenswrapper[4950]: I0318 20:12:02.708780 4950 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="8c799952-528f-4f92-b3fd-0329bc4e03a6" Mar 18 20:12:03 crc kubenswrapper[4950]: I0318 20:12:03.709858 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-6b5f774455-xd699_63b375d7-9a06-47aa-8c26-754e61bf2b91/oauth-openshift/2.log" Mar 18 20:12:03 crc kubenswrapper[4950]: I0318 20:12:03.710706 4950 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="831bf4ca-f84c-48d2-9054-410480eb986b" Mar 18 20:12:03 crc kubenswrapper[4950]: I0318 20:12:03.710737 4950 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="831bf4ca-f84c-48d2-9054-410480eb986b" Mar 18 20:12:03 crc kubenswrapper[4950]: I0318 20:12:03.717391 4950 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="8c799952-528f-4f92-b3fd-0329bc4e03a6" Mar 18 20:12:07 crc kubenswrapper[4950]: I0318 20:12:07.543782 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:12:07 crc kubenswrapper[4950]: I0318 20:12:07.544199 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:12:07 crc kubenswrapper[4950]: I0318 20:12:07.544861 4950 scope.go:117] "RemoveContainer" containerID="ff6841998f2a9e3b30df4ec89243e8597a2db6945b2c46cadec1fd5193c8adf8" Mar 18 20:12:07 crc kubenswrapper[4950]: E0318 20:12:07.545196 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 20s restarting failed container=oauth-openshift pod=oauth-openshift-6b5f774455-xd699_openshift-authentication(63b375d7-9a06-47aa-8c26-754e61bf2b91)\"" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" podUID="63b375d7-9a06-47aa-8c26-754e61bf2b91" Mar 18 20:12:08 crc kubenswrapper[4950]: I0318 20:12:08.300809 4950 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 18 20:12:08 crc kubenswrapper[4950]: I0318 20:12:08.301159 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 18 20:12:10 crc kubenswrapper[4950]: I0318 20:12:10.871290 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 18 20:12:10 crc kubenswrapper[4950]: I0318 20:12:10.886905 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 18 20:12:10 crc kubenswrapper[4950]: I0318 20:12:10.902557 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 18 20:12:11 crc kubenswrapper[4950]: I0318 20:12:11.472073 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 18 20:12:11 crc kubenswrapper[4950]: I0318 20:12:11.665559 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 18 20:12:11 crc kubenswrapper[4950]: I0318 20:12:11.788659 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 18 20:12:12 crc kubenswrapper[4950]: I0318 20:12:12.050138 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 18 20:12:12 crc kubenswrapper[4950]: I0318 20:12:12.118330 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 18 20:12:12 crc kubenswrapper[4950]: I0318 20:12:12.164756 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 18 20:12:12 crc kubenswrapper[4950]: I0318 20:12:12.307930 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 18 20:12:12 crc kubenswrapper[4950]: I0318 20:12:12.400964 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 18 20:12:12 crc kubenswrapper[4950]: I0318 20:12:12.690085 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 18 20:12:12 crc kubenswrapper[4950]: I0318 20:12:12.836547 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 18 20:12:12 crc kubenswrapper[4950]: I0318 20:12:12.845866 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 18 20:12:13 crc kubenswrapper[4950]: I0318 20:12:13.099002 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 18 20:12:13 crc kubenswrapper[4950]: I0318 20:12:13.206462 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 18 20:12:13 crc kubenswrapper[4950]: I0318 20:12:13.221621 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 18 20:12:13 crc kubenswrapper[4950]: I0318 20:12:13.223344 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 18 20:12:13 crc kubenswrapper[4950]: I0318 20:12:13.716804 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 18 20:12:13 crc kubenswrapper[4950]: I0318 20:12:13.752693 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 18 20:12:13 crc kubenswrapper[4950]: I0318 20:12:13.799014 4950 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 18 20:12:14 crc kubenswrapper[4950]: I0318 20:12:14.161043 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 18 20:12:14 crc kubenswrapper[4950]: I0318 20:12:14.181676 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 18 20:12:14 crc kubenswrapper[4950]: I0318 20:12:14.367908 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 18 20:12:14 crc kubenswrapper[4950]: I0318 20:12:14.374784 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 18 20:12:14 crc kubenswrapper[4950]: I0318 20:12:14.462381 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 18 20:12:14 crc kubenswrapper[4950]: I0318 20:12:14.634475 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 18 20:12:14 crc kubenswrapper[4950]: I0318 20:12:14.654609 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 18 20:12:14 crc kubenswrapper[4950]: I0318 20:12:14.708035 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 18 20:12:14 crc kubenswrapper[4950]: I0318 20:12:14.818604 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 18 20:12:14 crc kubenswrapper[4950]: I0318 20:12:14.872463 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.071887 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.154692 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.155113 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.233931 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.337855 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.380202 4950 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.384932 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.391241 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.391312 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.399047 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.400670 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.407806 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=14.407789229 podStartE2EDuration="14.407789229s" podCreationTimestamp="2026-03-18 20:12:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:12:15.407481321 +0000 UTC m=+348.648323189" watchObservedRunningTime="2026-03-18 20:12:15.407789229 +0000 UTC m=+348.648631097" Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.424598 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.437293 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.465698 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.470233 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.474293 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.491108 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.707172 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.779862 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.839626 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.878764 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.887251 4950 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.920737 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.925997 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 18 20:12:15 crc kubenswrapper[4950]: I0318 20:12:15.931116 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 18 20:12:16 crc kubenswrapper[4950]: I0318 20:12:16.036696 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 18 20:12:16 crc kubenswrapper[4950]: I0318 20:12:16.039382 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 18 20:12:16 crc kubenswrapper[4950]: I0318 20:12:16.077038 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 18 20:12:16 crc kubenswrapper[4950]: I0318 20:12:16.129578 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 18 20:12:16 crc kubenswrapper[4950]: I0318 20:12:16.153723 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 18 20:12:16 crc kubenswrapper[4950]: I0318 20:12:16.154521 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 18 20:12:16 crc kubenswrapper[4950]: I0318 20:12:16.267130 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 18 20:12:16 crc kubenswrapper[4950]: I0318 20:12:16.310099 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 18 20:12:16 crc kubenswrapper[4950]: I0318 20:12:16.338062 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 18 20:12:16 crc kubenswrapper[4950]: I0318 20:12:16.416340 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 18 20:12:16 crc kubenswrapper[4950]: I0318 20:12:16.441205 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 18 20:12:16 crc kubenswrapper[4950]: I0318 20:12:16.469893 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 18 20:12:16 crc kubenswrapper[4950]: I0318 20:12:16.600899 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 18 20:12:16 crc kubenswrapper[4950]: I0318 20:12:16.750209 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 18 20:12:16 crc kubenswrapper[4950]: I0318 20:12:16.780734 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 18 20:12:16 crc kubenswrapper[4950]: I0318 20:12:16.786821 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 18 20:12:16 crc kubenswrapper[4950]: I0318 20:12:16.881626 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 18 20:12:16 crc kubenswrapper[4950]: I0318 20:12:16.904228 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 18 20:12:17 crc kubenswrapper[4950]: I0318 20:12:17.001630 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 18 20:12:17 crc kubenswrapper[4950]: I0318 20:12:17.208930 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 18 20:12:17 crc kubenswrapper[4950]: I0318 20:12:17.221995 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 18 20:12:17 crc kubenswrapper[4950]: I0318 20:12:17.306740 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 18 20:12:17 crc kubenswrapper[4950]: I0318 20:12:17.357723 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 18 20:12:17 crc kubenswrapper[4950]: I0318 20:12:17.390760 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 18 20:12:17 crc kubenswrapper[4950]: I0318 20:12:17.430927 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 18 20:12:17 crc kubenswrapper[4950]: I0318 20:12:17.497821 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 18 20:12:17 crc kubenswrapper[4950]: I0318 20:12:17.579400 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 18 20:12:17 crc kubenswrapper[4950]: I0318 20:12:17.586464 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 18 20:12:17 crc kubenswrapper[4950]: I0318 20:12:17.853143 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 18 20:12:17 crc kubenswrapper[4950]: I0318 20:12:17.917390 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 18 20:12:17 crc kubenswrapper[4950]: I0318 20:12:17.932693 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 18 20:12:17 crc kubenswrapper[4950]: I0318 20:12:17.948103 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 18 20:12:18 crc kubenswrapper[4950]: I0318 20:12:18.072809 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 18 20:12:18 crc kubenswrapper[4950]: I0318 20:12:18.104973 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 18 20:12:18 crc kubenswrapper[4950]: I0318 20:12:18.134840 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 18 20:12:18 crc kubenswrapper[4950]: I0318 20:12:18.146740 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 18 20:12:18 crc kubenswrapper[4950]: I0318 20:12:18.299897 4950 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 18 20:12:18 crc kubenswrapper[4950]: I0318 20:12:18.299950 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 18 20:12:18 crc kubenswrapper[4950]: I0318 20:12:18.300004 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:12:18 crc kubenswrapper[4950]: I0318 20:12:18.300676 4950 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"09ea3abb3c96a15302297ba6f39c8f02316ce89c93d1415fda445ae61800eed6"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Mar 18 20:12:18 crc kubenswrapper[4950]: I0318 20:12:18.300819 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://09ea3abb3c96a15302297ba6f39c8f02316ce89c93d1415fda445ae61800eed6" gracePeriod=30 Mar 18 20:12:18 crc kubenswrapper[4950]: I0318 20:12:18.329573 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 18 20:12:18 crc kubenswrapper[4950]: I0318 20:12:18.341759 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 18 20:12:18 crc kubenswrapper[4950]: I0318 20:12:18.350176 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 18 20:12:18 crc kubenswrapper[4950]: I0318 20:12:18.426457 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 18 20:12:18 crc kubenswrapper[4950]: I0318 20:12:18.521662 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 18 20:12:18 crc kubenswrapper[4950]: I0318 20:12:18.589644 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 18 20:12:18 crc kubenswrapper[4950]: I0318 20:12:18.800443 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 18 20:12:18 crc kubenswrapper[4950]: I0318 20:12:18.821687 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 18 20:12:18 crc kubenswrapper[4950]: I0318 20:12:18.959523 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 18 20:12:19 crc kubenswrapper[4950]: I0318 20:12:19.062589 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 18 20:12:19 crc kubenswrapper[4950]: I0318 20:12:19.063632 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 18 20:12:19 crc kubenswrapper[4950]: I0318 20:12:19.154815 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 18 20:12:19 crc kubenswrapper[4950]: I0318 20:12:19.266769 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 18 20:12:19 crc kubenswrapper[4950]: I0318 20:12:19.314544 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 18 20:12:19 crc kubenswrapper[4950]: I0318 20:12:19.336276 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 18 20:12:19 crc kubenswrapper[4950]: I0318 20:12:19.338105 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 18 20:12:19 crc kubenswrapper[4950]: I0318 20:12:19.361399 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 18 20:12:19 crc kubenswrapper[4950]: I0318 20:12:19.479217 4950 scope.go:117] "RemoveContainer" containerID="ff6841998f2a9e3b30df4ec89243e8597a2db6945b2c46cadec1fd5193c8adf8" Mar 18 20:12:19 crc kubenswrapper[4950]: E0318 20:12:19.479492 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 20s restarting failed container=oauth-openshift pod=oauth-openshift-6b5f774455-xd699_openshift-authentication(63b375d7-9a06-47aa-8c26-754e61bf2b91)\"" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" podUID="63b375d7-9a06-47aa-8c26-754e61bf2b91" Mar 18 20:12:19 crc kubenswrapper[4950]: I0318 20:12:19.552809 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 18 20:12:19 crc kubenswrapper[4950]: I0318 20:12:19.741220 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 18 20:12:19 crc kubenswrapper[4950]: I0318 20:12:19.744138 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 18 20:12:19 crc kubenswrapper[4950]: I0318 20:12:19.777958 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 18 20:12:19 crc kubenswrapper[4950]: I0318 20:12:19.859765 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 18 20:12:19 crc kubenswrapper[4950]: I0318 20:12:19.865525 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 18 20:12:19 crc kubenswrapper[4950]: I0318 20:12:19.925777 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 18 20:12:19 crc kubenswrapper[4950]: I0318 20:12:19.934442 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 18 20:12:19 crc kubenswrapper[4950]: I0318 20:12:19.960573 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 18 20:12:19 crc kubenswrapper[4950]: I0318 20:12:19.961924 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 18 20:12:20 crc kubenswrapper[4950]: I0318 20:12:20.226788 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 18 20:12:20 crc kubenswrapper[4950]: I0318 20:12:20.242705 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 18 20:12:20 crc kubenswrapper[4950]: I0318 20:12:20.269814 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 18 20:12:20 crc kubenswrapper[4950]: I0318 20:12:20.302113 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 18 20:12:20 crc kubenswrapper[4950]: I0318 20:12:20.358973 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 18 20:12:20 crc kubenswrapper[4950]: I0318 20:12:20.364676 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 18 20:12:20 crc kubenswrapper[4950]: I0318 20:12:20.398869 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 18 20:12:20 crc kubenswrapper[4950]: I0318 20:12:20.419685 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 18 20:12:20 crc kubenswrapper[4950]: I0318 20:12:20.530471 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 18 20:12:20 crc kubenswrapper[4950]: I0318 20:12:20.540677 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 18 20:12:20 crc kubenswrapper[4950]: I0318 20:12:20.770615 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 18 20:12:20 crc kubenswrapper[4950]: I0318 20:12:20.800837 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 18 20:12:20 crc kubenswrapper[4950]: I0318 20:12:20.802674 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 18 20:12:20 crc kubenswrapper[4950]: I0318 20:12:20.845498 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 18 20:12:20 crc kubenswrapper[4950]: I0318 20:12:20.856762 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 18 20:12:20 crc kubenswrapper[4950]: I0318 20:12:20.876801 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 18 20:12:20 crc kubenswrapper[4950]: I0318 20:12:20.892074 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 18 20:12:20 crc kubenswrapper[4950]: I0318 20:12:20.924573 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 18 20:12:20 crc kubenswrapper[4950]: I0318 20:12:20.964578 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 18 20:12:20 crc kubenswrapper[4950]: I0318 20:12:20.980604 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 18 20:12:21 crc kubenswrapper[4950]: I0318 20:12:21.024118 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 18 20:12:21 crc kubenswrapper[4950]: I0318 20:12:21.116904 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 18 20:12:21 crc kubenswrapper[4950]: I0318 20:12:21.122109 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 18 20:12:21 crc kubenswrapper[4950]: I0318 20:12:21.167340 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 18 20:12:21 crc kubenswrapper[4950]: I0318 20:12:21.197699 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 18 20:12:21 crc kubenswrapper[4950]: I0318 20:12:21.302719 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 18 20:12:21 crc kubenswrapper[4950]: I0318 20:12:21.366141 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 18 20:12:21 crc kubenswrapper[4950]: I0318 20:12:21.374981 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 18 20:12:21 crc kubenswrapper[4950]: I0318 20:12:21.420106 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 18 20:12:21 crc kubenswrapper[4950]: I0318 20:12:21.532906 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 18 20:12:21 crc kubenswrapper[4950]: I0318 20:12:21.538055 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 18 20:12:21 crc kubenswrapper[4950]: I0318 20:12:21.543544 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 18 20:12:21 crc kubenswrapper[4950]: I0318 20:12:21.570020 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 18 20:12:21 crc kubenswrapper[4950]: I0318 20:12:21.582658 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 18 20:12:21 crc kubenswrapper[4950]: I0318 20:12:21.589306 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 18 20:12:21 crc kubenswrapper[4950]: I0318 20:12:21.608865 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 18 20:12:21 crc kubenswrapper[4950]: I0318 20:12:21.661999 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 18 20:12:21 crc kubenswrapper[4950]: I0318 20:12:21.789861 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 18 20:12:21 crc kubenswrapper[4950]: I0318 20:12:21.794957 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 18 20:12:21 crc kubenswrapper[4950]: I0318 20:12:21.844401 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 18 20:12:21 crc kubenswrapper[4950]: I0318 20:12:21.868937 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 18 20:12:21 crc kubenswrapper[4950]: I0318 20:12:21.955704 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 18 20:12:22 crc kubenswrapper[4950]: I0318 20:12:22.070909 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 18 20:12:22 crc kubenswrapper[4950]: I0318 20:12:22.160652 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 18 20:12:22 crc kubenswrapper[4950]: I0318 20:12:22.168511 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 18 20:12:22 crc kubenswrapper[4950]: I0318 20:12:22.237047 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 18 20:12:22 crc kubenswrapper[4950]: I0318 20:12:22.237848 4950 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 18 20:12:22 crc kubenswrapper[4950]: I0318 20:12:22.265391 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 18 20:12:22 crc kubenswrapper[4950]: I0318 20:12:22.323446 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 18 20:12:22 crc kubenswrapper[4950]: I0318 20:12:22.365271 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 18 20:12:22 crc kubenswrapper[4950]: I0318 20:12:22.422079 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 18 20:12:22 crc kubenswrapper[4950]: I0318 20:12:22.495513 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 18 20:12:22 crc kubenswrapper[4950]: I0318 20:12:22.600578 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 18 20:12:22 crc kubenswrapper[4950]: I0318 20:12:22.645884 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 18 20:12:22 crc kubenswrapper[4950]: I0318 20:12:22.810098 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 18 20:12:22 crc kubenswrapper[4950]: I0318 20:12:22.818156 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 18 20:12:22 crc kubenswrapper[4950]: I0318 20:12:22.846792 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 18 20:12:23 crc kubenswrapper[4950]: I0318 20:12:23.080825 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 18 20:12:23 crc kubenswrapper[4950]: I0318 20:12:23.118780 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 18 20:12:23 crc kubenswrapper[4950]: I0318 20:12:23.144365 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 18 20:12:23 crc kubenswrapper[4950]: I0318 20:12:23.163277 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 18 20:12:23 crc kubenswrapper[4950]: I0318 20:12:23.398356 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 18 20:12:23 crc kubenswrapper[4950]: I0318 20:12:23.440157 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 18 20:12:23 crc kubenswrapper[4950]: I0318 20:12:23.457111 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 18 20:12:23 crc kubenswrapper[4950]: I0318 20:12:23.481902 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 18 20:12:23 crc kubenswrapper[4950]: I0318 20:12:23.486834 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 18 20:12:23 crc kubenswrapper[4950]: I0318 20:12:23.585394 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 18 20:12:23 crc kubenswrapper[4950]: I0318 20:12:23.701103 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 18 20:12:23 crc kubenswrapper[4950]: I0318 20:12:23.736258 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 18 20:12:23 crc kubenswrapper[4950]: I0318 20:12:23.797061 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 18 20:12:23 crc kubenswrapper[4950]: I0318 20:12:23.803566 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 18 20:12:23 crc kubenswrapper[4950]: I0318 20:12:23.958093 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 18 20:12:23 crc kubenswrapper[4950]: I0318 20:12:23.994978 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 18 20:12:23 crc kubenswrapper[4950]: I0318 20:12:23.996630 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 18 20:12:24 crc kubenswrapper[4950]: I0318 20:12:24.122504 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 18 20:12:24 crc kubenswrapper[4950]: I0318 20:12:24.151254 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 18 20:12:24 crc kubenswrapper[4950]: I0318 20:12:24.154862 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 18 20:12:24 crc kubenswrapper[4950]: I0318 20:12:24.166842 4950 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 18 20:12:24 crc kubenswrapper[4950]: I0318 20:12:24.167062 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://86227fc700ceea5952784deaf8f32fa67d24f9a70021929a4850d6c3f0bc5806" gracePeriod=5 Mar 18 20:12:24 crc kubenswrapper[4950]: I0318 20:12:24.227856 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 18 20:12:24 crc kubenswrapper[4950]: I0318 20:12:24.388342 4950 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 18 20:12:24 crc kubenswrapper[4950]: I0318 20:12:24.448338 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 18 20:12:24 crc kubenswrapper[4950]: I0318 20:12:24.589260 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 18 20:12:24 crc kubenswrapper[4950]: I0318 20:12:24.627211 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 18 20:12:24 crc kubenswrapper[4950]: I0318 20:12:24.688282 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 18 20:12:24 crc kubenswrapper[4950]: I0318 20:12:24.700795 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 18 20:12:24 crc kubenswrapper[4950]: I0318 20:12:24.741954 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 18 20:12:24 crc kubenswrapper[4950]: I0318 20:12:24.863085 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 18 20:12:24 crc kubenswrapper[4950]: I0318 20:12:24.866391 4950 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 18 20:12:24 crc kubenswrapper[4950]: I0318 20:12:24.942861 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 18 20:12:24 crc kubenswrapper[4950]: I0318 20:12:24.974215 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 18 20:12:25 crc kubenswrapper[4950]: I0318 20:12:25.070682 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 18 20:12:25 crc kubenswrapper[4950]: I0318 20:12:25.409309 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 18 20:12:25 crc kubenswrapper[4950]: I0318 20:12:25.477301 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 18 20:12:25 crc kubenswrapper[4950]: I0318 20:12:25.567289 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 18 20:12:25 crc kubenswrapper[4950]: I0318 20:12:25.588694 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 18 20:12:25 crc kubenswrapper[4950]: I0318 20:12:25.595132 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 18 20:12:25 crc kubenswrapper[4950]: I0318 20:12:25.596031 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 18 20:12:25 crc kubenswrapper[4950]: I0318 20:12:25.640244 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 18 20:12:25 crc kubenswrapper[4950]: I0318 20:12:25.693288 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 18 20:12:25 crc kubenswrapper[4950]: I0318 20:12:25.858720 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 18 20:12:25 crc kubenswrapper[4950]: I0318 20:12:25.957813 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 18 20:12:26 crc kubenswrapper[4950]: I0318 20:12:26.053912 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 18 20:12:26 crc kubenswrapper[4950]: I0318 20:12:26.105746 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 18 20:12:26 crc kubenswrapper[4950]: I0318 20:12:26.288223 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 18 20:12:26 crc kubenswrapper[4950]: I0318 20:12:26.297796 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 18 20:12:26 crc kubenswrapper[4950]: I0318 20:12:26.396482 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 18 20:12:26 crc kubenswrapper[4950]: I0318 20:12:26.412649 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 18 20:12:26 crc kubenswrapper[4950]: I0318 20:12:26.613163 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 18 20:12:26 crc kubenswrapper[4950]: I0318 20:12:26.635618 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 18 20:12:26 crc kubenswrapper[4950]: I0318 20:12:26.654808 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 18 20:12:26 crc kubenswrapper[4950]: I0318 20:12:26.680458 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 18 20:12:26 crc kubenswrapper[4950]: I0318 20:12:26.725514 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 18 20:12:26 crc kubenswrapper[4950]: I0318 20:12:26.750945 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 18 20:12:26 crc kubenswrapper[4950]: I0318 20:12:26.763617 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 18 20:12:26 crc kubenswrapper[4950]: I0318 20:12:26.783980 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 18 20:12:26 crc kubenswrapper[4950]: I0318 20:12:26.895440 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 18 20:12:27 crc kubenswrapper[4950]: I0318 20:12:27.071946 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 18 20:12:27 crc kubenswrapper[4950]: I0318 20:12:27.220901 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 18 20:12:27 crc kubenswrapper[4950]: I0318 20:12:27.416328 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 18 20:12:29 crc kubenswrapper[4950]: I0318 20:12:29.772771 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 18 20:12:29 crc kubenswrapper[4950]: I0318 20:12:29.773167 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 20:12:29 crc kubenswrapper[4950]: I0318 20:12:29.854257 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 18 20:12:29 crc kubenswrapper[4950]: I0318 20:12:29.854336 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 18 20:12:29 crc kubenswrapper[4950]: I0318 20:12:29.854385 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 18 20:12:29 crc kubenswrapper[4950]: I0318 20:12:29.854469 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:12:29 crc kubenswrapper[4950]: I0318 20:12:29.854502 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:12:29 crc kubenswrapper[4950]: I0318 20:12:29.854579 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 18 20:12:29 crc kubenswrapper[4950]: I0318 20:12:29.854635 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 18 20:12:29 crc kubenswrapper[4950]: I0318 20:12:29.854674 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:12:29 crc kubenswrapper[4950]: I0318 20:12:29.854769 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:12:29 crc kubenswrapper[4950]: I0318 20:12:29.855024 4950 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 18 20:12:29 crc kubenswrapper[4950]: I0318 20:12:29.855059 4950 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 18 20:12:29 crc kubenswrapper[4950]: I0318 20:12:29.855078 4950 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 18 20:12:29 crc kubenswrapper[4950]: I0318 20:12:29.855096 4950 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 18 20:12:29 crc kubenswrapper[4950]: I0318 20:12:29.868944 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:12:29 crc kubenswrapper[4950]: I0318 20:12:29.881203 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 18 20:12:29 crc kubenswrapper[4950]: I0318 20:12:29.881321 4950 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="86227fc700ceea5952784deaf8f32fa67d24f9a70021929a4850d6c3f0bc5806" exitCode=137 Mar 18 20:12:29 crc kubenswrapper[4950]: I0318 20:12:29.881376 4950 scope.go:117] "RemoveContainer" containerID="86227fc700ceea5952784deaf8f32fa67d24f9a70021929a4850d6c3f0bc5806" Mar 18 20:12:29 crc kubenswrapper[4950]: I0318 20:12:29.881486 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 18 20:12:29 crc kubenswrapper[4950]: I0318 20:12:29.923767 4950 scope.go:117] "RemoveContainer" containerID="86227fc700ceea5952784deaf8f32fa67d24f9a70021929a4850d6c3f0bc5806" Mar 18 20:12:29 crc kubenswrapper[4950]: E0318 20:12:29.924481 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86227fc700ceea5952784deaf8f32fa67d24f9a70021929a4850d6c3f0bc5806\": container with ID starting with 86227fc700ceea5952784deaf8f32fa67d24f9a70021929a4850d6c3f0bc5806 not found: ID does not exist" containerID="86227fc700ceea5952784deaf8f32fa67d24f9a70021929a4850d6c3f0bc5806" Mar 18 20:12:29 crc kubenswrapper[4950]: I0318 20:12:29.924536 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86227fc700ceea5952784deaf8f32fa67d24f9a70021929a4850d6c3f0bc5806"} err="failed to get container status \"86227fc700ceea5952784deaf8f32fa67d24f9a70021929a4850d6c3f0bc5806\": rpc error: code = NotFound desc = could not find container \"86227fc700ceea5952784deaf8f32fa67d24f9a70021929a4850d6c3f0bc5806\": container with ID starting with 86227fc700ceea5952784deaf8f32fa67d24f9a70021929a4850d6c3f0bc5806 not found: ID does not exist" Mar 18 20:12:29 crc kubenswrapper[4950]: I0318 20:12:29.956640 4950 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 18 20:12:31 crc kubenswrapper[4950]: I0318 20:12:31.480021 4950 scope.go:117] "RemoveContainer" containerID="ff6841998f2a9e3b30df4ec89243e8597a2db6945b2c46cadec1fd5193c8adf8" Mar 18 20:12:31 crc kubenswrapper[4950]: I0318 20:12:31.488147 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 18 20:12:31 crc kubenswrapper[4950]: I0318 20:12:31.896344 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-6b5f774455-xd699_63b375d7-9a06-47aa-8c26-754e61bf2b91/oauth-openshift/2.log" Mar 18 20:12:31 crc kubenswrapper[4950]: I0318 20:12:31.896742 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" event={"ID":"63b375d7-9a06-47aa-8c26-754e61bf2b91","Type":"ContainerStarted","Data":"06d122f56501c622ed37cbcad6864ece3f64fecdab6428a714b45a5d233f5a7c"} Mar 18 20:12:31 crc kubenswrapper[4950]: I0318 20:12:31.897226 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:12:31 crc kubenswrapper[4950]: I0318 20:12:31.922036 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" podStartSLOduration=88.922020727 podStartE2EDuration="1m28.922020727s" podCreationTimestamp="2026-03-18 20:11:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:12:31.920164464 +0000 UTC m=+365.161006332" watchObservedRunningTime="2026-03-18 20:12:31.922020727 +0000 UTC m=+365.162862595" Mar 18 20:12:32 crc kubenswrapper[4950]: I0318 20:12:32.154642 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6b5f774455-xd699" Mar 18 20:12:38 crc kubenswrapper[4950]: I0318 20:12:38.456615 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 18 20:12:39 crc kubenswrapper[4950]: I0318 20:12:39.113758 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 18 20:12:41 crc kubenswrapper[4950]: I0318 20:12:41.651400 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 18 20:12:43 crc kubenswrapper[4950]: I0318 20:12:43.177847 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 18 20:12:44 crc kubenswrapper[4950]: I0318 20:12:44.604801 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 18 20:12:44 crc kubenswrapper[4950]: I0318 20:12:44.928200 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 18 20:12:45 crc kubenswrapper[4950]: I0318 20:12:45.703334 4950 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-8nhwt container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Mar 18 20:12:45 crc kubenswrapper[4950]: I0318 20:12:45.703740 4950 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-8nhwt container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Mar 18 20:12:45 crc kubenswrapper[4950]: I0318 20:12:45.704256 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" podUID="a2456057-5ff6-458f-af94-53edca249b7f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" Mar 18 20:12:45 crc kubenswrapper[4950]: I0318 20:12:45.704209 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" podUID="a2456057-5ff6-458f-af94-53edca249b7f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" Mar 18 20:12:45 crc kubenswrapper[4950]: I0318 20:12:45.993264 4950 generic.go:334] "Generic (PLEG): container finished" podID="a2456057-5ff6-458f-af94-53edca249b7f" containerID="bb43772a7f0be4bd5df4460e9a3438e9991f1cee7c7ca14fcaf9f7e315f4121e" exitCode=0 Mar 18 20:12:45 crc kubenswrapper[4950]: I0318 20:12:45.993341 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" event={"ID":"a2456057-5ff6-458f-af94-53edca249b7f","Type":"ContainerDied","Data":"bb43772a7f0be4bd5df4460e9a3438e9991f1cee7c7ca14fcaf9f7e315f4121e"} Mar 18 20:12:45 crc kubenswrapper[4950]: I0318 20:12:45.994519 4950 scope.go:117] "RemoveContainer" containerID="bb43772a7f0be4bd5df4460e9a3438e9991f1cee7c7ca14fcaf9f7e315f4121e" Mar 18 20:12:46 crc kubenswrapper[4950]: I0318 20:12:46.143270 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 18 20:12:47 crc kubenswrapper[4950]: I0318 20:12:47.002582 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" event={"ID":"a2456057-5ff6-458f-af94-53edca249b7f","Type":"ContainerStarted","Data":"7164a8254b584d66dd328ab9294f1df8785a45c6f9bdd3507b1d426b574815f0"} Mar 18 20:12:47 crc kubenswrapper[4950]: I0318 20:12:47.003889 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" Mar 18 20:12:47 crc kubenswrapper[4950]: I0318 20:12:47.008528 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" Mar 18 20:12:48 crc kubenswrapper[4950]: I0318 20:12:48.058610 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 18 20:12:49 crc kubenswrapper[4950]: I0318 20:12:49.014579 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Mar 18 20:12:49 crc kubenswrapper[4950]: I0318 20:12:49.016717 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 18 20:12:49 crc kubenswrapper[4950]: I0318 20:12:49.017299 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 18 20:12:49 crc kubenswrapper[4950]: I0318 20:12:49.017351 4950 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="09ea3abb3c96a15302297ba6f39c8f02316ce89c93d1415fda445ae61800eed6" exitCode=137 Mar 18 20:12:49 crc kubenswrapper[4950]: I0318 20:12:49.017471 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"09ea3abb3c96a15302297ba6f39c8f02316ce89c93d1415fda445ae61800eed6"} Mar 18 20:12:49 crc kubenswrapper[4950]: I0318 20:12:49.017531 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b56e4cfdc7bc5a89021aed88b3f7f692e797d1370684ec582ccbab1767b2acaf"} Mar 18 20:12:49 crc kubenswrapper[4950]: I0318 20:12:49.017554 4950 scope.go:117] "RemoveContainer" containerID="5e662b624a83fcb0b0eab1048193fd00ce4a744f1a4653d776e0584aec856c49" Mar 18 20:12:49 crc kubenswrapper[4950]: I0318 20:12:49.600204 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5fnpx"] Mar 18 20:12:49 crc kubenswrapper[4950]: I0318 20:12:49.602851 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5fnpx" podUID="d3ae2e86-be5d-4546-97a0-88301381868b" containerName="registry-server" containerID="cri-o://3752f560059d7127c4b227d082fedc6224ab2e2c7a96fab6dd6f9ab0b1f652d1" gracePeriod=2 Mar 18 20:12:49 crc kubenswrapper[4950]: I0318 20:12:49.751773 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 18 20:12:49 crc kubenswrapper[4950]: I0318 20:12:49.784045 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.016107 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5fnpx" Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.023960 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.025086 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.027348 4950 generic.go:334] "Generic (PLEG): container finished" podID="d3ae2e86-be5d-4546-97a0-88301381868b" containerID="3752f560059d7127c4b227d082fedc6224ab2e2c7a96fab6dd6f9ab0b1f652d1" exitCode=0 Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.027380 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5fnpx" Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.027389 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fnpx" event={"ID":"d3ae2e86-be5d-4546-97a0-88301381868b","Type":"ContainerDied","Data":"3752f560059d7127c4b227d082fedc6224ab2e2c7a96fab6dd6f9ab0b1f652d1"} Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.027432 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5fnpx" event={"ID":"d3ae2e86-be5d-4546-97a0-88301381868b","Type":"ContainerDied","Data":"7666b6d629d01dd8ff0cea990e91a38418786e06ba35febd7d3ec855d70babc3"} Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.027456 4950 scope.go:117] "RemoveContainer" containerID="3752f560059d7127c4b227d082fedc6224ab2e2c7a96fab6dd6f9ab0b1f652d1" Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.049618 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.078572 4950 scope.go:117] "RemoveContainer" containerID="7308adae68168391c9ddefc29042a8b6c7245e1cda57a4b8f90ea857c8fe20a3" Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.094138 4950 scope.go:117] "RemoveContainer" containerID="69309d7fec17ae4fbdf231e4dac66b9f7c29ca46eedc368111f5776bcd93cf45" Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.110455 4950 scope.go:117] "RemoveContainer" containerID="3752f560059d7127c4b227d082fedc6224ab2e2c7a96fab6dd6f9ab0b1f652d1" Mar 18 20:12:50 crc kubenswrapper[4950]: E0318 20:12:50.110798 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3752f560059d7127c4b227d082fedc6224ab2e2c7a96fab6dd6f9ab0b1f652d1\": container with ID starting with 3752f560059d7127c4b227d082fedc6224ab2e2c7a96fab6dd6f9ab0b1f652d1 not found: ID does not exist" containerID="3752f560059d7127c4b227d082fedc6224ab2e2c7a96fab6dd6f9ab0b1f652d1" Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.110829 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3752f560059d7127c4b227d082fedc6224ab2e2c7a96fab6dd6f9ab0b1f652d1"} err="failed to get container status \"3752f560059d7127c4b227d082fedc6224ab2e2c7a96fab6dd6f9ab0b1f652d1\": rpc error: code = NotFound desc = could not find container \"3752f560059d7127c4b227d082fedc6224ab2e2c7a96fab6dd6f9ab0b1f652d1\": container with ID starting with 3752f560059d7127c4b227d082fedc6224ab2e2c7a96fab6dd6f9ab0b1f652d1 not found: ID does not exist" Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.110850 4950 scope.go:117] "RemoveContainer" containerID="7308adae68168391c9ddefc29042a8b6c7245e1cda57a4b8f90ea857c8fe20a3" Mar 18 20:12:50 crc kubenswrapper[4950]: E0318 20:12:50.111198 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7308adae68168391c9ddefc29042a8b6c7245e1cda57a4b8f90ea857c8fe20a3\": container with ID starting with 7308adae68168391c9ddefc29042a8b6c7245e1cda57a4b8f90ea857c8fe20a3 not found: ID does not exist" containerID="7308adae68168391c9ddefc29042a8b6c7245e1cda57a4b8f90ea857c8fe20a3" Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.111239 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7308adae68168391c9ddefc29042a8b6c7245e1cda57a4b8f90ea857c8fe20a3"} err="failed to get container status \"7308adae68168391c9ddefc29042a8b6c7245e1cda57a4b8f90ea857c8fe20a3\": rpc error: code = NotFound desc = could not find container \"7308adae68168391c9ddefc29042a8b6c7245e1cda57a4b8f90ea857c8fe20a3\": container with ID starting with 7308adae68168391c9ddefc29042a8b6c7245e1cda57a4b8f90ea857c8fe20a3 not found: ID does not exist" Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.111287 4950 scope.go:117] "RemoveContainer" containerID="69309d7fec17ae4fbdf231e4dac66b9f7c29ca46eedc368111f5776bcd93cf45" Mar 18 20:12:50 crc kubenswrapper[4950]: E0318 20:12:50.111764 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69309d7fec17ae4fbdf231e4dac66b9f7c29ca46eedc368111f5776bcd93cf45\": container with ID starting with 69309d7fec17ae4fbdf231e4dac66b9f7c29ca46eedc368111f5776bcd93cf45 not found: ID does not exist" containerID="69309d7fec17ae4fbdf231e4dac66b9f7c29ca46eedc368111f5776bcd93cf45" Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.111791 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69309d7fec17ae4fbdf231e4dac66b9f7c29ca46eedc368111f5776bcd93cf45"} err="failed to get container status \"69309d7fec17ae4fbdf231e4dac66b9f7c29ca46eedc368111f5776bcd93cf45\": rpc error: code = NotFound desc = could not find container \"69309d7fec17ae4fbdf231e4dac66b9f7c29ca46eedc368111f5776bcd93cf45\": container with ID starting with 69309d7fec17ae4fbdf231e4dac66b9f7c29ca46eedc368111f5776bcd93cf45 not found: ID does not exist" Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.170038 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbj6d\" (UniqueName: \"kubernetes.io/projected/d3ae2e86-be5d-4546-97a0-88301381868b-kube-api-access-lbj6d\") pod \"d3ae2e86-be5d-4546-97a0-88301381868b\" (UID: \"d3ae2e86-be5d-4546-97a0-88301381868b\") " Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.170116 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3ae2e86-be5d-4546-97a0-88301381868b-catalog-content\") pod \"d3ae2e86-be5d-4546-97a0-88301381868b\" (UID: \"d3ae2e86-be5d-4546-97a0-88301381868b\") " Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.170156 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3ae2e86-be5d-4546-97a0-88301381868b-utilities\") pod \"d3ae2e86-be5d-4546-97a0-88301381868b\" (UID: \"d3ae2e86-be5d-4546-97a0-88301381868b\") " Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.171061 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3ae2e86-be5d-4546-97a0-88301381868b-utilities" (OuterVolumeSpecName: "utilities") pod "d3ae2e86-be5d-4546-97a0-88301381868b" (UID: "d3ae2e86-be5d-4546-97a0-88301381868b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.175567 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3ae2e86-be5d-4546-97a0-88301381868b-kube-api-access-lbj6d" (OuterVolumeSpecName: "kube-api-access-lbj6d") pod "d3ae2e86-be5d-4546-97a0-88301381868b" (UID: "d3ae2e86-be5d-4546-97a0-88301381868b"). InnerVolumeSpecName "kube-api-access-lbj6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.271104 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbj6d\" (UniqueName: \"kubernetes.io/projected/d3ae2e86-be5d-4546-97a0-88301381868b-kube-api-access-lbj6d\") on node \"crc\" DevicePath \"\"" Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.271139 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3ae2e86-be5d-4546-97a0-88301381868b-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.318830 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3ae2e86-be5d-4546-97a0-88301381868b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d3ae2e86-be5d-4546-97a0-88301381868b" (UID: "d3ae2e86-be5d-4546-97a0-88301381868b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.355564 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5fnpx"] Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.357784 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5fnpx"] Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.372204 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3ae2e86-be5d-4546-97a0-88301381868b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:12:50 crc kubenswrapper[4950]: I0318 20:12:50.886371 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 18 20:12:51 crc kubenswrapper[4950]: I0318 20:12:51.050511 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 18 20:12:51 crc kubenswrapper[4950]: I0318 20:12:51.416736 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 18 20:12:51 crc kubenswrapper[4950]: I0318 20:12:51.491248 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3ae2e86-be5d-4546-97a0-88301381868b" path="/var/lib/kubelet/pods/d3ae2e86-be5d-4546-97a0-88301381868b/volumes" Mar 18 20:12:51 crc kubenswrapper[4950]: I0318 20:12:51.595062 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 18 20:12:55 crc kubenswrapper[4950]: I0318 20:12:55.457722 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:12:55 crc kubenswrapper[4950]: I0318 20:12:55.517248 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 18 20:12:58 crc kubenswrapper[4950]: I0318 20:12:58.300271 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:12:58 crc kubenswrapper[4950]: I0318 20:12:58.306021 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:12:59 crc kubenswrapper[4950]: I0318 20:12:59.084485 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 18 20:13:05 crc kubenswrapper[4950]: I0318 20:13:05.212404 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 18 20:13:08 crc kubenswrapper[4950]: I0318 20:13:08.157606 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 18 20:13:14 crc kubenswrapper[4950]: I0318 20:13:14.719882 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564412-s92bw"] Mar 18 20:13:14 crc kubenswrapper[4950]: E0318 20:13:14.720537 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3ae2e86-be5d-4546-97a0-88301381868b" containerName="extract-content" Mar 18 20:13:14 crc kubenswrapper[4950]: I0318 20:13:14.720551 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3ae2e86-be5d-4546-97a0-88301381868b" containerName="extract-content" Mar 18 20:13:14 crc kubenswrapper[4950]: E0318 20:13:14.720564 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 18 20:13:14 crc kubenswrapper[4950]: I0318 20:13:14.720570 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 18 20:13:14 crc kubenswrapper[4950]: E0318 20:13:14.720579 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3ae2e86-be5d-4546-97a0-88301381868b" containerName="registry-server" Mar 18 20:13:14 crc kubenswrapper[4950]: I0318 20:13:14.720585 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3ae2e86-be5d-4546-97a0-88301381868b" containerName="registry-server" Mar 18 20:13:14 crc kubenswrapper[4950]: E0318 20:13:14.720597 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3ae2e86-be5d-4546-97a0-88301381868b" containerName="extract-utilities" Mar 18 20:13:14 crc kubenswrapper[4950]: I0318 20:13:14.720603 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3ae2e86-be5d-4546-97a0-88301381868b" containerName="extract-utilities" Mar 18 20:13:14 crc kubenswrapper[4950]: E0318 20:13:14.720612 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7e1cabf-6954-4269-a638-16fb70a90850" containerName="installer" Mar 18 20:13:14 crc kubenswrapper[4950]: I0318 20:13:14.720617 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7e1cabf-6954-4269-a638-16fb70a90850" containerName="installer" Mar 18 20:13:14 crc kubenswrapper[4950]: I0318 20:13:14.720702 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7e1cabf-6954-4269-a638-16fb70a90850" containerName="installer" Mar 18 20:13:14 crc kubenswrapper[4950]: I0318 20:13:14.720710 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3ae2e86-be5d-4546-97a0-88301381868b" containerName="registry-server" Mar 18 20:13:14 crc kubenswrapper[4950]: I0318 20:13:14.720721 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 18 20:13:14 crc kubenswrapper[4950]: I0318 20:13:14.721049 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564412-s92bw" Mar 18 20:13:14 crc kubenswrapper[4950]: I0318 20:13:14.723729 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:13:14 crc kubenswrapper[4950]: I0318 20:13:14.723886 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:13:14 crc kubenswrapper[4950]: I0318 20:13:14.728317 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:13:14 crc kubenswrapper[4950]: I0318 20:13:14.728951 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564412-s92bw"] Mar 18 20:13:14 crc kubenswrapper[4950]: I0318 20:13:14.791862 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srx8t\" (UniqueName: \"kubernetes.io/projected/b7e9b7e8-14e9-422a-8b97-437f88df2595-kube-api-access-srx8t\") pod \"auto-csr-approver-29564412-s92bw\" (UID: \"b7e9b7e8-14e9-422a-8b97-437f88df2595\") " pod="openshift-infra/auto-csr-approver-29564412-s92bw" Mar 18 20:13:14 crc kubenswrapper[4950]: I0318 20:13:14.892974 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srx8t\" (UniqueName: \"kubernetes.io/projected/b7e9b7e8-14e9-422a-8b97-437f88df2595-kube-api-access-srx8t\") pod \"auto-csr-approver-29564412-s92bw\" (UID: \"b7e9b7e8-14e9-422a-8b97-437f88df2595\") " pod="openshift-infra/auto-csr-approver-29564412-s92bw" Mar 18 20:13:14 crc kubenswrapper[4950]: I0318 20:13:14.933704 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srx8t\" (UniqueName: \"kubernetes.io/projected/b7e9b7e8-14e9-422a-8b97-437f88df2595-kube-api-access-srx8t\") pod \"auto-csr-approver-29564412-s92bw\" (UID: \"b7e9b7e8-14e9-422a-8b97-437f88df2595\") " pod="openshift-infra/auto-csr-approver-29564412-s92bw" Mar 18 20:13:15 crc kubenswrapper[4950]: I0318 20:13:15.035698 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564412-s92bw" Mar 18 20:13:15 crc kubenswrapper[4950]: I0318 20:13:15.497700 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564412-s92bw"] Mar 18 20:13:16 crc kubenswrapper[4950]: I0318 20:13:16.175681 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564412-s92bw" event={"ID":"b7e9b7e8-14e9-422a-8b97-437f88df2595","Type":"ContainerStarted","Data":"7a7fd0f88eb388acf446d795d26c931b56a6505c7b9540dae4527fe558c3a9f4"} Mar 18 20:13:17 crc kubenswrapper[4950]: I0318 20:13:17.181250 4950 generic.go:334] "Generic (PLEG): container finished" podID="b7e9b7e8-14e9-422a-8b97-437f88df2595" containerID="21071767903abe4cc9037d6899818f503456804578d981c8dcfc413f8ec77034" exitCode=0 Mar 18 20:13:17 crc kubenswrapper[4950]: I0318 20:13:17.181423 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564412-s92bw" event={"ID":"b7e9b7e8-14e9-422a-8b97-437f88df2595","Type":"ContainerDied","Data":"21071767903abe4cc9037d6899818f503456804578d981c8dcfc413f8ec77034"} Mar 18 20:13:18 crc kubenswrapper[4950]: I0318 20:13:18.498301 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564412-s92bw" Mar 18 20:13:18 crc kubenswrapper[4950]: I0318 20:13:18.636502 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srx8t\" (UniqueName: \"kubernetes.io/projected/b7e9b7e8-14e9-422a-8b97-437f88df2595-kube-api-access-srx8t\") pod \"b7e9b7e8-14e9-422a-8b97-437f88df2595\" (UID: \"b7e9b7e8-14e9-422a-8b97-437f88df2595\") " Mar 18 20:13:18 crc kubenswrapper[4950]: I0318 20:13:18.641881 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7e9b7e8-14e9-422a-8b97-437f88df2595-kube-api-access-srx8t" (OuterVolumeSpecName: "kube-api-access-srx8t") pod "b7e9b7e8-14e9-422a-8b97-437f88df2595" (UID: "b7e9b7e8-14e9-422a-8b97-437f88df2595"). InnerVolumeSpecName "kube-api-access-srx8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:13:18 crc kubenswrapper[4950]: I0318 20:13:18.739195 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srx8t\" (UniqueName: \"kubernetes.io/projected/b7e9b7e8-14e9-422a-8b97-437f88df2595-kube-api-access-srx8t\") on node \"crc\" DevicePath \"\"" Mar 18 20:13:19 crc kubenswrapper[4950]: I0318 20:13:19.199042 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564412-s92bw" event={"ID":"b7e9b7e8-14e9-422a-8b97-437f88df2595","Type":"ContainerDied","Data":"7a7fd0f88eb388acf446d795d26c931b56a6505c7b9540dae4527fe558c3a9f4"} Mar 18 20:13:19 crc kubenswrapper[4950]: I0318 20:13:19.199084 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a7fd0f88eb388acf446d795d26c931b56a6505c7b9540dae4527fe558c3a9f4" Mar 18 20:13:19 crc kubenswrapper[4950]: I0318 20:13:19.199143 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564412-s92bw" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.221246 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8kwk7"] Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.226227 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8kwk7" podUID="4a386d81-3e2c-4df7-988d-5fb1758be616" containerName="registry-server" containerID="cri-o://81f4ca959b3b2d24b3c009f7e373b147ee24bca5aaaf63ab6988946a341b5cc6" gracePeriod=30 Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.263057 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gzj5q"] Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.263118 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8nhwt"] Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.263319 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" podUID="a2456057-5ff6-458f-af94-53edca249b7f" containerName="marketplace-operator" containerID="cri-o://7164a8254b584d66dd328ab9294f1df8785a45c6f9bdd3507b1d426b574815f0" gracePeriod=30 Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.263800 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gzj5q" podUID="044d9d2a-6b99-49ee-b00d-728c7832815c" containerName="registry-server" containerID="cri-o://a9132437997afe87402a5c278ffd8f59b66a8f2a8fa686e53a6c74092124a735" gracePeriod=30 Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.266442 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7dzz6"] Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.266711 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7dzz6" podUID="bbeed1b2-6abb-49df-94d9-7d9fe6e5a563" containerName="registry-server" containerID="cri-o://1fd9ccb388e588608b540567d2013892106bacdc049f5bcd80b2066229e558f3" gracePeriod=30 Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.276585 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nb7fx"] Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.276777 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nb7fx" podUID="33a99c2b-09d6-4ee7-b875-509ef5e144eb" containerName="registry-server" containerID="cri-o://866301b46a0fa42f5ddd84e80734a99c7495f2e22720901af0847f2ada691dcb" gracePeriod=30 Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.288224 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c6tfb"] Mar 18 20:13:27 crc kubenswrapper[4950]: E0318 20:13:27.288477 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7e9b7e8-14e9-422a-8b97-437f88df2595" containerName="oc" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.288490 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7e9b7e8-14e9-422a-8b97-437f88df2595" containerName="oc" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.288580 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7e9b7e8-14e9-422a-8b97-437f88df2595" containerName="oc" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.288928 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-c6tfb" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.293725 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c6tfb"] Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.347985 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27ccz\" (UniqueName: \"kubernetes.io/projected/51b0360c-046e-4360-b240-fdfd8d857a4b-kube-api-access-27ccz\") pod \"marketplace-operator-79b997595-c6tfb\" (UID: \"51b0360c-046e-4360-b240-fdfd8d857a4b\") " pod="openshift-marketplace/marketplace-operator-79b997595-c6tfb" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.348032 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/51b0360c-046e-4360-b240-fdfd8d857a4b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-c6tfb\" (UID: \"51b0360c-046e-4360-b240-fdfd8d857a4b\") " pod="openshift-marketplace/marketplace-operator-79b997595-c6tfb" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.348106 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/51b0360c-046e-4360-b240-fdfd8d857a4b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-c6tfb\" (UID: \"51b0360c-046e-4360-b240-fdfd8d857a4b\") " pod="openshift-marketplace/marketplace-operator-79b997595-c6tfb" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.449725 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/51b0360c-046e-4360-b240-fdfd8d857a4b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-c6tfb\" (UID: \"51b0360c-046e-4360-b240-fdfd8d857a4b\") " pod="openshift-marketplace/marketplace-operator-79b997595-c6tfb" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.450074 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27ccz\" (UniqueName: \"kubernetes.io/projected/51b0360c-046e-4360-b240-fdfd8d857a4b-kube-api-access-27ccz\") pod \"marketplace-operator-79b997595-c6tfb\" (UID: \"51b0360c-046e-4360-b240-fdfd8d857a4b\") " pod="openshift-marketplace/marketplace-operator-79b997595-c6tfb" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.450103 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/51b0360c-046e-4360-b240-fdfd8d857a4b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-c6tfb\" (UID: \"51b0360c-046e-4360-b240-fdfd8d857a4b\") " pod="openshift-marketplace/marketplace-operator-79b997595-c6tfb" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.452961 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/51b0360c-046e-4360-b240-fdfd8d857a4b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-c6tfb\" (UID: \"51b0360c-046e-4360-b240-fdfd8d857a4b\") " pod="openshift-marketplace/marketplace-operator-79b997595-c6tfb" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.457784 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/51b0360c-046e-4360-b240-fdfd8d857a4b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-c6tfb\" (UID: \"51b0360c-046e-4360-b240-fdfd8d857a4b\") " pod="openshift-marketplace/marketplace-operator-79b997595-c6tfb" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.471707 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27ccz\" (UniqueName: \"kubernetes.io/projected/51b0360c-046e-4360-b240-fdfd8d857a4b-kube-api-access-27ccz\") pod \"marketplace-operator-79b997595-c6tfb\" (UID: \"51b0360c-046e-4360-b240-fdfd8d857a4b\") " pod="openshift-marketplace/marketplace-operator-79b997595-c6tfb" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.527870 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-c6tfb" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.593825 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8kwk7" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.714350 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gzj5q" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.720514 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7dzz6" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.754928 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a386d81-3e2c-4df7-988d-5fb1758be616-utilities\") pod \"4a386d81-3e2c-4df7-988d-5fb1758be616\" (UID: \"4a386d81-3e2c-4df7-988d-5fb1758be616\") " Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.755003 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgcwj\" (UniqueName: \"kubernetes.io/projected/4a386d81-3e2c-4df7-988d-5fb1758be616-kube-api-access-bgcwj\") pod \"4a386d81-3e2c-4df7-988d-5fb1758be616\" (UID: \"4a386d81-3e2c-4df7-988d-5fb1758be616\") " Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.755037 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a386d81-3e2c-4df7-988d-5fb1758be616-catalog-content\") pod \"4a386d81-3e2c-4df7-988d-5fb1758be616\" (UID: \"4a386d81-3e2c-4df7-988d-5fb1758be616\") " Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.756333 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a386d81-3e2c-4df7-988d-5fb1758be616-utilities" (OuterVolumeSpecName: "utilities") pod "4a386d81-3e2c-4df7-988d-5fb1758be616" (UID: "4a386d81-3e2c-4df7-988d-5fb1758be616"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.777818 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a386d81-3e2c-4df7-988d-5fb1758be616-kube-api-access-bgcwj" (OuterVolumeSpecName: "kube-api-access-bgcwj") pod "4a386d81-3e2c-4df7-988d-5fb1758be616" (UID: "4a386d81-3e2c-4df7-988d-5fb1758be616"). InnerVolumeSpecName "kube-api-access-bgcwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.855605 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/044d9d2a-6b99-49ee-b00d-728c7832815c-catalog-content\") pod \"044d9d2a-6b99-49ee-b00d-728c7832815c\" (UID: \"044d9d2a-6b99-49ee-b00d-728c7832815c\") " Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.855666 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slf9m\" (UniqueName: \"kubernetes.io/projected/044d9d2a-6b99-49ee-b00d-728c7832815c-kube-api-access-slf9m\") pod \"044d9d2a-6b99-49ee-b00d-728c7832815c\" (UID: \"044d9d2a-6b99-49ee-b00d-728c7832815c\") " Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.855718 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbeed1b2-6abb-49df-94d9-7d9fe6e5a563-catalog-content\") pod \"bbeed1b2-6abb-49df-94d9-7d9fe6e5a563\" (UID: \"bbeed1b2-6abb-49df-94d9-7d9fe6e5a563\") " Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.855766 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/044d9d2a-6b99-49ee-b00d-728c7832815c-utilities\") pod \"044d9d2a-6b99-49ee-b00d-728c7832815c\" (UID: \"044d9d2a-6b99-49ee-b00d-728c7832815c\") " Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.855834 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbeed1b2-6abb-49df-94d9-7d9fe6e5a563-utilities\") pod \"bbeed1b2-6abb-49df-94d9-7d9fe6e5a563\" (UID: \"bbeed1b2-6abb-49df-94d9-7d9fe6e5a563\") " Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.855868 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9n22\" (UniqueName: \"kubernetes.io/projected/bbeed1b2-6abb-49df-94d9-7d9fe6e5a563-kube-api-access-x9n22\") pod \"bbeed1b2-6abb-49df-94d9-7d9fe6e5a563\" (UID: \"bbeed1b2-6abb-49df-94d9-7d9fe6e5a563\") " Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.856083 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a386d81-3e2c-4df7-988d-5fb1758be616-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.856097 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgcwj\" (UniqueName: \"kubernetes.io/projected/4a386d81-3e2c-4df7-988d-5fb1758be616-kube-api-access-bgcwj\") on node \"crc\" DevicePath \"\"" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.857579 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbeed1b2-6abb-49df-94d9-7d9fe6e5a563-utilities" (OuterVolumeSpecName: "utilities") pod "bbeed1b2-6abb-49df-94d9-7d9fe6e5a563" (UID: "bbeed1b2-6abb-49df-94d9-7d9fe6e5a563"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.857617 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/044d9d2a-6b99-49ee-b00d-728c7832815c-utilities" (OuterVolumeSpecName: "utilities") pod "044d9d2a-6b99-49ee-b00d-728c7832815c" (UID: "044d9d2a-6b99-49ee-b00d-728c7832815c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.860220 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/044d9d2a-6b99-49ee-b00d-728c7832815c-kube-api-access-slf9m" (OuterVolumeSpecName: "kube-api-access-slf9m") pod "044d9d2a-6b99-49ee-b00d-728c7832815c" (UID: "044d9d2a-6b99-49ee-b00d-728c7832815c"). InnerVolumeSpecName "kube-api-access-slf9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.861578 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbeed1b2-6abb-49df-94d9-7d9fe6e5a563-kube-api-access-x9n22" (OuterVolumeSpecName: "kube-api-access-x9n22") pod "bbeed1b2-6abb-49df-94d9-7d9fe6e5a563" (UID: "bbeed1b2-6abb-49df-94d9-7d9fe6e5a563"). InnerVolumeSpecName "kube-api-access-x9n22". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.904193 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.910801 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nb7fx" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.920117 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbeed1b2-6abb-49df-94d9-7d9fe6e5a563-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bbeed1b2-6abb-49df-94d9-7d9fe6e5a563" (UID: "bbeed1b2-6abb-49df-94d9-7d9fe6e5a563"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.951109 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a386d81-3e2c-4df7-988d-5fb1758be616-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4a386d81-3e2c-4df7-988d-5fb1758be616" (UID: "4a386d81-3e2c-4df7-988d-5fb1758be616"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.951804 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/044d9d2a-6b99-49ee-b00d-728c7832815c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "044d9d2a-6b99-49ee-b00d-728c7832815c" (UID: "044d9d2a-6b99-49ee-b00d-728c7832815c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.957452 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a2456057-5ff6-458f-af94-53edca249b7f-marketplace-operator-metrics\") pod \"a2456057-5ff6-458f-af94-53edca249b7f\" (UID: \"a2456057-5ff6-458f-af94-53edca249b7f\") " Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.957631 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a2456057-5ff6-458f-af94-53edca249b7f-marketplace-trusted-ca\") pod \"a2456057-5ff6-458f-af94-53edca249b7f\" (UID: \"a2456057-5ff6-458f-af94-53edca249b7f\") " Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.957745 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2sgrs\" (UniqueName: \"kubernetes.io/projected/a2456057-5ff6-458f-af94-53edca249b7f-kube-api-access-2sgrs\") pod \"a2456057-5ff6-458f-af94-53edca249b7f\" (UID: \"a2456057-5ff6-458f-af94-53edca249b7f\") " Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.958860 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2456057-5ff6-458f-af94-53edca249b7f-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "a2456057-5ff6-458f-af94-53edca249b7f" (UID: "a2456057-5ff6-458f-af94-53edca249b7f"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.964013 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2456057-5ff6-458f-af94-53edca249b7f-kube-api-access-2sgrs" (OuterVolumeSpecName: "kube-api-access-2sgrs") pod "a2456057-5ff6-458f-af94-53edca249b7f" (UID: "a2456057-5ff6-458f-af94-53edca249b7f"). InnerVolumeSpecName "kube-api-access-2sgrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.964573 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2456057-5ff6-458f-af94-53edca249b7f-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "a2456057-5ff6-458f-af94-53edca249b7f" (UID: "a2456057-5ff6-458f-af94-53edca249b7f"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.986777 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9n22\" (UniqueName: \"kubernetes.io/projected/bbeed1b2-6abb-49df-94d9-7d9fe6e5a563-kube-api-access-x9n22\") on node \"crc\" DevicePath \"\"" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.986816 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/044d9d2a-6b99-49ee-b00d-728c7832815c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.986826 4950 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a2456057-5ff6-458f-af94-53edca249b7f-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.986835 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slf9m\" (UniqueName: \"kubernetes.io/projected/044d9d2a-6b99-49ee-b00d-728c7832815c-kube-api-access-slf9m\") on node \"crc\" DevicePath \"\"" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.986845 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbeed1b2-6abb-49df-94d9-7d9fe6e5a563-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.986854 4950 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a2456057-5ff6-458f-af94-53edca249b7f-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.986862 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/044d9d2a-6b99-49ee-b00d-728c7832815c-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.986871 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2sgrs\" (UniqueName: \"kubernetes.io/projected/a2456057-5ff6-458f-af94-53edca249b7f-kube-api-access-2sgrs\") on node \"crc\" DevicePath \"\"" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.986880 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a386d81-3e2c-4df7-988d-5fb1758be616-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:13:27 crc kubenswrapper[4950]: I0318 20:13:27.986887 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbeed1b2-6abb-49df-94d9-7d9fe6e5a563-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.087711 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33a99c2b-09d6-4ee7-b875-509ef5e144eb-catalog-content\") pod \"33a99c2b-09d6-4ee7-b875-509ef5e144eb\" (UID: \"33a99c2b-09d6-4ee7-b875-509ef5e144eb\") " Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.087774 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33a99c2b-09d6-4ee7-b875-509ef5e144eb-utilities\") pod \"33a99c2b-09d6-4ee7-b875-509ef5e144eb\" (UID: \"33a99c2b-09d6-4ee7-b875-509ef5e144eb\") " Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.087832 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7zmg\" (UniqueName: \"kubernetes.io/projected/33a99c2b-09d6-4ee7-b875-509ef5e144eb-kube-api-access-k7zmg\") pod \"33a99c2b-09d6-4ee7-b875-509ef5e144eb\" (UID: \"33a99c2b-09d6-4ee7-b875-509ef5e144eb\") " Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.088567 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33a99c2b-09d6-4ee7-b875-509ef5e144eb-utilities" (OuterVolumeSpecName: "utilities") pod "33a99c2b-09d6-4ee7-b875-509ef5e144eb" (UID: "33a99c2b-09d6-4ee7-b875-509ef5e144eb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.092510 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33a99c2b-09d6-4ee7-b875-509ef5e144eb-kube-api-access-k7zmg" (OuterVolumeSpecName: "kube-api-access-k7zmg") pod "33a99c2b-09d6-4ee7-b875-509ef5e144eb" (UID: "33a99c2b-09d6-4ee7-b875-509ef5e144eb"). InnerVolumeSpecName "kube-api-access-k7zmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.171473 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c6tfb"] Mar 18 20:13:28 crc kubenswrapper[4950]: W0318 20:13:28.176178 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51b0360c_046e_4360_b240_fdfd8d857a4b.slice/crio-76c021e64cfcea7afd3b8fed92f96111ff0dafd7a26a7cad48fa3be24d23687b WatchSource:0}: Error finding container 76c021e64cfcea7afd3b8fed92f96111ff0dafd7a26a7cad48fa3be24d23687b: Status 404 returned error can't find the container with id 76c021e64cfcea7afd3b8fed92f96111ff0dafd7a26a7cad48fa3be24d23687b Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.188828 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33a99c2b-09d6-4ee7-b875-509ef5e144eb-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.188857 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7zmg\" (UniqueName: \"kubernetes.io/projected/33a99c2b-09d6-4ee7-b875-509ef5e144eb-kube-api-access-k7zmg\") on node \"crc\" DevicePath \"\"" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.217797 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33a99c2b-09d6-4ee7-b875-509ef5e144eb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "33a99c2b-09d6-4ee7-b875-509ef5e144eb" (UID: "33a99c2b-09d6-4ee7-b875-509ef5e144eb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.259837 4950 generic.go:334] "Generic (PLEG): container finished" podID="044d9d2a-6b99-49ee-b00d-728c7832815c" containerID="a9132437997afe87402a5c278ffd8f59b66a8f2a8fa686e53a6c74092124a735" exitCode=0 Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.260045 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gzj5q" event={"ID":"044d9d2a-6b99-49ee-b00d-728c7832815c","Type":"ContainerDied","Data":"a9132437997afe87402a5c278ffd8f59b66a8f2a8fa686e53a6c74092124a735"} Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.261169 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gzj5q" event={"ID":"044d9d2a-6b99-49ee-b00d-728c7832815c","Type":"ContainerDied","Data":"d7f50bce278012a7d1417f1bf553c4a3468fa213298eb3369f211c89c4dc71f0"} Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.261253 4950 scope.go:117] "RemoveContainer" containerID="a9132437997afe87402a5c278ffd8f59b66a8f2a8fa686e53a6c74092124a735" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.260123 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gzj5q" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.276227 4950 generic.go:334] "Generic (PLEG): container finished" podID="4a386d81-3e2c-4df7-988d-5fb1758be616" containerID="81f4ca959b3b2d24b3c009f7e373b147ee24bca5aaaf63ab6988946a341b5cc6" exitCode=0 Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.276328 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8kwk7" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.276365 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8kwk7" event={"ID":"4a386d81-3e2c-4df7-988d-5fb1758be616","Type":"ContainerDied","Data":"81f4ca959b3b2d24b3c009f7e373b147ee24bca5aaaf63ab6988946a341b5cc6"} Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.276444 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8kwk7" event={"ID":"4a386d81-3e2c-4df7-988d-5fb1758be616","Type":"ContainerDied","Data":"c1062526df43c854543930497f0effd3048ed2c3014191448f91af0d4dd5f4a5"} Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.280846 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-c6tfb" event={"ID":"51b0360c-046e-4360-b240-fdfd8d857a4b","Type":"ContainerStarted","Data":"76c021e64cfcea7afd3b8fed92f96111ff0dafd7a26a7cad48fa3be24d23687b"} Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.287222 4950 generic.go:334] "Generic (PLEG): container finished" podID="a2456057-5ff6-458f-af94-53edca249b7f" containerID="7164a8254b584d66dd328ab9294f1df8785a45c6f9bdd3507b1d426b574815f0" exitCode=0 Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.287319 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" event={"ID":"a2456057-5ff6-458f-af94-53edca249b7f","Type":"ContainerDied","Data":"7164a8254b584d66dd328ab9294f1df8785a45c6f9bdd3507b1d426b574815f0"} Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.287360 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" event={"ID":"a2456057-5ff6-458f-af94-53edca249b7f","Type":"ContainerDied","Data":"f923a6528bbab2ec46d1fd4396ff4afd6429d0bf4de07c2b95d05fe8f0abf876"} Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.287447 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8nhwt" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.289864 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33a99c2b-09d6-4ee7-b875-509ef5e144eb-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.301590 4950 scope.go:117] "RemoveContainer" containerID="e2be46f0ac137f088abf60e3ad3627ab19697959bf03f1e4948fe205b6b7016c" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.315860 4950 generic.go:334] "Generic (PLEG): container finished" podID="33a99c2b-09d6-4ee7-b875-509ef5e144eb" containerID="866301b46a0fa42f5ddd84e80734a99c7495f2e22720901af0847f2ada691dcb" exitCode=0 Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.316024 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nb7fx" event={"ID":"33a99c2b-09d6-4ee7-b875-509ef5e144eb","Type":"ContainerDied","Data":"866301b46a0fa42f5ddd84e80734a99c7495f2e22720901af0847f2ada691dcb"} Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.316119 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nb7fx" event={"ID":"33a99c2b-09d6-4ee7-b875-509ef5e144eb","Type":"ContainerDied","Data":"2afa4c113494611ccc861e335d5e09baf952c0b3c4f5f984e300a8f5521a735b"} Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.316313 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nb7fx" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.318491 4950 generic.go:334] "Generic (PLEG): container finished" podID="bbeed1b2-6abb-49df-94d9-7d9fe6e5a563" containerID="1fd9ccb388e588608b540567d2013892106bacdc049f5bcd80b2066229e558f3" exitCode=0 Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.318514 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7dzz6" event={"ID":"bbeed1b2-6abb-49df-94d9-7d9fe6e5a563","Type":"ContainerDied","Data":"1fd9ccb388e588608b540567d2013892106bacdc049f5bcd80b2066229e558f3"} Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.318529 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7dzz6" event={"ID":"bbeed1b2-6abb-49df-94d9-7d9fe6e5a563","Type":"ContainerDied","Data":"26e4af3d58e72ec29192677ecc2533933ed679944d6e39fcec71ae83e8109102"} Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.318594 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7dzz6" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.328816 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gzj5q"] Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.332157 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gzj5q"] Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.334111 4950 scope.go:117] "RemoveContainer" containerID="4cceb3b27880b43731e742f438b4a81d0515623897cd0916352fcfac501a9f70" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.336388 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8nhwt"] Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.341676 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8nhwt"] Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.344493 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8kwk7"] Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.363759 4950 scope.go:117] "RemoveContainer" containerID="a9132437997afe87402a5c278ffd8f59b66a8f2a8fa686e53a6c74092124a735" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.366439 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9132437997afe87402a5c278ffd8f59b66a8f2a8fa686e53a6c74092124a735\": container with ID starting with a9132437997afe87402a5c278ffd8f59b66a8f2a8fa686e53a6c74092124a735 not found: ID does not exist" containerID="a9132437997afe87402a5c278ffd8f59b66a8f2a8fa686e53a6c74092124a735" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.366511 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9132437997afe87402a5c278ffd8f59b66a8f2a8fa686e53a6c74092124a735"} err="failed to get container status \"a9132437997afe87402a5c278ffd8f59b66a8f2a8fa686e53a6c74092124a735\": rpc error: code = NotFound desc = could not find container \"a9132437997afe87402a5c278ffd8f59b66a8f2a8fa686e53a6c74092124a735\": container with ID starting with a9132437997afe87402a5c278ffd8f59b66a8f2a8fa686e53a6c74092124a735 not found: ID does not exist" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.377992 4950 scope.go:117] "RemoveContainer" containerID="e2be46f0ac137f088abf60e3ad3627ab19697959bf03f1e4948fe205b6b7016c" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.373156 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8kwk7"] Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.379752 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2be46f0ac137f088abf60e3ad3627ab19697959bf03f1e4948fe205b6b7016c\": container with ID starting with e2be46f0ac137f088abf60e3ad3627ab19697959bf03f1e4948fe205b6b7016c not found: ID does not exist" containerID="e2be46f0ac137f088abf60e3ad3627ab19697959bf03f1e4948fe205b6b7016c" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.379957 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2be46f0ac137f088abf60e3ad3627ab19697959bf03f1e4948fe205b6b7016c"} err="failed to get container status \"e2be46f0ac137f088abf60e3ad3627ab19697959bf03f1e4948fe205b6b7016c\": rpc error: code = NotFound desc = could not find container \"e2be46f0ac137f088abf60e3ad3627ab19697959bf03f1e4948fe205b6b7016c\": container with ID starting with e2be46f0ac137f088abf60e3ad3627ab19697959bf03f1e4948fe205b6b7016c not found: ID does not exist" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.379992 4950 scope.go:117] "RemoveContainer" containerID="4cceb3b27880b43731e742f438b4a81d0515623897cd0916352fcfac501a9f70" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.381283 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cceb3b27880b43731e742f438b4a81d0515623897cd0916352fcfac501a9f70\": container with ID starting with 4cceb3b27880b43731e742f438b4a81d0515623897cd0916352fcfac501a9f70 not found: ID does not exist" containerID="4cceb3b27880b43731e742f438b4a81d0515623897cd0916352fcfac501a9f70" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.381311 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cceb3b27880b43731e742f438b4a81d0515623897cd0916352fcfac501a9f70"} err="failed to get container status \"4cceb3b27880b43731e742f438b4a81d0515623897cd0916352fcfac501a9f70\": rpc error: code = NotFound desc = could not find container \"4cceb3b27880b43731e742f438b4a81d0515623897cd0916352fcfac501a9f70\": container with ID starting with 4cceb3b27880b43731e742f438b4a81d0515623897cd0916352fcfac501a9f70 not found: ID does not exist" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.381325 4950 scope.go:117] "RemoveContainer" containerID="81f4ca959b3b2d24b3c009f7e373b147ee24bca5aaaf63ab6988946a341b5cc6" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.394756 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7dzz6"] Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.401838 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7dzz6"] Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.405186 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nb7fx"] Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.409094 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nb7fx"] Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.428742 4950 scope.go:117] "RemoveContainer" containerID="b2559e3af7e229f9289ed9781209cece46a38140f4695a205450c808f8116617" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.441262 4950 scope.go:117] "RemoveContainer" containerID="83b4b8d10e11d03ddade158312b75de67ffc876388cad834fc3fcdff654df10c" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.459671 4950 scope.go:117] "RemoveContainer" containerID="81f4ca959b3b2d24b3c009f7e373b147ee24bca5aaaf63ab6988946a341b5cc6" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.460111 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81f4ca959b3b2d24b3c009f7e373b147ee24bca5aaaf63ab6988946a341b5cc6\": container with ID starting with 81f4ca959b3b2d24b3c009f7e373b147ee24bca5aaaf63ab6988946a341b5cc6 not found: ID does not exist" containerID="81f4ca959b3b2d24b3c009f7e373b147ee24bca5aaaf63ab6988946a341b5cc6" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.460159 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81f4ca959b3b2d24b3c009f7e373b147ee24bca5aaaf63ab6988946a341b5cc6"} err="failed to get container status \"81f4ca959b3b2d24b3c009f7e373b147ee24bca5aaaf63ab6988946a341b5cc6\": rpc error: code = NotFound desc = could not find container \"81f4ca959b3b2d24b3c009f7e373b147ee24bca5aaaf63ab6988946a341b5cc6\": container with ID starting with 81f4ca959b3b2d24b3c009f7e373b147ee24bca5aaaf63ab6988946a341b5cc6 not found: ID does not exist" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.460182 4950 scope.go:117] "RemoveContainer" containerID="b2559e3af7e229f9289ed9781209cece46a38140f4695a205450c808f8116617" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.460773 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2559e3af7e229f9289ed9781209cece46a38140f4695a205450c808f8116617\": container with ID starting with b2559e3af7e229f9289ed9781209cece46a38140f4695a205450c808f8116617 not found: ID does not exist" containerID="b2559e3af7e229f9289ed9781209cece46a38140f4695a205450c808f8116617" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.460813 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2559e3af7e229f9289ed9781209cece46a38140f4695a205450c808f8116617"} err="failed to get container status \"b2559e3af7e229f9289ed9781209cece46a38140f4695a205450c808f8116617\": rpc error: code = NotFound desc = could not find container \"b2559e3af7e229f9289ed9781209cece46a38140f4695a205450c808f8116617\": container with ID starting with b2559e3af7e229f9289ed9781209cece46a38140f4695a205450c808f8116617 not found: ID does not exist" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.460826 4950 scope.go:117] "RemoveContainer" containerID="83b4b8d10e11d03ddade158312b75de67ffc876388cad834fc3fcdff654df10c" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.465458 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83b4b8d10e11d03ddade158312b75de67ffc876388cad834fc3fcdff654df10c\": container with ID starting with 83b4b8d10e11d03ddade158312b75de67ffc876388cad834fc3fcdff654df10c not found: ID does not exist" containerID="83b4b8d10e11d03ddade158312b75de67ffc876388cad834fc3fcdff654df10c" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.465528 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83b4b8d10e11d03ddade158312b75de67ffc876388cad834fc3fcdff654df10c"} err="failed to get container status \"83b4b8d10e11d03ddade158312b75de67ffc876388cad834fc3fcdff654df10c\": rpc error: code = NotFound desc = could not find container \"83b4b8d10e11d03ddade158312b75de67ffc876388cad834fc3fcdff654df10c\": container with ID starting with 83b4b8d10e11d03ddade158312b75de67ffc876388cad834fc3fcdff654df10c not found: ID does not exist" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.465545 4950 scope.go:117] "RemoveContainer" containerID="7164a8254b584d66dd328ab9294f1df8785a45c6f9bdd3507b1d426b574815f0" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.484277 4950 scope.go:117] "RemoveContainer" containerID="bb43772a7f0be4bd5df4460e9a3438e9991f1cee7c7ca14fcaf9f7e315f4121e" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.500627 4950 scope.go:117] "RemoveContainer" containerID="7164a8254b584d66dd328ab9294f1df8785a45c6f9bdd3507b1d426b574815f0" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.501147 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7164a8254b584d66dd328ab9294f1df8785a45c6f9bdd3507b1d426b574815f0\": container with ID starting with 7164a8254b584d66dd328ab9294f1df8785a45c6f9bdd3507b1d426b574815f0 not found: ID does not exist" containerID="7164a8254b584d66dd328ab9294f1df8785a45c6f9bdd3507b1d426b574815f0" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.501195 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7164a8254b584d66dd328ab9294f1df8785a45c6f9bdd3507b1d426b574815f0"} err="failed to get container status \"7164a8254b584d66dd328ab9294f1df8785a45c6f9bdd3507b1d426b574815f0\": rpc error: code = NotFound desc = could not find container \"7164a8254b584d66dd328ab9294f1df8785a45c6f9bdd3507b1d426b574815f0\": container with ID starting with 7164a8254b584d66dd328ab9294f1df8785a45c6f9bdd3507b1d426b574815f0 not found: ID does not exist" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.501219 4950 scope.go:117] "RemoveContainer" containerID="bb43772a7f0be4bd5df4460e9a3438e9991f1cee7c7ca14fcaf9f7e315f4121e" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.501702 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb43772a7f0be4bd5df4460e9a3438e9991f1cee7c7ca14fcaf9f7e315f4121e\": container with ID starting with bb43772a7f0be4bd5df4460e9a3438e9991f1cee7c7ca14fcaf9f7e315f4121e not found: ID does not exist" containerID="bb43772a7f0be4bd5df4460e9a3438e9991f1cee7c7ca14fcaf9f7e315f4121e" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.501724 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb43772a7f0be4bd5df4460e9a3438e9991f1cee7c7ca14fcaf9f7e315f4121e"} err="failed to get container status \"bb43772a7f0be4bd5df4460e9a3438e9991f1cee7c7ca14fcaf9f7e315f4121e\": rpc error: code = NotFound desc = could not find container \"bb43772a7f0be4bd5df4460e9a3438e9991f1cee7c7ca14fcaf9f7e315f4121e\": container with ID starting with bb43772a7f0be4bd5df4460e9a3438e9991f1cee7c7ca14fcaf9f7e315f4121e not found: ID does not exist" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.501766 4950 scope.go:117] "RemoveContainer" containerID="866301b46a0fa42f5ddd84e80734a99c7495f2e22720901af0847f2ada691dcb" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.523062 4950 scope.go:117] "RemoveContainer" containerID="786e809c5bbec43680ae9e1cb62efe0945f0c7a82d0c1f4e42c481597f4b2820" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.539998 4950 scope.go:117] "RemoveContainer" containerID="14fc2e56f2022499c8bde7db5aa8430bf11cae9929d29b3e716e2ca6a0a8217c" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.582688 4950 scope.go:117] "RemoveContainer" containerID="866301b46a0fa42f5ddd84e80734a99c7495f2e22720901af0847f2ada691dcb" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.583373 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"866301b46a0fa42f5ddd84e80734a99c7495f2e22720901af0847f2ada691dcb\": container with ID starting with 866301b46a0fa42f5ddd84e80734a99c7495f2e22720901af0847f2ada691dcb not found: ID does not exist" containerID="866301b46a0fa42f5ddd84e80734a99c7495f2e22720901af0847f2ada691dcb" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.583495 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"866301b46a0fa42f5ddd84e80734a99c7495f2e22720901af0847f2ada691dcb"} err="failed to get container status \"866301b46a0fa42f5ddd84e80734a99c7495f2e22720901af0847f2ada691dcb\": rpc error: code = NotFound desc = could not find container \"866301b46a0fa42f5ddd84e80734a99c7495f2e22720901af0847f2ada691dcb\": container with ID starting with 866301b46a0fa42f5ddd84e80734a99c7495f2e22720901af0847f2ada691dcb not found: ID does not exist" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.583529 4950 scope.go:117] "RemoveContainer" containerID="786e809c5bbec43680ae9e1cb62efe0945f0c7a82d0c1f4e42c481597f4b2820" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.584221 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"786e809c5bbec43680ae9e1cb62efe0945f0c7a82d0c1f4e42c481597f4b2820\": container with ID starting with 786e809c5bbec43680ae9e1cb62efe0945f0c7a82d0c1f4e42c481597f4b2820 not found: ID does not exist" containerID="786e809c5bbec43680ae9e1cb62efe0945f0c7a82d0c1f4e42c481597f4b2820" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.584329 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"786e809c5bbec43680ae9e1cb62efe0945f0c7a82d0c1f4e42c481597f4b2820"} err="failed to get container status \"786e809c5bbec43680ae9e1cb62efe0945f0c7a82d0c1f4e42c481597f4b2820\": rpc error: code = NotFound desc = could not find container \"786e809c5bbec43680ae9e1cb62efe0945f0c7a82d0c1f4e42c481597f4b2820\": container with ID starting with 786e809c5bbec43680ae9e1cb62efe0945f0c7a82d0c1f4e42c481597f4b2820 not found: ID does not exist" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.584577 4950 scope.go:117] "RemoveContainer" containerID="14fc2e56f2022499c8bde7db5aa8430bf11cae9929d29b3e716e2ca6a0a8217c" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.584999 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14fc2e56f2022499c8bde7db5aa8430bf11cae9929d29b3e716e2ca6a0a8217c\": container with ID starting with 14fc2e56f2022499c8bde7db5aa8430bf11cae9929d29b3e716e2ca6a0a8217c not found: ID does not exist" containerID="14fc2e56f2022499c8bde7db5aa8430bf11cae9929d29b3e716e2ca6a0a8217c" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.585085 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14fc2e56f2022499c8bde7db5aa8430bf11cae9929d29b3e716e2ca6a0a8217c"} err="failed to get container status \"14fc2e56f2022499c8bde7db5aa8430bf11cae9929d29b3e716e2ca6a0a8217c\": rpc error: code = NotFound desc = could not find container \"14fc2e56f2022499c8bde7db5aa8430bf11cae9929d29b3e716e2ca6a0a8217c\": container with ID starting with 14fc2e56f2022499c8bde7db5aa8430bf11cae9929d29b3e716e2ca6a0a8217c not found: ID does not exist" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.585151 4950 scope.go:117] "RemoveContainer" containerID="1fd9ccb388e588608b540567d2013892106bacdc049f5bcd80b2066229e558f3" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.603489 4950 scope.go:117] "RemoveContainer" containerID="633af43d8823abb96bf78609a47d9f3ff3e8c5ce993efdca967ceb91288247a5" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.618777 4950 scope.go:117] "RemoveContainer" containerID="e73af9bce8b03e6b8ea69a8eb5cba1504a5c451c568cb4b6135277a0623070f3" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.629838 4950 scope.go:117] "RemoveContainer" containerID="1fd9ccb388e588608b540567d2013892106bacdc049f5bcd80b2066229e558f3" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.630272 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fd9ccb388e588608b540567d2013892106bacdc049f5bcd80b2066229e558f3\": container with ID starting with 1fd9ccb388e588608b540567d2013892106bacdc049f5bcd80b2066229e558f3 not found: ID does not exist" containerID="1fd9ccb388e588608b540567d2013892106bacdc049f5bcd80b2066229e558f3" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.630299 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fd9ccb388e588608b540567d2013892106bacdc049f5bcd80b2066229e558f3"} err="failed to get container status \"1fd9ccb388e588608b540567d2013892106bacdc049f5bcd80b2066229e558f3\": rpc error: code = NotFound desc = could not find container \"1fd9ccb388e588608b540567d2013892106bacdc049f5bcd80b2066229e558f3\": container with ID starting with 1fd9ccb388e588608b540567d2013892106bacdc049f5bcd80b2066229e558f3 not found: ID does not exist" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.630319 4950 scope.go:117] "RemoveContainer" containerID="633af43d8823abb96bf78609a47d9f3ff3e8c5ce993efdca967ceb91288247a5" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.630602 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"633af43d8823abb96bf78609a47d9f3ff3e8c5ce993efdca967ceb91288247a5\": container with ID starting with 633af43d8823abb96bf78609a47d9f3ff3e8c5ce993efdca967ceb91288247a5 not found: ID does not exist" containerID="633af43d8823abb96bf78609a47d9f3ff3e8c5ce993efdca967ceb91288247a5" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.630705 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"633af43d8823abb96bf78609a47d9f3ff3e8c5ce993efdca967ceb91288247a5"} err="failed to get container status \"633af43d8823abb96bf78609a47d9f3ff3e8c5ce993efdca967ceb91288247a5\": rpc error: code = NotFound desc = could not find container \"633af43d8823abb96bf78609a47d9f3ff3e8c5ce993efdca967ceb91288247a5\": container with ID starting with 633af43d8823abb96bf78609a47d9f3ff3e8c5ce993efdca967ceb91288247a5 not found: ID does not exist" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.630797 4950 scope.go:117] "RemoveContainer" containerID="e73af9bce8b03e6b8ea69a8eb5cba1504a5c451c568cb4b6135277a0623070f3" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.631176 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e73af9bce8b03e6b8ea69a8eb5cba1504a5c451c568cb4b6135277a0623070f3\": container with ID starting with e73af9bce8b03e6b8ea69a8eb5cba1504a5c451c568cb4b6135277a0623070f3 not found: ID does not exist" containerID="e73af9bce8b03e6b8ea69a8eb5cba1504a5c451c568cb4b6135277a0623070f3" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.631209 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e73af9bce8b03e6b8ea69a8eb5cba1504a5c451c568cb4b6135277a0623070f3"} err="failed to get container status \"e73af9bce8b03e6b8ea69a8eb5cba1504a5c451c568cb4b6135277a0623070f3\": rpc error: code = NotFound desc = could not find container \"e73af9bce8b03e6b8ea69a8eb5cba1504a5c451c568cb4b6135277a0623070f3\": container with ID starting with e73af9bce8b03e6b8ea69a8eb5cba1504a5c451c568cb4b6135277a0623070f3 not found: ID does not exist" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.823591 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xv6kx"] Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.823802 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33a99c2b-09d6-4ee7-b875-509ef5e144eb" containerName="extract-content" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.823815 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="33a99c2b-09d6-4ee7-b875-509ef5e144eb" containerName="extract-content" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.823827 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2456057-5ff6-458f-af94-53edca249b7f" containerName="marketplace-operator" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.823835 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2456057-5ff6-458f-af94-53edca249b7f" containerName="marketplace-operator" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.823846 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a386d81-3e2c-4df7-988d-5fb1758be616" containerName="extract-content" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.823854 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a386d81-3e2c-4df7-988d-5fb1758be616" containerName="extract-content" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.823864 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="044d9d2a-6b99-49ee-b00d-728c7832815c" containerName="registry-server" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.823871 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="044d9d2a-6b99-49ee-b00d-728c7832815c" containerName="registry-server" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.823880 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a386d81-3e2c-4df7-988d-5fb1758be616" containerName="registry-server" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.823888 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a386d81-3e2c-4df7-988d-5fb1758be616" containerName="registry-server" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.823897 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2456057-5ff6-458f-af94-53edca249b7f" containerName="marketplace-operator" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.823905 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2456057-5ff6-458f-af94-53edca249b7f" containerName="marketplace-operator" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.823916 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbeed1b2-6abb-49df-94d9-7d9fe6e5a563" containerName="registry-server" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.823923 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbeed1b2-6abb-49df-94d9-7d9fe6e5a563" containerName="registry-server" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.823932 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a386d81-3e2c-4df7-988d-5fb1758be616" containerName="extract-utilities" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.823938 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a386d81-3e2c-4df7-988d-5fb1758be616" containerName="extract-utilities" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.823947 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33a99c2b-09d6-4ee7-b875-509ef5e144eb" containerName="registry-server" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.823955 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="33a99c2b-09d6-4ee7-b875-509ef5e144eb" containerName="registry-server" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.823968 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbeed1b2-6abb-49df-94d9-7d9fe6e5a563" containerName="extract-content" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.823975 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbeed1b2-6abb-49df-94d9-7d9fe6e5a563" containerName="extract-content" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.823986 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="044d9d2a-6b99-49ee-b00d-728c7832815c" containerName="extract-content" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.823993 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="044d9d2a-6b99-49ee-b00d-728c7832815c" containerName="extract-content" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.824003 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbeed1b2-6abb-49df-94d9-7d9fe6e5a563" containerName="extract-utilities" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.824010 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbeed1b2-6abb-49df-94d9-7d9fe6e5a563" containerName="extract-utilities" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.824022 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33a99c2b-09d6-4ee7-b875-509ef5e144eb" containerName="extract-utilities" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.824029 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="33a99c2b-09d6-4ee7-b875-509ef5e144eb" containerName="extract-utilities" Mar 18 20:13:28 crc kubenswrapper[4950]: E0318 20:13:28.824038 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="044d9d2a-6b99-49ee-b00d-728c7832815c" containerName="extract-utilities" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.824046 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="044d9d2a-6b99-49ee-b00d-728c7832815c" containerName="extract-utilities" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.824150 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="044d9d2a-6b99-49ee-b00d-728c7832815c" containerName="registry-server" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.824162 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2456057-5ff6-458f-af94-53edca249b7f" containerName="marketplace-operator" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.824173 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2456057-5ff6-458f-af94-53edca249b7f" containerName="marketplace-operator" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.824185 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbeed1b2-6abb-49df-94d9-7d9fe6e5a563" containerName="registry-server" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.824196 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="33a99c2b-09d6-4ee7-b875-509ef5e144eb" containerName="registry-server" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.824210 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a386d81-3e2c-4df7-988d-5fb1758be616" containerName="registry-server" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.824904 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xv6kx" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.827511 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.843174 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xv6kx"] Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.900459 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df5f1fd5-ec77-45be-84de-d0d8a1568a73-utilities\") pod \"certified-operators-xv6kx\" (UID: \"df5f1fd5-ec77-45be-84de-d0d8a1568a73\") " pod="openshift-marketplace/certified-operators-xv6kx" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.900504 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzm4x\" (UniqueName: \"kubernetes.io/projected/df5f1fd5-ec77-45be-84de-d0d8a1568a73-kube-api-access-wzm4x\") pod \"certified-operators-xv6kx\" (UID: \"df5f1fd5-ec77-45be-84de-d0d8a1568a73\") " pod="openshift-marketplace/certified-operators-xv6kx" Mar 18 20:13:28 crc kubenswrapper[4950]: I0318 20:13:28.900543 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df5f1fd5-ec77-45be-84de-d0d8a1568a73-catalog-content\") pod \"certified-operators-xv6kx\" (UID: \"df5f1fd5-ec77-45be-84de-d0d8a1568a73\") " pod="openshift-marketplace/certified-operators-xv6kx" Mar 18 20:13:29 crc kubenswrapper[4950]: I0318 20:13:29.001892 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df5f1fd5-ec77-45be-84de-d0d8a1568a73-utilities\") pod \"certified-operators-xv6kx\" (UID: \"df5f1fd5-ec77-45be-84de-d0d8a1568a73\") " pod="openshift-marketplace/certified-operators-xv6kx" Mar 18 20:13:29 crc kubenswrapper[4950]: I0318 20:13:29.002219 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzm4x\" (UniqueName: \"kubernetes.io/projected/df5f1fd5-ec77-45be-84de-d0d8a1568a73-kube-api-access-wzm4x\") pod \"certified-operators-xv6kx\" (UID: \"df5f1fd5-ec77-45be-84de-d0d8a1568a73\") " pod="openshift-marketplace/certified-operators-xv6kx" Mar 18 20:13:29 crc kubenswrapper[4950]: I0318 20:13:29.002365 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df5f1fd5-ec77-45be-84de-d0d8a1568a73-catalog-content\") pod \"certified-operators-xv6kx\" (UID: \"df5f1fd5-ec77-45be-84de-d0d8a1568a73\") " pod="openshift-marketplace/certified-operators-xv6kx" Mar 18 20:13:29 crc kubenswrapper[4950]: I0318 20:13:29.002497 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df5f1fd5-ec77-45be-84de-d0d8a1568a73-utilities\") pod \"certified-operators-xv6kx\" (UID: \"df5f1fd5-ec77-45be-84de-d0d8a1568a73\") " pod="openshift-marketplace/certified-operators-xv6kx" Mar 18 20:13:29 crc kubenswrapper[4950]: I0318 20:13:29.002796 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df5f1fd5-ec77-45be-84de-d0d8a1568a73-catalog-content\") pod \"certified-operators-xv6kx\" (UID: \"df5f1fd5-ec77-45be-84de-d0d8a1568a73\") " pod="openshift-marketplace/certified-operators-xv6kx" Mar 18 20:13:29 crc kubenswrapper[4950]: I0318 20:13:29.027270 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzm4x\" (UniqueName: \"kubernetes.io/projected/df5f1fd5-ec77-45be-84de-d0d8a1568a73-kube-api-access-wzm4x\") pod \"certified-operators-xv6kx\" (UID: \"df5f1fd5-ec77-45be-84de-d0d8a1568a73\") " pod="openshift-marketplace/certified-operators-xv6kx" Mar 18 20:13:29 crc kubenswrapper[4950]: I0318 20:13:29.138542 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xv6kx" Mar 18 20:13:29 crc kubenswrapper[4950]: I0318 20:13:29.330621 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xv6kx"] Mar 18 20:13:29 crc kubenswrapper[4950]: I0318 20:13:29.338066 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-c6tfb" event={"ID":"51b0360c-046e-4360-b240-fdfd8d857a4b","Type":"ContainerStarted","Data":"1c3d9aadf3832e111ba871bb52977a0e608bbcb62114e4e95e6a30ea88627385"} Mar 18 20:13:29 crc kubenswrapper[4950]: I0318 20:13:29.338343 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-c6tfb" Mar 18 20:13:29 crc kubenswrapper[4950]: I0318 20:13:29.343766 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-c6tfb" Mar 18 20:13:29 crc kubenswrapper[4950]: I0318 20:13:29.355801 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-c6tfb" podStartSLOduration=2.355773132 podStartE2EDuration="2.355773132s" podCreationTimestamp="2026-03-18 20:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:13:29.355277308 +0000 UTC m=+422.596119196" watchObservedRunningTime="2026-03-18 20:13:29.355773132 +0000 UTC m=+422.596615000" Mar 18 20:13:29 crc kubenswrapper[4950]: I0318 20:13:29.487095 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="044d9d2a-6b99-49ee-b00d-728c7832815c" path="/var/lib/kubelet/pods/044d9d2a-6b99-49ee-b00d-728c7832815c/volumes" Mar 18 20:13:29 crc kubenswrapper[4950]: I0318 20:13:29.488187 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33a99c2b-09d6-4ee7-b875-509ef5e144eb" path="/var/lib/kubelet/pods/33a99c2b-09d6-4ee7-b875-509ef5e144eb/volumes" Mar 18 20:13:29 crc kubenswrapper[4950]: I0318 20:13:29.488911 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a386d81-3e2c-4df7-988d-5fb1758be616" path="/var/lib/kubelet/pods/4a386d81-3e2c-4df7-988d-5fb1758be616/volumes" Mar 18 20:13:29 crc kubenswrapper[4950]: I0318 20:13:29.490381 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2456057-5ff6-458f-af94-53edca249b7f" path="/var/lib/kubelet/pods/a2456057-5ff6-458f-af94-53edca249b7f/volumes" Mar 18 20:13:29 crc kubenswrapper[4950]: I0318 20:13:29.490976 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbeed1b2-6abb-49df-94d9-7d9fe6e5a563" path="/var/lib/kubelet/pods/bbeed1b2-6abb-49df-94d9-7d9fe6e5a563/volumes" Mar 18 20:13:30 crc kubenswrapper[4950]: I0318 20:13:30.227605 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-shgnx"] Mar 18 20:13:30 crc kubenswrapper[4950]: I0318 20:13:30.229858 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-shgnx" Mar 18 20:13:30 crc kubenswrapper[4950]: I0318 20:13:30.232341 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 18 20:13:30 crc kubenswrapper[4950]: I0318 20:13:30.240452 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-shgnx"] Mar 18 20:13:30 crc kubenswrapper[4950]: I0318 20:13:30.329049 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pzdh\" (UniqueName: \"kubernetes.io/projected/19ec358c-0c71-4b43-b13f-688f2103d895-kube-api-access-2pzdh\") pod \"redhat-marketplace-shgnx\" (UID: \"19ec358c-0c71-4b43-b13f-688f2103d895\") " pod="openshift-marketplace/redhat-marketplace-shgnx" Mar 18 20:13:30 crc kubenswrapper[4950]: I0318 20:13:30.329106 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19ec358c-0c71-4b43-b13f-688f2103d895-utilities\") pod \"redhat-marketplace-shgnx\" (UID: \"19ec358c-0c71-4b43-b13f-688f2103d895\") " pod="openshift-marketplace/redhat-marketplace-shgnx" Mar 18 20:13:30 crc kubenswrapper[4950]: I0318 20:13:30.329171 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19ec358c-0c71-4b43-b13f-688f2103d895-catalog-content\") pod \"redhat-marketplace-shgnx\" (UID: \"19ec358c-0c71-4b43-b13f-688f2103d895\") " pod="openshift-marketplace/redhat-marketplace-shgnx" Mar 18 20:13:30 crc kubenswrapper[4950]: I0318 20:13:30.353981 4950 generic.go:334] "Generic (PLEG): container finished" podID="df5f1fd5-ec77-45be-84de-d0d8a1568a73" containerID="71b8d30ca050d368d8f440092748638364e2f3c8b229f90a8f690c4af5eae848" exitCode=0 Mar 18 20:13:30 crc kubenswrapper[4950]: I0318 20:13:30.354047 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xv6kx" event={"ID":"df5f1fd5-ec77-45be-84de-d0d8a1568a73","Type":"ContainerDied","Data":"71b8d30ca050d368d8f440092748638364e2f3c8b229f90a8f690c4af5eae848"} Mar 18 20:13:30 crc kubenswrapper[4950]: I0318 20:13:30.354113 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xv6kx" event={"ID":"df5f1fd5-ec77-45be-84de-d0d8a1568a73","Type":"ContainerStarted","Data":"4d212e08086a76c089a2d806116877eb6a0b5117da038b4cd53134e16d662b67"} Mar 18 20:13:30 crc kubenswrapper[4950]: I0318 20:13:30.430343 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19ec358c-0c71-4b43-b13f-688f2103d895-catalog-content\") pod \"redhat-marketplace-shgnx\" (UID: \"19ec358c-0c71-4b43-b13f-688f2103d895\") " pod="openshift-marketplace/redhat-marketplace-shgnx" Mar 18 20:13:30 crc kubenswrapper[4950]: I0318 20:13:30.430394 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pzdh\" (UniqueName: \"kubernetes.io/projected/19ec358c-0c71-4b43-b13f-688f2103d895-kube-api-access-2pzdh\") pod \"redhat-marketplace-shgnx\" (UID: \"19ec358c-0c71-4b43-b13f-688f2103d895\") " pod="openshift-marketplace/redhat-marketplace-shgnx" Mar 18 20:13:30 crc kubenswrapper[4950]: I0318 20:13:30.430455 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19ec358c-0c71-4b43-b13f-688f2103d895-utilities\") pod \"redhat-marketplace-shgnx\" (UID: \"19ec358c-0c71-4b43-b13f-688f2103d895\") " pod="openshift-marketplace/redhat-marketplace-shgnx" Mar 18 20:13:30 crc kubenswrapper[4950]: I0318 20:13:30.431133 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19ec358c-0c71-4b43-b13f-688f2103d895-utilities\") pod \"redhat-marketplace-shgnx\" (UID: \"19ec358c-0c71-4b43-b13f-688f2103d895\") " pod="openshift-marketplace/redhat-marketplace-shgnx" Mar 18 20:13:30 crc kubenswrapper[4950]: I0318 20:13:30.431135 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19ec358c-0c71-4b43-b13f-688f2103d895-catalog-content\") pod \"redhat-marketplace-shgnx\" (UID: \"19ec358c-0c71-4b43-b13f-688f2103d895\") " pod="openshift-marketplace/redhat-marketplace-shgnx" Mar 18 20:13:30 crc kubenswrapper[4950]: I0318 20:13:30.453344 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pzdh\" (UniqueName: \"kubernetes.io/projected/19ec358c-0c71-4b43-b13f-688f2103d895-kube-api-access-2pzdh\") pod \"redhat-marketplace-shgnx\" (UID: \"19ec358c-0c71-4b43-b13f-688f2103d895\") " pod="openshift-marketplace/redhat-marketplace-shgnx" Mar 18 20:13:30 crc kubenswrapper[4950]: I0318 20:13:30.557685 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-shgnx" Mar 18 20:13:30 crc kubenswrapper[4950]: I0318 20:13:30.977211 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-shgnx"] Mar 18 20:13:30 crc kubenswrapper[4950]: W0318 20:13:30.988625 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19ec358c_0c71_4b43_b13f_688f2103d895.slice/crio-c51cfa7bb87756526644e8cd791d9c4178783ee50cf4bd4477296c2e1d908d2f WatchSource:0}: Error finding container c51cfa7bb87756526644e8cd791d9c4178783ee50cf4bd4477296c2e1d908d2f: Status 404 returned error can't find the container with id c51cfa7bb87756526644e8cd791d9c4178783ee50cf4bd4477296c2e1d908d2f Mar 18 20:13:31 crc kubenswrapper[4950]: I0318 20:13:31.235389 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ldrm5"] Mar 18 20:13:31 crc kubenswrapper[4950]: I0318 20:13:31.236565 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ldrm5" Mar 18 20:13:31 crc kubenswrapper[4950]: I0318 20:13:31.240122 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 18 20:13:31 crc kubenswrapper[4950]: I0318 20:13:31.255342 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ldrm5"] Mar 18 20:13:31 crc kubenswrapper[4950]: I0318 20:13:31.345691 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81fa390c-72e6-471b-ac97-8e5d5fc3b4d7-catalog-content\") pod \"redhat-operators-ldrm5\" (UID: \"81fa390c-72e6-471b-ac97-8e5d5fc3b4d7\") " pod="openshift-marketplace/redhat-operators-ldrm5" Mar 18 20:13:31 crc kubenswrapper[4950]: I0318 20:13:31.346075 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbllm\" (UniqueName: \"kubernetes.io/projected/81fa390c-72e6-471b-ac97-8e5d5fc3b4d7-kube-api-access-rbllm\") pod \"redhat-operators-ldrm5\" (UID: \"81fa390c-72e6-471b-ac97-8e5d5fc3b4d7\") " pod="openshift-marketplace/redhat-operators-ldrm5" Mar 18 20:13:31 crc kubenswrapper[4950]: I0318 20:13:31.346118 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81fa390c-72e6-471b-ac97-8e5d5fc3b4d7-utilities\") pod \"redhat-operators-ldrm5\" (UID: \"81fa390c-72e6-471b-ac97-8e5d5fc3b4d7\") " pod="openshift-marketplace/redhat-operators-ldrm5" Mar 18 20:13:31 crc kubenswrapper[4950]: I0318 20:13:31.387087 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xv6kx" event={"ID":"df5f1fd5-ec77-45be-84de-d0d8a1568a73","Type":"ContainerStarted","Data":"b3604b7a94c081d7230b1d801e1187cabf96229111a4be9a7e7024ba87205d16"} Mar 18 20:13:31 crc kubenswrapper[4950]: I0318 20:13:31.394698 4950 generic.go:334] "Generic (PLEG): container finished" podID="19ec358c-0c71-4b43-b13f-688f2103d895" containerID="bf47af8d70effec2680967ac8f49db06a68709c493c13112f3596bb33f7f3452" exitCode=0 Mar 18 20:13:31 crc kubenswrapper[4950]: I0318 20:13:31.394797 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-shgnx" event={"ID":"19ec358c-0c71-4b43-b13f-688f2103d895","Type":"ContainerDied","Data":"bf47af8d70effec2680967ac8f49db06a68709c493c13112f3596bb33f7f3452"} Mar 18 20:13:31 crc kubenswrapper[4950]: I0318 20:13:31.396902 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-shgnx" event={"ID":"19ec358c-0c71-4b43-b13f-688f2103d895","Type":"ContainerStarted","Data":"c51cfa7bb87756526644e8cd791d9c4178783ee50cf4bd4477296c2e1d908d2f"} Mar 18 20:13:31 crc kubenswrapper[4950]: I0318 20:13:31.446676 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81fa390c-72e6-471b-ac97-8e5d5fc3b4d7-catalog-content\") pod \"redhat-operators-ldrm5\" (UID: \"81fa390c-72e6-471b-ac97-8e5d5fc3b4d7\") " pod="openshift-marketplace/redhat-operators-ldrm5" Mar 18 20:13:31 crc kubenswrapper[4950]: I0318 20:13:31.446739 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbllm\" (UniqueName: \"kubernetes.io/projected/81fa390c-72e6-471b-ac97-8e5d5fc3b4d7-kube-api-access-rbllm\") pod \"redhat-operators-ldrm5\" (UID: \"81fa390c-72e6-471b-ac97-8e5d5fc3b4d7\") " pod="openshift-marketplace/redhat-operators-ldrm5" Mar 18 20:13:31 crc kubenswrapper[4950]: I0318 20:13:31.446771 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81fa390c-72e6-471b-ac97-8e5d5fc3b4d7-utilities\") pod \"redhat-operators-ldrm5\" (UID: \"81fa390c-72e6-471b-ac97-8e5d5fc3b4d7\") " pod="openshift-marketplace/redhat-operators-ldrm5" Mar 18 20:13:31 crc kubenswrapper[4950]: I0318 20:13:31.447232 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81fa390c-72e6-471b-ac97-8e5d5fc3b4d7-utilities\") pod \"redhat-operators-ldrm5\" (UID: \"81fa390c-72e6-471b-ac97-8e5d5fc3b4d7\") " pod="openshift-marketplace/redhat-operators-ldrm5" Mar 18 20:13:31 crc kubenswrapper[4950]: I0318 20:13:31.447479 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81fa390c-72e6-471b-ac97-8e5d5fc3b4d7-catalog-content\") pod \"redhat-operators-ldrm5\" (UID: \"81fa390c-72e6-471b-ac97-8e5d5fc3b4d7\") " pod="openshift-marketplace/redhat-operators-ldrm5" Mar 18 20:13:31 crc kubenswrapper[4950]: I0318 20:13:31.476213 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbllm\" (UniqueName: \"kubernetes.io/projected/81fa390c-72e6-471b-ac97-8e5d5fc3b4d7-kube-api-access-rbllm\") pod \"redhat-operators-ldrm5\" (UID: \"81fa390c-72e6-471b-ac97-8e5d5fc3b4d7\") " pod="openshift-marketplace/redhat-operators-ldrm5" Mar 18 20:13:31 crc kubenswrapper[4950]: I0318 20:13:31.549006 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ldrm5" Mar 18 20:13:31 crc kubenswrapper[4950]: E0318 20:13:31.580354 4950 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2456057_5ff6_458f_af94_53edca249b7f.slice/crio-7164a8254b584d66dd328ab9294f1df8785a45c6f9bdd3507b1d426b574815f0.scope\": RecentStats: unable to find data in memory cache]" Mar 18 20:13:31 crc kubenswrapper[4950]: I0318 20:13:31.985668 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ldrm5"] Mar 18 20:13:32 crc kubenswrapper[4950]: I0318 20:13:32.416078 4950 generic.go:334] "Generic (PLEG): container finished" podID="81fa390c-72e6-471b-ac97-8e5d5fc3b4d7" containerID="f3289f682a363386458f56a7c14afd5877d045a384db2ec36ea2e8fc0796af34" exitCode=0 Mar 18 20:13:32 crc kubenswrapper[4950]: I0318 20:13:32.416251 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldrm5" event={"ID":"81fa390c-72e6-471b-ac97-8e5d5fc3b4d7","Type":"ContainerDied","Data":"f3289f682a363386458f56a7c14afd5877d045a384db2ec36ea2e8fc0796af34"} Mar 18 20:13:32 crc kubenswrapper[4950]: I0318 20:13:32.416300 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldrm5" event={"ID":"81fa390c-72e6-471b-ac97-8e5d5fc3b4d7","Type":"ContainerStarted","Data":"112ea6205059f5c043f1fd27d088e60bf5cfc48ffa3a7aa1d5395821da8cda98"} Mar 18 20:13:32 crc kubenswrapper[4950]: I0318 20:13:32.422030 4950 generic.go:334] "Generic (PLEG): container finished" podID="df5f1fd5-ec77-45be-84de-d0d8a1568a73" containerID="b3604b7a94c081d7230b1d801e1187cabf96229111a4be9a7e7024ba87205d16" exitCode=0 Mar 18 20:13:32 crc kubenswrapper[4950]: I0318 20:13:32.422094 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xv6kx" event={"ID":"df5f1fd5-ec77-45be-84de-d0d8a1568a73","Type":"ContainerDied","Data":"b3604b7a94c081d7230b1d801e1187cabf96229111a4be9a7e7024ba87205d16"} Mar 18 20:13:32 crc kubenswrapper[4950]: I0318 20:13:32.632701 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zljc4"] Mar 18 20:13:32 crc kubenswrapper[4950]: I0318 20:13:32.636626 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zljc4" Mar 18 20:13:32 crc kubenswrapper[4950]: I0318 20:13:32.639011 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 18 20:13:32 crc kubenswrapper[4950]: I0318 20:13:32.646887 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zljc4"] Mar 18 20:13:32 crc kubenswrapper[4950]: I0318 20:13:32.765117 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40ff6ac2-e78c-4501-9fdd-8d612fd6d29d-catalog-content\") pod \"community-operators-zljc4\" (UID: \"40ff6ac2-e78c-4501-9fdd-8d612fd6d29d\") " pod="openshift-marketplace/community-operators-zljc4" Mar 18 20:13:32 crc kubenswrapper[4950]: I0318 20:13:32.765181 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40ff6ac2-e78c-4501-9fdd-8d612fd6d29d-utilities\") pod \"community-operators-zljc4\" (UID: \"40ff6ac2-e78c-4501-9fdd-8d612fd6d29d\") " pod="openshift-marketplace/community-operators-zljc4" Mar 18 20:13:32 crc kubenswrapper[4950]: I0318 20:13:32.765246 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw4hp\" (UniqueName: \"kubernetes.io/projected/40ff6ac2-e78c-4501-9fdd-8d612fd6d29d-kube-api-access-zw4hp\") pod \"community-operators-zljc4\" (UID: \"40ff6ac2-e78c-4501-9fdd-8d612fd6d29d\") " pod="openshift-marketplace/community-operators-zljc4" Mar 18 20:13:32 crc kubenswrapper[4950]: I0318 20:13:32.865969 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40ff6ac2-e78c-4501-9fdd-8d612fd6d29d-catalog-content\") pod \"community-operators-zljc4\" (UID: \"40ff6ac2-e78c-4501-9fdd-8d612fd6d29d\") " pod="openshift-marketplace/community-operators-zljc4" Mar 18 20:13:32 crc kubenswrapper[4950]: I0318 20:13:32.866047 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40ff6ac2-e78c-4501-9fdd-8d612fd6d29d-utilities\") pod \"community-operators-zljc4\" (UID: \"40ff6ac2-e78c-4501-9fdd-8d612fd6d29d\") " pod="openshift-marketplace/community-operators-zljc4" Mar 18 20:13:32 crc kubenswrapper[4950]: I0318 20:13:32.866070 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw4hp\" (UniqueName: \"kubernetes.io/projected/40ff6ac2-e78c-4501-9fdd-8d612fd6d29d-kube-api-access-zw4hp\") pod \"community-operators-zljc4\" (UID: \"40ff6ac2-e78c-4501-9fdd-8d612fd6d29d\") " pod="openshift-marketplace/community-operators-zljc4" Mar 18 20:13:32 crc kubenswrapper[4950]: I0318 20:13:32.866640 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40ff6ac2-e78c-4501-9fdd-8d612fd6d29d-catalog-content\") pod \"community-operators-zljc4\" (UID: \"40ff6ac2-e78c-4501-9fdd-8d612fd6d29d\") " pod="openshift-marketplace/community-operators-zljc4" Mar 18 20:13:32 crc kubenswrapper[4950]: I0318 20:13:32.866753 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40ff6ac2-e78c-4501-9fdd-8d612fd6d29d-utilities\") pod \"community-operators-zljc4\" (UID: \"40ff6ac2-e78c-4501-9fdd-8d612fd6d29d\") " pod="openshift-marketplace/community-operators-zljc4" Mar 18 20:13:32 crc kubenswrapper[4950]: I0318 20:13:32.891452 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw4hp\" (UniqueName: \"kubernetes.io/projected/40ff6ac2-e78c-4501-9fdd-8d612fd6d29d-kube-api-access-zw4hp\") pod \"community-operators-zljc4\" (UID: \"40ff6ac2-e78c-4501-9fdd-8d612fd6d29d\") " pod="openshift-marketplace/community-operators-zljc4" Mar 18 20:13:32 crc kubenswrapper[4950]: I0318 20:13:32.956557 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zljc4" Mar 18 20:13:33 crc kubenswrapper[4950]: I0318 20:13:33.379237 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zljc4"] Mar 18 20:13:33 crc kubenswrapper[4950]: W0318 20:13:33.389850 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40ff6ac2_e78c_4501_9fdd_8d612fd6d29d.slice/crio-eafe1c8c6cb766fa3211ec4222c715e79cb05a7b691122e61fc3ba8f35837325 WatchSource:0}: Error finding container eafe1c8c6cb766fa3211ec4222c715e79cb05a7b691122e61fc3ba8f35837325: Status 404 returned error can't find the container with id eafe1c8c6cb766fa3211ec4222c715e79cb05a7b691122e61fc3ba8f35837325 Mar 18 20:13:33 crc kubenswrapper[4950]: I0318 20:13:33.446093 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xv6kx" event={"ID":"df5f1fd5-ec77-45be-84de-d0d8a1568a73","Type":"ContainerStarted","Data":"d1ebc8c7739487b24059e070ad1749ab1d229fc40e7cf2ee1a89c36dfe5d4d17"} Mar 18 20:13:33 crc kubenswrapper[4950]: I0318 20:13:33.447702 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zljc4" event={"ID":"40ff6ac2-e78c-4501-9fdd-8d612fd6d29d","Type":"ContainerStarted","Data":"eafe1c8c6cb766fa3211ec4222c715e79cb05a7b691122e61fc3ba8f35837325"} Mar 18 20:13:33 crc kubenswrapper[4950]: I0318 20:13:33.450456 4950 generic.go:334] "Generic (PLEG): container finished" podID="19ec358c-0c71-4b43-b13f-688f2103d895" containerID="b0ad742aceee551a0e56a24a8eacc8be2aec6e7d3baa4ac4f133f312cfc4003e" exitCode=0 Mar 18 20:13:33 crc kubenswrapper[4950]: I0318 20:13:33.450517 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-shgnx" event={"ID":"19ec358c-0c71-4b43-b13f-688f2103d895","Type":"ContainerDied","Data":"b0ad742aceee551a0e56a24a8eacc8be2aec6e7d3baa4ac4f133f312cfc4003e"} Mar 18 20:13:33 crc kubenswrapper[4950]: I0318 20:13:33.514735 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xv6kx" podStartSLOduration=3.006175458 podStartE2EDuration="5.514694541s" podCreationTimestamp="2026-03-18 20:13:28 +0000 UTC" firstStartedPulling="2026-03-18 20:13:30.355790687 +0000 UTC m=+423.596632555" lastFinishedPulling="2026-03-18 20:13:32.86430976 +0000 UTC m=+426.105151638" observedRunningTime="2026-03-18 20:13:33.483834538 +0000 UTC m=+426.724676406" watchObservedRunningTime="2026-03-18 20:13:33.514694541 +0000 UTC m=+426.755536409" Mar 18 20:13:34 crc kubenswrapper[4950]: I0318 20:13:34.456685 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldrm5" event={"ID":"81fa390c-72e6-471b-ac97-8e5d5fc3b4d7","Type":"ContainerStarted","Data":"346c36d7293d3fafd4d85fd5487209b524537b640bce764bf8408c27935c20c5"} Mar 18 20:13:34 crc kubenswrapper[4950]: I0318 20:13:34.457744 4950 generic.go:334] "Generic (PLEG): container finished" podID="40ff6ac2-e78c-4501-9fdd-8d612fd6d29d" containerID="01ee45d4034e1047bcde3cb5620e4261bdc8f61a276d86d580d502fe1e334414" exitCode=0 Mar 18 20:13:34 crc kubenswrapper[4950]: I0318 20:13:34.457801 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zljc4" event={"ID":"40ff6ac2-e78c-4501-9fdd-8d612fd6d29d","Type":"ContainerDied","Data":"01ee45d4034e1047bcde3cb5620e4261bdc8f61a276d86d580d502fe1e334414"} Mar 18 20:13:34 crc kubenswrapper[4950]: I0318 20:13:34.461225 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-shgnx" event={"ID":"19ec358c-0c71-4b43-b13f-688f2103d895","Type":"ContainerStarted","Data":"8a84ba90f608533a641be6dfdf96f9dcb5f62289443ef865cb7541945dd4c5df"} Mar 18 20:13:34 crc kubenswrapper[4950]: I0318 20:13:34.503902 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-shgnx" podStartSLOduration=2.030725132 podStartE2EDuration="4.50388431s" podCreationTimestamp="2026-03-18 20:13:30 +0000 UTC" firstStartedPulling="2026-03-18 20:13:31.408215523 +0000 UTC m=+424.649057401" lastFinishedPulling="2026-03-18 20:13:33.881374681 +0000 UTC m=+427.122216579" observedRunningTime="2026-03-18 20:13:34.500928599 +0000 UTC m=+427.741770467" watchObservedRunningTime="2026-03-18 20:13:34.50388431 +0000 UTC m=+427.744726168" Mar 18 20:13:35 crc kubenswrapper[4950]: I0318 20:13:35.468557 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zljc4" event={"ID":"40ff6ac2-e78c-4501-9fdd-8d612fd6d29d","Type":"ContainerStarted","Data":"a5fdcea53c9859485ec43a289d4c5e450dddea8eba47e8813511804874ad950c"} Mar 18 20:13:35 crc kubenswrapper[4950]: I0318 20:13:35.470903 4950 generic.go:334] "Generic (PLEG): container finished" podID="81fa390c-72e6-471b-ac97-8e5d5fc3b4d7" containerID="346c36d7293d3fafd4d85fd5487209b524537b640bce764bf8408c27935c20c5" exitCode=0 Mar 18 20:13:35 crc kubenswrapper[4950]: I0318 20:13:35.471402 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldrm5" event={"ID":"81fa390c-72e6-471b-ac97-8e5d5fc3b4d7","Type":"ContainerDied","Data":"346c36d7293d3fafd4d85fd5487209b524537b640bce764bf8408c27935c20c5"} Mar 18 20:13:36 crc kubenswrapper[4950]: I0318 20:13:36.477135 4950 generic.go:334] "Generic (PLEG): container finished" podID="40ff6ac2-e78c-4501-9fdd-8d612fd6d29d" containerID="a5fdcea53c9859485ec43a289d4c5e450dddea8eba47e8813511804874ad950c" exitCode=0 Mar 18 20:13:36 crc kubenswrapper[4950]: I0318 20:13:36.477216 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zljc4" event={"ID":"40ff6ac2-e78c-4501-9fdd-8d612fd6d29d","Type":"ContainerDied","Data":"a5fdcea53c9859485ec43a289d4c5e450dddea8eba47e8813511804874ad950c"} Mar 18 20:13:36 crc kubenswrapper[4950]: I0318 20:13:36.483266 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldrm5" event={"ID":"81fa390c-72e6-471b-ac97-8e5d5fc3b4d7","Type":"ContainerStarted","Data":"faa09adf7d326e2afe7514093466e5cd0922ed36ed761b54afb1c4b3a8ad9764"} Mar 18 20:13:37 crc kubenswrapper[4950]: I0318 20:13:37.490717 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zljc4" event={"ID":"40ff6ac2-e78c-4501-9fdd-8d612fd6d29d","Type":"ContainerStarted","Data":"6f1f4636513697f5cbe38cc6c35151b895b4ba36c750c7f303d889eb2ea8da1d"} Mar 18 20:13:37 crc kubenswrapper[4950]: I0318 20:13:37.507150 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ldrm5" podStartSLOduration=2.978515455 podStartE2EDuration="6.507134932s" podCreationTimestamp="2026-03-18 20:13:31 +0000 UTC" firstStartedPulling="2026-03-18 20:13:32.419192657 +0000 UTC m=+425.660034555" lastFinishedPulling="2026-03-18 20:13:35.947812154 +0000 UTC m=+429.188654032" observedRunningTime="2026-03-18 20:13:36.514811457 +0000 UTC m=+429.755653325" watchObservedRunningTime="2026-03-18 20:13:37.507134932 +0000 UTC m=+430.747976800" Mar 18 20:13:37 crc kubenswrapper[4950]: I0318 20:13:37.509170 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zljc4" podStartSLOduration=3.02140071 podStartE2EDuration="5.509164797s" podCreationTimestamp="2026-03-18 20:13:32 +0000 UTC" firstStartedPulling="2026-03-18 20:13:34.45888403 +0000 UTC m=+427.699725898" lastFinishedPulling="2026-03-18 20:13:36.946648117 +0000 UTC m=+430.187489985" observedRunningTime="2026-03-18 20:13:37.506269938 +0000 UTC m=+430.747111806" watchObservedRunningTime="2026-03-18 20:13:37.509164797 +0000 UTC m=+430.750006665" Mar 18 20:13:39 crc kubenswrapper[4950]: I0318 20:13:39.139262 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xv6kx" Mar 18 20:13:39 crc kubenswrapper[4950]: I0318 20:13:39.139649 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xv6kx" Mar 18 20:13:39 crc kubenswrapper[4950]: I0318 20:13:39.199033 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xv6kx" Mar 18 20:13:39 crc kubenswrapper[4950]: I0318 20:13:39.542208 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xv6kx" Mar 18 20:13:41 crc kubenswrapper[4950]: I0318 20:13:40.558737 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-shgnx" Mar 18 20:13:41 crc kubenswrapper[4950]: I0318 20:13:40.559084 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-shgnx" Mar 18 20:13:41 crc kubenswrapper[4950]: I0318 20:13:40.615296 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-shgnx" Mar 18 20:13:41 crc kubenswrapper[4950]: I0318 20:13:41.549852 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ldrm5" Mar 18 20:13:41 crc kubenswrapper[4950]: I0318 20:13:41.550196 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ldrm5" Mar 18 20:13:41 crc kubenswrapper[4950]: I0318 20:13:41.568552 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-shgnx" Mar 18 20:13:41 crc kubenswrapper[4950]: E0318 20:13:41.688537 4950 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2456057_5ff6_458f_af94_53edca249b7f.slice/crio-7164a8254b584d66dd328ab9294f1df8785a45c6f9bdd3507b1d426b574815f0.scope\": RecentStats: unable to find data in memory cache]" Mar 18 20:13:42 crc kubenswrapper[4950]: I0318 20:13:42.599552 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ldrm5" podUID="81fa390c-72e6-471b-ac97-8e5d5fc3b4d7" containerName="registry-server" probeResult="failure" output=< Mar 18 20:13:42 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 20:13:42 crc kubenswrapper[4950]: > Mar 18 20:13:42 crc kubenswrapper[4950]: I0318 20:13:42.956947 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zljc4" Mar 18 20:13:42 crc kubenswrapper[4950]: I0318 20:13:42.956985 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zljc4" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.026266 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-mhj9r"] Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.026987 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.049580 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zljc4" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.058126 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-mhj9r"] Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.093155 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w82m\" (UniqueName: \"kubernetes.io/projected/f5e72dda-7220-4abb-b703-7ffbfa001af5-kube-api-access-6w82m\") pod \"image-registry-66df7c8f76-mhj9r\" (UID: \"f5e72dda-7220-4abb-b703-7ffbfa001af5\") " pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.093202 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f5e72dda-7220-4abb-b703-7ffbfa001af5-registry-tls\") pod \"image-registry-66df7c8f76-mhj9r\" (UID: \"f5e72dda-7220-4abb-b703-7ffbfa001af5\") " pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.093222 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f5e72dda-7220-4abb-b703-7ffbfa001af5-installation-pull-secrets\") pod \"image-registry-66df7c8f76-mhj9r\" (UID: \"f5e72dda-7220-4abb-b703-7ffbfa001af5\") " pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.093349 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-mhj9r\" (UID: \"f5e72dda-7220-4abb-b703-7ffbfa001af5\") " pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.093430 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f5e72dda-7220-4abb-b703-7ffbfa001af5-ca-trust-extracted\") pod \"image-registry-66df7c8f76-mhj9r\" (UID: \"f5e72dda-7220-4abb-b703-7ffbfa001af5\") " pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.093506 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f5e72dda-7220-4abb-b703-7ffbfa001af5-bound-sa-token\") pod \"image-registry-66df7c8f76-mhj9r\" (UID: \"f5e72dda-7220-4abb-b703-7ffbfa001af5\") " pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.093545 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f5e72dda-7220-4abb-b703-7ffbfa001af5-trusted-ca\") pod \"image-registry-66df7c8f76-mhj9r\" (UID: \"f5e72dda-7220-4abb-b703-7ffbfa001af5\") " pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.093563 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f5e72dda-7220-4abb-b703-7ffbfa001af5-registry-certificates\") pod \"image-registry-66df7c8f76-mhj9r\" (UID: \"f5e72dda-7220-4abb-b703-7ffbfa001af5\") " pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.142462 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-mhj9r\" (UID: \"f5e72dda-7220-4abb-b703-7ffbfa001af5\") " pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.194257 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f5e72dda-7220-4abb-b703-7ffbfa001af5-bound-sa-token\") pod \"image-registry-66df7c8f76-mhj9r\" (UID: \"f5e72dda-7220-4abb-b703-7ffbfa001af5\") " pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.194477 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f5e72dda-7220-4abb-b703-7ffbfa001af5-trusted-ca\") pod \"image-registry-66df7c8f76-mhj9r\" (UID: \"f5e72dda-7220-4abb-b703-7ffbfa001af5\") " pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.194572 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f5e72dda-7220-4abb-b703-7ffbfa001af5-registry-certificates\") pod \"image-registry-66df7c8f76-mhj9r\" (UID: \"f5e72dda-7220-4abb-b703-7ffbfa001af5\") " pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.194696 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w82m\" (UniqueName: \"kubernetes.io/projected/f5e72dda-7220-4abb-b703-7ffbfa001af5-kube-api-access-6w82m\") pod \"image-registry-66df7c8f76-mhj9r\" (UID: \"f5e72dda-7220-4abb-b703-7ffbfa001af5\") " pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.194790 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f5e72dda-7220-4abb-b703-7ffbfa001af5-registry-tls\") pod \"image-registry-66df7c8f76-mhj9r\" (UID: \"f5e72dda-7220-4abb-b703-7ffbfa001af5\") " pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.194875 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f5e72dda-7220-4abb-b703-7ffbfa001af5-installation-pull-secrets\") pod \"image-registry-66df7c8f76-mhj9r\" (UID: \"f5e72dda-7220-4abb-b703-7ffbfa001af5\") " pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.194968 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f5e72dda-7220-4abb-b703-7ffbfa001af5-ca-trust-extracted\") pod \"image-registry-66df7c8f76-mhj9r\" (UID: \"f5e72dda-7220-4abb-b703-7ffbfa001af5\") " pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.195445 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f5e72dda-7220-4abb-b703-7ffbfa001af5-ca-trust-extracted\") pod \"image-registry-66df7c8f76-mhj9r\" (UID: \"f5e72dda-7220-4abb-b703-7ffbfa001af5\") " pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.195875 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f5e72dda-7220-4abb-b703-7ffbfa001af5-registry-certificates\") pod \"image-registry-66df7c8f76-mhj9r\" (UID: \"f5e72dda-7220-4abb-b703-7ffbfa001af5\") " pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.196950 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f5e72dda-7220-4abb-b703-7ffbfa001af5-trusted-ca\") pod \"image-registry-66df7c8f76-mhj9r\" (UID: \"f5e72dda-7220-4abb-b703-7ffbfa001af5\") " pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.200320 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f5e72dda-7220-4abb-b703-7ffbfa001af5-registry-tls\") pod \"image-registry-66df7c8f76-mhj9r\" (UID: \"f5e72dda-7220-4abb-b703-7ffbfa001af5\") " pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.200732 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f5e72dda-7220-4abb-b703-7ffbfa001af5-installation-pull-secrets\") pod \"image-registry-66df7c8f76-mhj9r\" (UID: \"f5e72dda-7220-4abb-b703-7ffbfa001af5\") " pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.209293 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f5e72dda-7220-4abb-b703-7ffbfa001af5-bound-sa-token\") pod \"image-registry-66df7c8f76-mhj9r\" (UID: \"f5e72dda-7220-4abb-b703-7ffbfa001af5\") " pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.217093 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w82m\" (UniqueName: \"kubernetes.io/projected/f5e72dda-7220-4abb-b703-7ffbfa001af5-kube-api-access-6w82m\") pod \"image-registry-66df7c8f76-mhj9r\" (UID: \"f5e72dda-7220-4abb-b703-7ffbfa001af5\") " pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.342490 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.584032 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zljc4" Mar 18 20:13:43 crc kubenswrapper[4950]: I0318 20:13:43.617596 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-mhj9r"] Mar 18 20:13:43 crc kubenswrapper[4950]: W0318 20:13:43.623446 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5e72dda_7220_4abb_b703_7ffbfa001af5.slice/crio-dde82114f8fca8762465ea2104235a371f13427868c97f23b51eff33d3ac5f40 WatchSource:0}: Error finding container dde82114f8fca8762465ea2104235a371f13427868c97f23b51eff33d3ac5f40: Status 404 returned error can't find the container with id dde82114f8fca8762465ea2104235a371f13427868c97f23b51eff33d3ac5f40 Mar 18 20:13:44 crc kubenswrapper[4950]: I0318 20:13:44.536243 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" event={"ID":"f5e72dda-7220-4abb-b703-7ffbfa001af5","Type":"ContainerStarted","Data":"c11e4126a839eeadd1e76d3fb281c824d509eda574870c9b20475de448d7be7d"} Mar 18 20:13:44 crc kubenswrapper[4950]: I0318 20:13:44.536516 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" event={"ID":"f5e72dda-7220-4abb-b703-7ffbfa001af5","Type":"ContainerStarted","Data":"dde82114f8fca8762465ea2104235a371f13427868c97f23b51eff33d3ac5f40"} Mar 18 20:13:44 crc kubenswrapper[4950]: I0318 20:13:44.536550 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:13:44 crc kubenswrapper[4950]: I0318 20:13:44.555727 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" podStartSLOduration=2.555712239 podStartE2EDuration="2.555712239s" podCreationTimestamp="2026-03-18 20:13:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:13:44.555494413 +0000 UTC m=+437.796336281" watchObservedRunningTime="2026-03-18 20:13:44.555712239 +0000 UTC m=+437.796554107" Mar 18 20:13:51 crc kubenswrapper[4950]: I0318 20:13:51.597505 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ldrm5" Mar 18 20:13:51 crc kubenswrapper[4950]: I0318 20:13:51.671677 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ldrm5" Mar 18 20:13:51 crc kubenswrapper[4950]: E0318 20:13:51.790809 4950 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2456057_5ff6_458f_af94_53edca249b7f.slice/crio-7164a8254b584d66dd328ab9294f1df8785a45c6f9bdd3507b1d426b574815f0.scope\": RecentStats: unable to find data in memory cache]" Mar 18 20:14:00 crc kubenswrapper[4950]: I0318 20:14:00.130215 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564414-nf9v5"] Mar 18 20:14:00 crc kubenswrapper[4950]: I0318 20:14:00.131470 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564414-nf9v5" Mar 18 20:14:00 crc kubenswrapper[4950]: I0318 20:14:00.133273 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:14:00 crc kubenswrapper[4950]: I0318 20:14:00.133587 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:14:00 crc kubenswrapper[4950]: I0318 20:14:00.133767 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:14:00 crc kubenswrapper[4950]: I0318 20:14:00.134915 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564414-nf9v5"] Mar 18 20:14:00 crc kubenswrapper[4950]: I0318 20:14:00.258002 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqfks\" (UniqueName: \"kubernetes.io/projected/88c51f5c-d48e-4779-879a-f7169556d9bd-kube-api-access-nqfks\") pod \"auto-csr-approver-29564414-nf9v5\" (UID: \"88c51f5c-d48e-4779-879a-f7169556d9bd\") " pod="openshift-infra/auto-csr-approver-29564414-nf9v5" Mar 18 20:14:00 crc kubenswrapper[4950]: I0318 20:14:00.359228 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqfks\" (UniqueName: \"kubernetes.io/projected/88c51f5c-d48e-4779-879a-f7169556d9bd-kube-api-access-nqfks\") pod \"auto-csr-approver-29564414-nf9v5\" (UID: \"88c51f5c-d48e-4779-879a-f7169556d9bd\") " pod="openshift-infra/auto-csr-approver-29564414-nf9v5" Mar 18 20:14:00 crc kubenswrapper[4950]: I0318 20:14:00.378089 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqfks\" (UniqueName: \"kubernetes.io/projected/88c51f5c-d48e-4779-879a-f7169556d9bd-kube-api-access-nqfks\") pod \"auto-csr-approver-29564414-nf9v5\" (UID: \"88c51f5c-d48e-4779-879a-f7169556d9bd\") " pod="openshift-infra/auto-csr-approver-29564414-nf9v5" Mar 18 20:14:00 crc kubenswrapper[4950]: I0318 20:14:00.457210 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564414-nf9v5" Mar 18 20:14:00 crc kubenswrapper[4950]: I0318 20:14:00.720353 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564414-nf9v5"] Mar 18 20:14:01 crc kubenswrapper[4950]: I0318 20:14:01.641256 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564414-nf9v5" event={"ID":"88c51f5c-d48e-4779-879a-f7169556d9bd","Type":"ContainerStarted","Data":"6dbc891a972fdf7cad6a6e1986ffd88ba269e65a36b044a95b7ea9c8f201f56e"} Mar 18 20:14:01 crc kubenswrapper[4950]: E0318 20:14:01.920281 4950 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2456057_5ff6_458f_af94_53edca249b7f.slice/crio-7164a8254b584d66dd328ab9294f1df8785a45c6f9bdd3507b1d426b574815f0.scope\": RecentStats: unable to find data in memory cache]" Mar 18 20:14:02 crc kubenswrapper[4950]: I0318 20:14:02.648055 4950 generic.go:334] "Generic (PLEG): container finished" podID="88c51f5c-d48e-4779-879a-f7169556d9bd" containerID="606b0b7a879af47936c0166efea78e8d4cb57aa6954e2812f7432305854e6981" exitCode=0 Mar 18 20:14:02 crc kubenswrapper[4950]: I0318 20:14:02.648119 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564414-nf9v5" event={"ID":"88c51f5c-d48e-4779-879a-f7169556d9bd","Type":"ContainerDied","Data":"606b0b7a879af47936c0166efea78e8d4cb57aa6954e2812f7432305854e6981"} Mar 18 20:14:03 crc kubenswrapper[4950]: I0318 20:14:03.348325 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-mhj9r" Mar 18 20:14:03 crc kubenswrapper[4950]: I0318 20:14:03.430966 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2kz74"] Mar 18 20:14:03 crc kubenswrapper[4950]: I0318 20:14:03.836020 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:14:03 crc kubenswrapper[4950]: I0318 20:14:03.836570 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:14:03 crc kubenswrapper[4950]: I0318 20:14:03.862192 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564414-nf9v5" Mar 18 20:14:03 crc kubenswrapper[4950]: I0318 20:14:03.905432 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqfks\" (UniqueName: \"kubernetes.io/projected/88c51f5c-d48e-4779-879a-f7169556d9bd-kube-api-access-nqfks\") pod \"88c51f5c-d48e-4779-879a-f7169556d9bd\" (UID: \"88c51f5c-d48e-4779-879a-f7169556d9bd\") " Mar 18 20:14:03 crc kubenswrapper[4950]: I0318 20:14:03.913934 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88c51f5c-d48e-4779-879a-f7169556d9bd-kube-api-access-nqfks" (OuterVolumeSpecName: "kube-api-access-nqfks") pod "88c51f5c-d48e-4779-879a-f7169556d9bd" (UID: "88c51f5c-d48e-4779-879a-f7169556d9bd"). InnerVolumeSpecName "kube-api-access-nqfks". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:14:04 crc kubenswrapper[4950]: I0318 20:14:04.006656 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqfks\" (UniqueName: \"kubernetes.io/projected/88c51f5c-d48e-4779-879a-f7169556d9bd-kube-api-access-nqfks\") on node \"crc\" DevicePath \"\"" Mar 18 20:14:04 crc kubenswrapper[4950]: I0318 20:14:04.659452 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564414-nf9v5" event={"ID":"88c51f5c-d48e-4779-879a-f7169556d9bd","Type":"ContainerDied","Data":"6dbc891a972fdf7cad6a6e1986ffd88ba269e65a36b044a95b7ea9c8f201f56e"} Mar 18 20:14:04 crc kubenswrapper[4950]: I0318 20:14:04.659675 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6dbc891a972fdf7cad6a6e1986ffd88ba269e65a36b044a95b7ea9c8f201f56e" Mar 18 20:14:04 crc kubenswrapper[4950]: I0318 20:14:04.659497 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564414-nf9v5" Mar 18 20:14:04 crc kubenswrapper[4950]: I0318 20:14:04.916332 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564408-2658d"] Mar 18 20:14:04 crc kubenswrapper[4950]: I0318 20:14:04.919765 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564408-2658d"] Mar 18 20:14:05 crc kubenswrapper[4950]: I0318 20:14:05.486574 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97cb884c-5f55-4ca0-8639-193c2f1df139" path="/var/lib/kubelet/pods/97cb884c-5f55-4ca0-8639-193c2f1df139/volumes" Mar 18 20:14:12 crc kubenswrapper[4950]: E0318 20:14:12.024442 4950 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2456057_5ff6_458f_af94_53edca249b7f.slice/crio-7164a8254b584d66dd328ab9294f1df8785a45c6f9bdd3507b1d426b574815f0.scope\": RecentStats: unable to find data in memory cache]" Mar 18 20:14:22 crc kubenswrapper[4950]: E0318 20:14:22.187560 4950 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2456057_5ff6_458f_af94_53edca249b7f.slice/crio-7164a8254b584d66dd328ab9294f1df8785a45c6f9bdd3507b1d426b574815f0.scope\": RecentStats: unable to find data in memory cache]" Mar 18 20:14:27 crc kubenswrapper[4950]: E0318 20:14:27.455335 4950 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/f19f721124c35b91a8e40e9e73950f5c06790d4b8aab0a787ac06063d3ad5b28/diff" to get inode usage: stat /var/lib/containers/storage/overlay/f19f721124c35b91a8e40e9e73950f5c06790d4b8aab0a787ac06063d3ad5b28/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-8nhwt_a2456057-5ff6-458f-af94-53edca249b7f/marketplace-operator/1.log" to get inode usage: stat /var/log/pods/openshift-marketplace_marketplace-operator-79b997595-8nhwt_a2456057-5ff6-458f-af94-53edca249b7f/marketplace-operator/1.log: no such file or directory Mar 18 20:14:28 crc kubenswrapper[4950]: I0318 20:14:28.474855 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" podUID="28e264f3-49f6-406a-b700-213e196574c1" containerName="registry" containerID="cri-o://3ceb89f25f617ed8ec72818a52205813cc4d8914af7fdcfad00cf1e9365d94ee" gracePeriod=30 Mar 18 20:14:28 crc kubenswrapper[4950]: I0318 20:14:28.819640 4950 generic.go:334] "Generic (PLEG): container finished" podID="28e264f3-49f6-406a-b700-213e196574c1" containerID="3ceb89f25f617ed8ec72818a52205813cc4d8914af7fdcfad00cf1e9365d94ee" exitCode=0 Mar 18 20:14:28 crc kubenswrapper[4950]: I0318 20:14:28.820203 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" event={"ID":"28e264f3-49f6-406a-b700-213e196574c1","Type":"ContainerDied","Data":"3ceb89f25f617ed8ec72818a52205813cc4d8914af7fdcfad00cf1e9365d94ee"} Mar 18 20:14:28 crc kubenswrapper[4950]: I0318 20:14:28.901221 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.065917 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"28e264f3-49f6-406a-b700-213e196574c1\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.066119 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pg8mb\" (UniqueName: \"kubernetes.io/projected/28e264f3-49f6-406a-b700-213e196574c1-kube-api-access-pg8mb\") pod \"28e264f3-49f6-406a-b700-213e196574c1\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.066165 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/28e264f3-49f6-406a-b700-213e196574c1-bound-sa-token\") pod \"28e264f3-49f6-406a-b700-213e196574c1\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.066239 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/28e264f3-49f6-406a-b700-213e196574c1-ca-trust-extracted\") pod \"28e264f3-49f6-406a-b700-213e196574c1\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.066318 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/28e264f3-49f6-406a-b700-213e196574c1-trusted-ca\") pod \"28e264f3-49f6-406a-b700-213e196574c1\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.066467 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/28e264f3-49f6-406a-b700-213e196574c1-installation-pull-secrets\") pod \"28e264f3-49f6-406a-b700-213e196574c1\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.066534 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/28e264f3-49f6-406a-b700-213e196574c1-registry-tls\") pod \"28e264f3-49f6-406a-b700-213e196574c1\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.066581 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/28e264f3-49f6-406a-b700-213e196574c1-registry-certificates\") pod \"28e264f3-49f6-406a-b700-213e196574c1\" (UID: \"28e264f3-49f6-406a-b700-213e196574c1\") " Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.067665 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28e264f3-49f6-406a-b700-213e196574c1-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "28e264f3-49f6-406a-b700-213e196574c1" (UID: "28e264f3-49f6-406a-b700-213e196574c1"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.067903 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28e264f3-49f6-406a-b700-213e196574c1-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "28e264f3-49f6-406a-b700-213e196574c1" (UID: "28e264f3-49f6-406a-b700-213e196574c1"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.079202 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28e264f3-49f6-406a-b700-213e196574c1-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "28e264f3-49f6-406a-b700-213e196574c1" (UID: "28e264f3-49f6-406a-b700-213e196574c1"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.081636 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28e264f3-49f6-406a-b700-213e196574c1-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "28e264f3-49f6-406a-b700-213e196574c1" (UID: "28e264f3-49f6-406a-b700-213e196574c1"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.081834 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28e264f3-49f6-406a-b700-213e196574c1-kube-api-access-pg8mb" (OuterVolumeSpecName: "kube-api-access-pg8mb") pod "28e264f3-49f6-406a-b700-213e196574c1" (UID: "28e264f3-49f6-406a-b700-213e196574c1"). InnerVolumeSpecName "kube-api-access-pg8mb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.084087 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "28e264f3-49f6-406a-b700-213e196574c1" (UID: "28e264f3-49f6-406a-b700-213e196574c1"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.087829 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28e264f3-49f6-406a-b700-213e196574c1-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "28e264f3-49f6-406a-b700-213e196574c1" (UID: "28e264f3-49f6-406a-b700-213e196574c1"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.095592 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28e264f3-49f6-406a-b700-213e196574c1-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "28e264f3-49f6-406a-b700-213e196574c1" (UID: "28e264f3-49f6-406a-b700-213e196574c1"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.167884 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pg8mb\" (UniqueName: \"kubernetes.io/projected/28e264f3-49f6-406a-b700-213e196574c1-kube-api-access-pg8mb\") on node \"crc\" DevicePath \"\"" Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.167927 4950 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/28e264f3-49f6-406a-b700-213e196574c1-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.167943 4950 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/28e264f3-49f6-406a-b700-213e196574c1-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.167957 4950 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/28e264f3-49f6-406a-b700-213e196574c1-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.167974 4950 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/28e264f3-49f6-406a-b700-213e196574c1-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.167990 4950 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/28e264f3-49f6-406a-b700-213e196574c1-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.168002 4950 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/28e264f3-49f6-406a-b700-213e196574c1-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.830697 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" event={"ID":"28e264f3-49f6-406a-b700-213e196574c1","Type":"ContainerDied","Data":"4de76fd78aa150421888d2bdca5e4a9c772694329c1d994d47ff4b7d136ada6a"} Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.830774 4950 scope.go:117] "RemoveContainer" containerID="3ceb89f25f617ed8ec72818a52205813cc4d8914af7fdcfad00cf1e9365d94ee" Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.830942 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2kz74" Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.862225 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2kz74"] Mar 18 20:14:29 crc kubenswrapper[4950]: I0318 20:14:29.870281 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2kz74"] Mar 18 20:14:31 crc kubenswrapper[4950]: I0318 20:14:31.500658 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28e264f3-49f6-406a-b700-213e196574c1" path="/var/lib/kubelet/pods/28e264f3-49f6-406a-b700-213e196574c1/volumes" Mar 18 20:14:33 crc kubenswrapper[4950]: I0318 20:14:33.836869 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:14:33 crc kubenswrapper[4950]: I0318 20:14:33.836998 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:15:00 crc kubenswrapper[4950]: I0318 20:15:00.159517 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564415-rpc7c"] Mar 18 20:15:00 crc kubenswrapper[4950]: E0318 20:15:00.160515 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88c51f5c-d48e-4779-879a-f7169556d9bd" containerName="oc" Mar 18 20:15:00 crc kubenswrapper[4950]: I0318 20:15:00.160549 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="88c51f5c-d48e-4779-879a-f7169556d9bd" containerName="oc" Mar 18 20:15:00 crc kubenswrapper[4950]: E0318 20:15:00.160578 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28e264f3-49f6-406a-b700-213e196574c1" containerName="registry" Mar 18 20:15:00 crc kubenswrapper[4950]: I0318 20:15:00.160594 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="28e264f3-49f6-406a-b700-213e196574c1" containerName="registry" Mar 18 20:15:00 crc kubenswrapper[4950]: I0318 20:15:00.160786 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="88c51f5c-d48e-4779-879a-f7169556d9bd" containerName="oc" Mar 18 20:15:00 crc kubenswrapper[4950]: I0318 20:15:00.160819 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="28e264f3-49f6-406a-b700-213e196574c1" containerName="registry" Mar 18 20:15:00 crc kubenswrapper[4950]: I0318 20:15:00.161523 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564415-rpc7c" Mar 18 20:15:00 crc kubenswrapper[4950]: I0318 20:15:00.165777 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 18 20:15:00 crc kubenswrapper[4950]: I0318 20:15:00.170051 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 18 20:15:00 crc kubenswrapper[4950]: I0318 20:15:00.184809 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564415-rpc7c"] Mar 18 20:15:00 crc kubenswrapper[4950]: I0318 20:15:00.224478 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/49fb4fb7-6a60-4eeb-b1bf-349a1948b08f-config-volume\") pod \"collect-profiles-29564415-rpc7c\" (UID: \"49fb4fb7-6a60-4eeb-b1bf-349a1948b08f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564415-rpc7c" Mar 18 20:15:00 crc kubenswrapper[4950]: I0318 20:15:00.224810 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/49fb4fb7-6a60-4eeb-b1bf-349a1948b08f-secret-volume\") pod \"collect-profiles-29564415-rpc7c\" (UID: \"49fb4fb7-6a60-4eeb-b1bf-349a1948b08f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564415-rpc7c" Mar 18 20:15:00 crc kubenswrapper[4950]: I0318 20:15:00.326149 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/49fb4fb7-6a60-4eeb-b1bf-349a1948b08f-secret-volume\") pod \"collect-profiles-29564415-rpc7c\" (UID: \"49fb4fb7-6a60-4eeb-b1bf-349a1948b08f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564415-rpc7c" Mar 18 20:15:00 crc kubenswrapper[4950]: I0318 20:15:00.326210 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vt67\" (UniqueName: \"kubernetes.io/projected/49fb4fb7-6a60-4eeb-b1bf-349a1948b08f-kube-api-access-5vt67\") pod \"collect-profiles-29564415-rpc7c\" (UID: \"49fb4fb7-6a60-4eeb-b1bf-349a1948b08f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564415-rpc7c" Mar 18 20:15:00 crc kubenswrapper[4950]: I0318 20:15:00.326303 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/49fb4fb7-6a60-4eeb-b1bf-349a1948b08f-config-volume\") pod \"collect-profiles-29564415-rpc7c\" (UID: \"49fb4fb7-6a60-4eeb-b1bf-349a1948b08f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564415-rpc7c" Mar 18 20:15:00 crc kubenswrapper[4950]: I0318 20:15:00.327059 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/49fb4fb7-6a60-4eeb-b1bf-349a1948b08f-config-volume\") pod \"collect-profiles-29564415-rpc7c\" (UID: \"49fb4fb7-6a60-4eeb-b1bf-349a1948b08f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564415-rpc7c" Mar 18 20:15:00 crc kubenswrapper[4950]: I0318 20:15:00.331995 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/49fb4fb7-6a60-4eeb-b1bf-349a1948b08f-secret-volume\") pod \"collect-profiles-29564415-rpc7c\" (UID: \"49fb4fb7-6a60-4eeb-b1bf-349a1948b08f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564415-rpc7c" Mar 18 20:15:00 crc kubenswrapper[4950]: I0318 20:15:00.428212 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vt67\" (UniqueName: \"kubernetes.io/projected/49fb4fb7-6a60-4eeb-b1bf-349a1948b08f-kube-api-access-5vt67\") pod \"collect-profiles-29564415-rpc7c\" (UID: \"49fb4fb7-6a60-4eeb-b1bf-349a1948b08f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564415-rpc7c" Mar 18 20:15:00 crc kubenswrapper[4950]: I0318 20:15:00.450087 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vt67\" (UniqueName: \"kubernetes.io/projected/49fb4fb7-6a60-4eeb-b1bf-349a1948b08f-kube-api-access-5vt67\") pod \"collect-profiles-29564415-rpc7c\" (UID: \"49fb4fb7-6a60-4eeb-b1bf-349a1948b08f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564415-rpc7c" Mar 18 20:15:00 crc kubenswrapper[4950]: I0318 20:15:00.535273 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564415-rpc7c" Mar 18 20:15:00 crc kubenswrapper[4950]: I0318 20:15:00.737705 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564415-rpc7c"] Mar 18 20:15:01 crc kubenswrapper[4950]: I0318 20:15:01.099021 4950 generic.go:334] "Generic (PLEG): container finished" podID="49fb4fb7-6a60-4eeb-b1bf-349a1948b08f" containerID="1eb8b7086479d629bb79852b059d6109d1860922ae7b4b4422572d2d9d2d89bc" exitCode=0 Mar 18 20:15:01 crc kubenswrapper[4950]: I0318 20:15:01.099103 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564415-rpc7c" event={"ID":"49fb4fb7-6a60-4eeb-b1bf-349a1948b08f","Type":"ContainerDied","Data":"1eb8b7086479d629bb79852b059d6109d1860922ae7b4b4422572d2d9d2d89bc"} Mar 18 20:15:01 crc kubenswrapper[4950]: I0318 20:15:01.099362 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564415-rpc7c" event={"ID":"49fb4fb7-6a60-4eeb-b1bf-349a1948b08f","Type":"ContainerStarted","Data":"a066ecbac837609addc78f849777945009f7d9dcf994ae5ea68183e819dd4daa"} Mar 18 20:15:02 crc kubenswrapper[4950]: I0318 20:15:02.343719 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564415-rpc7c" Mar 18 20:15:02 crc kubenswrapper[4950]: I0318 20:15:02.464562 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vt67\" (UniqueName: \"kubernetes.io/projected/49fb4fb7-6a60-4eeb-b1bf-349a1948b08f-kube-api-access-5vt67\") pod \"49fb4fb7-6a60-4eeb-b1bf-349a1948b08f\" (UID: \"49fb4fb7-6a60-4eeb-b1bf-349a1948b08f\") " Mar 18 20:15:02 crc kubenswrapper[4950]: I0318 20:15:02.464627 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/49fb4fb7-6a60-4eeb-b1bf-349a1948b08f-secret-volume\") pod \"49fb4fb7-6a60-4eeb-b1bf-349a1948b08f\" (UID: \"49fb4fb7-6a60-4eeb-b1bf-349a1948b08f\") " Mar 18 20:15:02 crc kubenswrapper[4950]: I0318 20:15:02.464678 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/49fb4fb7-6a60-4eeb-b1bf-349a1948b08f-config-volume\") pod \"49fb4fb7-6a60-4eeb-b1bf-349a1948b08f\" (UID: \"49fb4fb7-6a60-4eeb-b1bf-349a1948b08f\") " Mar 18 20:15:02 crc kubenswrapper[4950]: I0318 20:15:02.465499 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49fb4fb7-6a60-4eeb-b1bf-349a1948b08f-config-volume" (OuterVolumeSpecName: "config-volume") pod "49fb4fb7-6a60-4eeb-b1bf-349a1948b08f" (UID: "49fb4fb7-6a60-4eeb-b1bf-349a1948b08f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:15:02 crc kubenswrapper[4950]: I0318 20:15:02.465680 4950 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/49fb4fb7-6a60-4eeb-b1bf-349a1948b08f-config-volume\") on node \"crc\" DevicePath \"\"" Mar 18 20:15:02 crc kubenswrapper[4950]: I0318 20:15:02.469874 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49fb4fb7-6a60-4eeb-b1bf-349a1948b08f-kube-api-access-5vt67" (OuterVolumeSpecName: "kube-api-access-5vt67") pod "49fb4fb7-6a60-4eeb-b1bf-349a1948b08f" (UID: "49fb4fb7-6a60-4eeb-b1bf-349a1948b08f"). InnerVolumeSpecName "kube-api-access-5vt67". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:15:02 crc kubenswrapper[4950]: I0318 20:15:02.473584 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49fb4fb7-6a60-4eeb-b1bf-349a1948b08f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "49fb4fb7-6a60-4eeb-b1bf-349a1948b08f" (UID: "49fb4fb7-6a60-4eeb-b1bf-349a1948b08f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:15:02 crc kubenswrapper[4950]: I0318 20:15:02.566998 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vt67\" (UniqueName: \"kubernetes.io/projected/49fb4fb7-6a60-4eeb-b1bf-349a1948b08f-kube-api-access-5vt67\") on node \"crc\" DevicePath \"\"" Mar 18 20:15:02 crc kubenswrapper[4950]: I0318 20:15:02.567053 4950 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/49fb4fb7-6a60-4eeb-b1bf-349a1948b08f-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 18 20:15:03 crc kubenswrapper[4950]: I0318 20:15:03.114094 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564415-rpc7c" event={"ID":"49fb4fb7-6a60-4eeb-b1bf-349a1948b08f","Type":"ContainerDied","Data":"a066ecbac837609addc78f849777945009f7d9dcf994ae5ea68183e819dd4daa"} Mar 18 20:15:03 crc kubenswrapper[4950]: I0318 20:15:03.114461 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a066ecbac837609addc78f849777945009f7d9dcf994ae5ea68183e819dd4daa" Mar 18 20:15:03 crc kubenswrapper[4950]: I0318 20:15:03.114194 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564415-rpc7c" Mar 18 20:15:03 crc kubenswrapper[4950]: I0318 20:15:03.836930 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:15:03 crc kubenswrapper[4950]: I0318 20:15:03.838172 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:15:03 crc kubenswrapper[4950]: I0318 20:15:03.838356 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:15:03 crc kubenswrapper[4950]: I0318 20:15:03.840693 4950 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"01edc7246d617f13699dceaa220ac50fb8349729df73606eb00a73d4e3ae8f6a"} pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 20:15:03 crc kubenswrapper[4950]: I0318 20:15:03.840959 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" containerID="cri-o://01edc7246d617f13699dceaa220ac50fb8349729df73606eb00a73d4e3ae8f6a" gracePeriod=600 Mar 18 20:15:04 crc kubenswrapper[4950]: I0318 20:15:04.121361 4950 generic.go:334] "Generic (PLEG): container finished" podID="4048b439-3266-46e7-9de0-22377efacc46" containerID="01edc7246d617f13699dceaa220ac50fb8349729df73606eb00a73d4e3ae8f6a" exitCode=0 Mar 18 20:15:04 crc kubenswrapper[4950]: I0318 20:15:04.121541 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerDied","Data":"01edc7246d617f13699dceaa220ac50fb8349729df73606eb00a73d4e3ae8f6a"} Mar 18 20:15:04 crc kubenswrapper[4950]: I0318 20:15:04.121998 4950 scope.go:117] "RemoveContainer" containerID="a7d1997cbdc12e287ca60347613b2e778e09e74b9c34651c4c21eb96076f0217" Mar 18 20:15:05 crc kubenswrapper[4950]: I0318 20:15:05.132950 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerStarted","Data":"7eafe5dcfcc22883e166beb04dbf1c20eecc93c473bcbfb2c29f6e0319e288f2"} Mar 18 20:16:00 crc kubenswrapper[4950]: I0318 20:16:00.154011 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564416-mcvh6"] Mar 18 20:16:00 crc kubenswrapper[4950]: E0318 20:16:00.155017 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49fb4fb7-6a60-4eeb-b1bf-349a1948b08f" containerName="collect-profiles" Mar 18 20:16:00 crc kubenswrapper[4950]: I0318 20:16:00.155037 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="49fb4fb7-6a60-4eeb-b1bf-349a1948b08f" containerName="collect-profiles" Mar 18 20:16:00 crc kubenswrapper[4950]: I0318 20:16:00.155175 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="49fb4fb7-6a60-4eeb-b1bf-349a1948b08f" containerName="collect-profiles" Mar 18 20:16:00 crc kubenswrapper[4950]: I0318 20:16:00.155715 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564416-mcvh6" Mar 18 20:16:00 crc kubenswrapper[4950]: I0318 20:16:00.160392 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:16:00 crc kubenswrapper[4950]: I0318 20:16:00.173931 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:16:00 crc kubenswrapper[4950]: I0318 20:16:00.178892 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564416-mcvh6"] Mar 18 20:16:00 crc kubenswrapper[4950]: I0318 20:16:00.179150 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:16:00 crc kubenswrapper[4950]: I0318 20:16:00.223758 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrsft\" (UniqueName: \"kubernetes.io/projected/13558485-efd4-47b9-a7ad-2a82ed08fe3b-kube-api-access-lrsft\") pod \"auto-csr-approver-29564416-mcvh6\" (UID: \"13558485-efd4-47b9-a7ad-2a82ed08fe3b\") " pod="openshift-infra/auto-csr-approver-29564416-mcvh6" Mar 18 20:16:00 crc kubenswrapper[4950]: I0318 20:16:00.325108 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrsft\" (UniqueName: \"kubernetes.io/projected/13558485-efd4-47b9-a7ad-2a82ed08fe3b-kube-api-access-lrsft\") pod \"auto-csr-approver-29564416-mcvh6\" (UID: \"13558485-efd4-47b9-a7ad-2a82ed08fe3b\") " pod="openshift-infra/auto-csr-approver-29564416-mcvh6" Mar 18 20:16:00 crc kubenswrapper[4950]: I0318 20:16:00.352752 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrsft\" (UniqueName: \"kubernetes.io/projected/13558485-efd4-47b9-a7ad-2a82ed08fe3b-kube-api-access-lrsft\") pod \"auto-csr-approver-29564416-mcvh6\" (UID: \"13558485-efd4-47b9-a7ad-2a82ed08fe3b\") " pod="openshift-infra/auto-csr-approver-29564416-mcvh6" Mar 18 20:16:00 crc kubenswrapper[4950]: I0318 20:16:00.483183 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564416-mcvh6" Mar 18 20:16:00 crc kubenswrapper[4950]: I0318 20:16:00.753097 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564416-mcvh6"] Mar 18 20:16:00 crc kubenswrapper[4950]: I0318 20:16:00.766180 4950 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 20:16:01 crc kubenswrapper[4950]: I0318 20:16:01.527770 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564416-mcvh6" event={"ID":"13558485-efd4-47b9-a7ad-2a82ed08fe3b","Type":"ContainerStarted","Data":"13a656dfad2e2656757da6041fe94b7a3d5d0ee644291a610ec0285241424ff4"} Mar 18 20:16:02 crc kubenswrapper[4950]: I0318 20:16:02.534158 4950 generic.go:334] "Generic (PLEG): container finished" podID="13558485-efd4-47b9-a7ad-2a82ed08fe3b" containerID="c501268b7c004983273092ea349fea475a7416cf2525156462cb37c5bdc76999" exitCode=0 Mar 18 20:16:02 crc kubenswrapper[4950]: I0318 20:16:02.534199 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564416-mcvh6" event={"ID":"13558485-efd4-47b9-a7ad-2a82ed08fe3b","Type":"ContainerDied","Data":"c501268b7c004983273092ea349fea475a7416cf2525156462cb37c5bdc76999"} Mar 18 20:16:03 crc kubenswrapper[4950]: I0318 20:16:03.751752 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564416-mcvh6" Mar 18 20:16:03 crc kubenswrapper[4950]: I0318 20:16:03.893528 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrsft\" (UniqueName: \"kubernetes.io/projected/13558485-efd4-47b9-a7ad-2a82ed08fe3b-kube-api-access-lrsft\") pod \"13558485-efd4-47b9-a7ad-2a82ed08fe3b\" (UID: \"13558485-efd4-47b9-a7ad-2a82ed08fe3b\") " Mar 18 20:16:03 crc kubenswrapper[4950]: I0318 20:16:03.898487 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13558485-efd4-47b9-a7ad-2a82ed08fe3b-kube-api-access-lrsft" (OuterVolumeSpecName: "kube-api-access-lrsft") pod "13558485-efd4-47b9-a7ad-2a82ed08fe3b" (UID: "13558485-efd4-47b9-a7ad-2a82ed08fe3b"). InnerVolumeSpecName "kube-api-access-lrsft". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:16:03 crc kubenswrapper[4950]: I0318 20:16:03.995131 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrsft\" (UniqueName: \"kubernetes.io/projected/13558485-efd4-47b9-a7ad-2a82ed08fe3b-kube-api-access-lrsft\") on node \"crc\" DevicePath \"\"" Mar 18 20:16:04 crc kubenswrapper[4950]: I0318 20:16:04.547817 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564416-mcvh6" event={"ID":"13558485-efd4-47b9-a7ad-2a82ed08fe3b","Type":"ContainerDied","Data":"13a656dfad2e2656757da6041fe94b7a3d5d0ee644291a610ec0285241424ff4"} Mar 18 20:16:04 crc kubenswrapper[4950]: I0318 20:16:04.547884 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13a656dfad2e2656757da6041fe94b7a3d5d0ee644291a610ec0285241424ff4" Mar 18 20:16:04 crc kubenswrapper[4950]: I0318 20:16:04.548202 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564416-mcvh6" Mar 18 20:16:04 crc kubenswrapper[4950]: I0318 20:16:04.819316 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564410-9hqrd"] Mar 18 20:16:04 crc kubenswrapper[4950]: I0318 20:16:04.832212 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564410-9hqrd"] Mar 18 20:16:05 crc kubenswrapper[4950]: I0318 20:16:05.493795 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="627e6f3d-cde5-4849-bd02-6437622f4978" path="/var/lib/kubelet/pods/627e6f3d-cde5-4849-bd02-6437622f4978/volumes" Mar 18 20:17:33 crc kubenswrapper[4950]: I0318 20:17:33.836905 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:17:33 crc kubenswrapper[4950]: I0318 20:17:33.837830 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:17:44 crc kubenswrapper[4950]: I0318 20:17:44.250210 4950 scope.go:117] "RemoveContainer" containerID="00dffb6edda0aae584ee703172b0a41786cf67c059f901a0afd2d75aeedf585d" Mar 18 20:17:44 crc kubenswrapper[4950]: I0318 20:17:44.297446 4950 scope.go:117] "RemoveContainer" containerID="d790f769ee6aced97a1576b8d64131e51abc022caf2a585431da9cc291585aeb" Mar 18 20:18:00 crc kubenswrapper[4950]: I0318 20:18:00.143532 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564418-bh82f"] Mar 18 20:18:00 crc kubenswrapper[4950]: E0318 20:18:00.144339 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13558485-efd4-47b9-a7ad-2a82ed08fe3b" containerName="oc" Mar 18 20:18:00 crc kubenswrapper[4950]: I0318 20:18:00.144352 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="13558485-efd4-47b9-a7ad-2a82ed08fe3b" containerName="oc" Mar 18 20:18:00 crc kubenswrapper[4950]: I0318 20:18:00.144488 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="13558485-efd4-47b9-a7ad-2a82ed08fe3b" containerName="oc" Mar 18 20:18:00 crc kubenswrapper[4950]: I0318 20:18:00.144890 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564418-bh82f" Mar 18 20:18:00 crc kubenswrapper[4950]: I0318 20:18:00.146883 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:18:00 crc kubenswrapper[4950]: I0318 20:18:00.147273 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:18:00 crc kubenswrapper[4950]: I0318 20:18:00.147870 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:18:00 crc kubenswrapper[4950]: I0318 20:18:00.156494 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564418-bh82f"] Mar 18 20:18:00 crc kubenswrapper[4950]: I0318 20:18:00.252448 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zf8c\" (UniqueName: \"kubernetes.io/projected/9f375d3f-628f-47a7-b2e5-d8afe6859aa8-kube-api-access-9zf8c\") pod \"auto-csr-approver-29564418-bh82f\" (UID: \"9f375d3f-628f-47a7-b2e5-d8afe6859aa8\") " pod="openshift-infra/auto-csr-approver-29564418-bh82f" Mar 18 20:18:00 crc kubenswrapper[4950]: I0318 20:18:00.353533 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zf8c\" (UniqueName: \"kubernetes.io/projected/9f375d3f-628f-47a7-b2e5-d8afe6859aa8-kube-api-access-9zf8c\") pod \"auto-csr-approver-29564418-bh82f\" (UID: \"9f375d3f-628f-47a7-b2e5-d8afe6859aa8\") " pod="openshift-infra/auto-csr-approver-29564418-bh82f" Mar 18 20:18:00 crc kubenswrapper[4950]: I0318 20:18:00.662924 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zf8c\" (UniqueName: \"kubernetes.io/projected/9f375d3f-628f-47a7-b2e5-d8afe6859aa8-kube-api-access-9zf8c\") pod \"auto-csr-approver-29564418-bh82f\" (UID: \"9f375d3f-628f-47a7-b2e5-d8afe6859aa8\") " pod="openshift-infra/auto-csr-approver-29564418-bh82f" Mar 18 20:18:00 crc kubenswrapper[4950]: I0318 20:18:00.767302 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564418-bh82f" Mar 18 20:18:01 crc kubenswrapper[4950]: I0318 20:18:01.016964 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564418-bh82f"] Mar 18 20:18:01 crc kubenswrapper[4950]: I0318 20:18:01.656386 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564418-bh82f" event={"ID":"9f375d3f-628f-47a7-b2e5-d8afe6859aa8","Type":"ContainerStarted","Data":"9b55e75c85b882f54ef92736fb9e396919f1da775bcff152aac952c9d3b3cd82"} Mar 18 20:18:02 crc kubenswrapper[4950]: I0318 20:18:02.668244 4950 generic.go:334] "Generic (PLEG): container finished" podID="9f375d3f-628f-47a7-b2e5-d8afe6859aa8" containerID="b9e558de791450ddb096532acb3ce036f8ba7ae1f3018bd784ce268015e46df9" exitCode=0 Mar 18 20:18:02 crc kubenswrapper[4950]: I0318 20:18:02.668294 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564418-bh82f" event={"ID":"9f375d3f-628f-47a7-b2e5-d8afe6859aa8","Type":"ContainerDied","Data":"b9e558de791450ddb096532acb3ce036f8ba7ae1f3018bd784ce268015e46df9"} Mar 18 20:18:03 crc kubenswrapper[4950]: I0318 20:18:03.837615 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:18:03 crc kubenswrapper[4950]: I0318 20:18:03.837693 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:18:03 crc kubenswrapper[4950]: I0318 20:18:03.954810 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564418-bh82f" Mar 18 20:18:04 crc kubenswrapper[4950]: I0318 20:18:04.105993 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zf8c\" (UniqueName: \"kubernetes.io/projected/9f375d3f-628f-47a7-b2e5-d8afe6859aa8-kube-api-access-9zf8c\") pod \"9f375d3f-628f-47a7-b2e5-d8afe6859aa8\" (UID: \"9f375d3f-628f-47a7-b2e5-d8afe6859aa8\") " Mar 18 20:18:04 crc kubenswrapper[4950]: I0318 20:18:04.320874 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f375d3f-628f-47a7-b2e5-d8afe6859aa8-kube-api-access-9zf8c" (OuterVolumeSpecName: "kube-api-access-9zf8c") pod "9f375d3f-628f-47a7-b2e5-d8afe6859aa8" (UID: "9f375d3f-628f-47a7-b2e5-d8afe6859aa8"). InnerVolumeSpecName "kube-api-access-9zf8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:18:04 crc kubenswrapper[4950]: I0318 20:18:04.410621 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zf8c\" (UniqueName: \"kubernetes.io/projected/9f375d3f-628f-47a7-b2e5-d8afe6859aa8-kube-api-access-9zf8c\") on node \"crc\" DevicePath \"\"" Mar 18 20:18:04 crc kubenswrapper[4950]: I0318 20:18:04.686073 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564418-bh82f" event={"ID":"9f375d3f-628f-47a7-b2e5-d8afe6859aa8","Type":"ContainerDied","Data":"9b55e75c85b882f54ef92736fb9e396919f1da775bcff152aac952c9d3b3cd82"} Mar 18 20:18:04 crc kubenswrapper[4950]: I0318 20:18:04.686131 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b55e75c85b882f54ef92736fb9e396919f1da775bcff152aac952c9d3b3cd82" Mar 18 20:18:04 crc kubenswrapper[4950]: I0318 20:18:04.686183 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564418-bh82f" Mar 18 20:18:05 crc kubenswrapper[4950]: I0318 20:18:05.044638 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564412-s92bw"] Mar 18 20:18:05 crc kubenswrapper[4950]: I0318 20:18:05.051455 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564412-s92bw"] Mar 18 20:18:05 crc kubenswrapper[4950]: I0318 20:18:05.494438 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7e9b7e8-14e9-422a-8b97-437f88df2595" path="/var/lib/kubelet/pods/b7e9b7e8-14e9-422a-8b97-437f88df2595/volumes" Mar 18 20:18:33 crc kubenswrapper[4950]: I0318 20:18:33.835951 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:18:33 crc kubenswrapper[4950]: I0318 20:18:33.836820 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:18:33 crc kubenswrapper[4950]: I0318 20:18:33.836893 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:18:33 crc kubenswrapper[4950]: I0318 20:18:33.837764 4950 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7eafe5dcfcc22883e166beb04dbf1c20eecc93c473bcbfb2c29f6e0319e288f2"} pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 20:18:33 crc kubenswrapper[4950]: I0318 20:18:33.837916 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" containerID="cri-o://7eafe5dcfcc22883e166beb04dbf1c20eecc93c473bcbfb2c29f6e0319e288f2" gracePeriod=600 Mar 18 20:18:34 crc kubenswrapper[4950]: I0318 20:18:34.890495 4950 generic.go:334] "Generic (PLEG): container finished" podID="4048b439-3266-46e7-9de0-22377efacc46" containerID="7eafe5dcfcc22883e166beb04dbf1c20eecc93c473bcbfb2c29f6e0319e288f2" exitCode=0 Mar 18 20:18:34 crc kubenswrapper[4950]: I0318 20:18:34.890579 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerDied","Data":"7eafe5dcfcc22883e166beb04dbf1c20eecc93c473bcbfb2c29f6e0319e288f2"} Mar 18 20:18:34 crc kubenswrapper[4950]: I0318 20:18:34.890915 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerStarted","Data":"bef1deca2fab314ca277f2c06d049ce203ead40181a917f46a9164a0f3e058c3"} Mar 18 20:18:34 crc kubenswrapper[4950]: I0318 20:18:34.890948 4950 scope.go:117] "RemoveContainer" containerID="01edc7246d617f13699dceaa220ac50fb8349729df73606eb00a73d4e3ae8f6a" Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.705284 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-njln9"] Mar 18 20:18:45 crc kubenswrapper[4950]: E0318 20:18:45.705997 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f375d3f-628f-47a7-b2e5-d8afe6859aa8" containerName="oc" Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.706009 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f375d3f-628f-47a7-b2e5-d8afe6859aa8" containerName="oc" Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.706103 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f375d3f-628f-47a7-b2e5-d8afe6859aa8" containerName="oc" Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.706519 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-njln9" Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.709902 4950 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-rfj4w" Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.709965 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.710002 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.718529 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-9pk5n"] Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.719923 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-9pk5n" Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.722880 4950 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-kwkd8" Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.730391 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-njln9"] Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.734666 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-9pk5n"] Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.764015 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-9gqqw"] Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.767158 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-9gqqw" Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.778236 4950 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-mxz7l" Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.792508 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlclr\" (UniqueName: \"kubernetes.io/projected/7eb3a8a5-cac1-4a6f-8dba-2cc19867e55c-kube-api-access-wlclr\") pod \"cert-manager-webhook-687f57d79b-9gqqw\" (UID: \"7eb3a8a5-cac1-4a6f-8dba-2cc19867e55c\") " pod="cert-manager/cert-manager-webhook-687f57d79b-9gqqw" Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.792547 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drznj\" (UniqueName: \"kubernetes.io/projected/5be9c32f-9b12-4d42-8dd1-3d08455965cc-kube-api-access-drznj\") pod \"cert-manager-858654f9db-9pk5n\" (UID: \"5be9c32f-9b12-4d42-8dd1-3d08455965cc\") " pod="cert-manager/cert-manager-858654f9db-9pk5n" Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.792604 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h98ph\" (UniqueName: \"kubernetes.io/projected/f1979115-0bc1-4a2a-9973-72fc24aa570e-kube-api-access-h98ph\") pod \"cert-manager-cainjector-cf98fcc89-njln9\" (UID: \"f1979115-0bc1-4a2a-9973-72fc24aa570e\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-njln9" Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.795049 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-9gqqw"] Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.893553 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h98ph\" (UniqueName: \"kubernetes.io/projected/f1979115-0bc1-4a2a-9973-72fc24aa570e-kube-api-access-h98ph\") pod \"cert-manager-cainjector-cf98fcc89-njln9\" (UID: \"f1979115-0bc1-4a2a-9973-72fc24aa570e\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-njln9" Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.893825 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlclr\" (UniqueName: \"kubernetes.io/projected/7eb3a8a5-cac1-4a6f-8dba-2cc19867e55c-kube-api-access-wlclr\") pod \"cert-manager-webhook-687f57d79b-9gqqw\" (UID: \"7eb3a8a5-cac1-4a6f-8dba-2cc19867e55c\") " pod="cert-manager/cert-manager-webhook-687f57d79b-9gqqw" Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.893912 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drznj\" (UniqueName: \"kubernetes.io/projected/5be9c32f-9b12-4d42-8dd1-3d08455965cc-kube-api-access-drznj\") pod \"cert-manager-858654f9db-9pk5n\" (UID: \"5be9c32f-9b12-4d42-8dd1-3d08455965cc\") " pod="cert-manager/cert-manager-858654f9db-9pk5n" Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.913848 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drznj\" (UniqueName: \"kubernetes.io/projected/5be9c32f-9b12-4d42-8dd1-3d08455965cc-kube-api-access-drznj\") pod \"cert-manager-858654f9db-9pk5n\" (UID: \"5be9c32f-9b12-4d42-8dd1-3d08455965cc\") " pod="cert-manager/cert-manager-858654f9db-9pk5n" Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.919780 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlclr\" (UniqueName: \"kubernetes.io/projected/7eb3a8a5-cac1-4a6f-8dba-2cc19867e55c-kube-api-access-wlclr\") pod \"cert-manager-webhook-687f57d79b-9gqqw\" (UID: \"7eb3a8a5-cac1-4a6f-8dba-2cc19867e55c\") " pod="cert-manager/cert-manager-webhook-687f57d79b-9gqqw" Mar 18 20:18:45 crc kubenswrapper[4950]: I0318 20:18:45.919837 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h98ph\" (UniqueName: \"kubernetes.io/projected/f1979115-0bc1-4a2a-9973-72fc24aa570e-kube-api-access-h98ph\") pod \"cert-manager-cainjector-cf98fcc89-njln9\" (UID: \"f1979115-0bc1-4a2a-9973-72fc24aa570e\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-njln9" Mar 18 20:18:46 crc kubenswrapper[4950]: I0318 20:18:46.027338 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-njln9" Mar 18 20:18:46 crc kubenswrapper[4950]: I0318 20:18:46.034800 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-9pk5n" Mar 18 20:18:46 crc kubenswrapper[4950]: I0318 20:18:46.088164 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-9gqqw" Mar 18 20:18:46 crc kubenswrapper[4950]: I0318 20:18:46.251079 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-njln9"] Mar 18 20:18:46 crc kubenswrapper[4950]: W0318 20:18:46.254585 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1979115_0bc1_4a2a_9973_72fc24aa570e.slice/crio-36611b8035cfe8c1e4b5132cc4201cf2dbf70343570730fdb32ea21bdd92a3e8 WatchSource:0}: Error finding container 36611b8035cfe8c1e4b5132cc4201cf2dbf70343570730fdb32ea21bdd92a3e8: Status 404 returned error can't find the container with id 36611b8035cfe8c1e4b5132cc4201cf2dbf70343570730fdb32ea21bdd92a3e8 Mar 18 20:18:46 crc kubenswrapper[4950]: I0318 20:18:46.303647 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-9pk5n"] Mar 18 20:18:46 crc kubenswrapper[4950]: W0318 20:18:46.306471 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5be9c32f_9b12_4d42_8dd1_3d08455965cc.slice/crio-e1072610bdf4a5835e15d7d6c2d91e38ea5530a9acb45e5b5f8533c70264ff2f WatchSource:0}: Error finding container e1072610bdf4a5835e15d7d6c2d91e38ea5530a9acb45e5b5f8533c70264ff2f: Status 404 returned error can't find the container with id e1072610bdf4a5835e15d7d6c2d91e38ea5530a9acb45e5b5f8533c70264ff2f Mar 18 20:18:46 crc kubenswrapper[4950]: W0318 20:18:46.544898 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7eb3a8a5_cac1_4a6f_8dba_2cc19867e55c.slice/crio-0168197527e222897fb4ea1ce6f0deb63f2bf309e924343d11fa643a08efab29 WatchSource:0}: Error finding container 0168197527e222897fb4ea1ce6f0deb63f2bf309e924343d11fa643a08efab29: Status 404 returned error can't find the container with id 0168197527e222897fb4ea1ce6f0deb63f2bf309e924343d11fa643a08efab29 Mar 18 20:18:46 crc kubenswrapper[4950]: I0318 20:18:46.546489 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-9gqqw"] Mar 18 20:18:46 crc kubenswrapper[4950]: I0318 20:18:46.971010 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-9pk5n" event={"ID":"5be9c32f-9b12-4d42-8dd1-3d08455965cc","Type":"ContainerStarted","Data":"e1072610bdf4a5835e15d7d6c2d91e38ea5530a9acb45e5b5f8533c70264ff2f"} Mar 18 20:18:46 crc kubenswrapper[4950]: I0318 20:18:46.971925 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-9gqqw" event={"ID":"7eb3a8a5-cac1-4a6f-8dba-2cc19867e55c","Type":"ContainerStarted","Data":"0168197527e222897fb4ea1ce6f0deb63f2bf309e924343d11fa643a08efab29"} Mar 18 20:18:46 crc kubenswrapper[4950]: I0318 20:18:46.972713 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-njln9" event={"ID":"f1979115-0bc1-4a2a-9973-72fc24aa570e","Type":"ContainerStarted","Data":"36611b8035cfe8c1e4b5132cc4201cf2dbf70343570730fdb32ea21bdd92a3e8"} Mar 18 20:18:49 crc kubenswrapper[4950]: I0318 20:18:49.990752 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-9pk5n" event={"ID":"5be9c32f-9b12-4d42-8dd1-3d08455965cc","Type":"ContainerStarted","Data":"2e6ede5d2c56c25807c8944e0a1f8c785fbf1db20f5709727acd8561b51445de"} Mar 18 20:18:49 crc kubenswrapper[4950]: I0318 20:18:49.993426 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-njln9" event={"ID":"f1979115-0bc1-4a2a-9973-72fc24aa570e","Type":"ContainerStarted","Data":"8e8033c95cd1f0fc07d57d15ef484f680374446f6b28556a3884936c825a1a58"} Mar 18 20:18:50 crc kubenswrapper[4950]: I0318 20:18:50.022024 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-9pk5n" podStartSLOduration=2.13416861 podStartE2EDuration="5.021991378s" podCreationTimestamp="2026-03-18 20:18:45 +0000 UTC" firstStartedPulling="2026-03-18 20:18:46.311374099 +0000 UTC m=+739.552215967" lastFinishedPulling="2026-03-18 20:18:49.199196867 +0000 UTC m=+742.440038735" observedRunningTime="2026-03-18 20:18:50.013319581 +0000 UTC m=+743.254161469" watchObservedRunningTime="2026-03-18 20:18:50.021991378 +0000 UTC m=+743.262833256" Mar 18 20:18:50 crc kubenswrapper[4950]: I0318 20:18:50.036462 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-njln9" podStartSLOduration=2.153404445 podStartE2EDuration="5.036440538s" podCreationTimestamp="2026-03-18 20:18:45 +0000 UTC" firstStartedPulling="2026-03-18 20:18:46.256892638 +0000 UTC m=+739.497734496" lastFinishedPulling="2026-03-18 20:18:49.139928711 +0000 UTC m=+742.380770589" observedRunningTime="2026-03-18 20:18:50.03461668 +0000 UTC m=+743.275458568" watchObservedRunningTime="2026-03-18 20:18:50.036440538 +0000 UTC m=+743.277282406" Mar 18 20:18:51 crc kubenswrapper[4950]: I0318 20:18:51.004100 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-9gqqw" event={"ID":"7eb3a8a5-cac1-4a6f-8dba-2cc19867e55c","Type":"ContainerStarted","Data":"bd15bd19db49005021afce985c015ffb3cf9460c839ef9873df64cd308241120"} Mar 18 20:18:51 crc kubenswrapper[4950]: I0318 20:18:51.032954 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-9gqqw" podStartSLOduration=2.371436212 podStartE2EDuration="6.032928072s" podCreationTimestamp="2026-03-18 20:18:45 +0000 UTC" firstStartedPulling="2026-03-18 20:18:46.547641714 +0000 UTC m=+739.788483622" lastFinishedPulling="2026-03-18 20:18:50.209133614 +0000 UTC m=+743.449975482" observedRunningTime="2026-03-18 20:18:51.030633762 +0000 UTC m=+744.271475670" watchObservedRunningTime="2026-03-18 20:18:51.032928072 +0000 UTC m=+744.273769970" Mar 18 20:18:51 crc kubenswrapper[4950]: I0318 20:18:51.088516 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-9gqqw" Mar 18 20:18:55 crc kubenswrapper[4950]: I0318 20:18:55.631731 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-p2zdp"] Mar 18 20:18:55 crc kubenswrapper[4950]: I0318 20:18:55.633785 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovn-controller" containerID="cri-o://6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2" gracePeriod=30 Mar 18 20:18:55 crc kubenswrapper[4950]: I0318 20:18:55.633873 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="nbdb" containerID="cri-o://ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8" gracePeriod=30 Mar 18 20:18:55 crc kubenswrapper[4950]: I0318 20:18:55.634105 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="sbdb" containerID="cri-o://50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc" gracePeriod=30 Mar 18 20:18:55 crc kubenswrapper[4950]: I0318 20:18:55.634312 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="kube-rbac-proxy-node" containerID="cri-o://6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1" gracePeriod=30 Mar 18 20:18:55 crc kubenswrapper[4950]: I0318 20:18:55.634450 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="northd" containerID="cri-o://59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f" gracePeriod=30 Mar 18 20:18:55 crc kubenswrapper[4950]: I0318 20:18:55.634521 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8" gracePeriod=30 Mar 18 20:18:55 crc kubenswrapper[4950]: I0318 20:18:55.634637 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovn-acl-logging" containerID="cri-o://8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61" gracePeriod=30 Mar 18 20:18:55 crc kubenswrapper[4950]: I0318 20:18:55.689169 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovnkube-controller" containerID="cri-o://c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d" gracePeriod=30 Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.048618 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lqkxm_981403b4-25e5-4ea0-9267-9584719853d6/kube-multus/2.log" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.049162 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lqkxm_981403b4-25e5-4ea0-9267-9584719853d6/kube-multus/1.log" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.049211 4950 generic.go:334] "Generic (PLEG): container finished" podID="981403b4-25e5-4ea0-9267-9584719853d6" containerID="25e6ec3b99e4fad5368ef138e2764c1c0dab606df101d76849a96e45014cf62c" exitCode=2 Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.049265 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lqkxm" event={"ID":"981403b4-25e5-4ea0-9267-9584719853d6","Type":"ContainerDied","Data":"25e6ec3b99e4fad5368ef138e2764c1c0dab606df101d76849a96e45014cf62c"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.049305 4950 scope.go:117] "RemoveContainer" containerID="8b42f10f95ebda7760ce304ba8c5dfbf0e891059750a239f5ea8d0c96b673165" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.049886 4950 scope.go:117] "RemoveContainer" containerID="25e6ec3b99e4fad5368ef138e2764c1c0dab606df101d76849a96e45014cf62c" Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.050114 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-lqkxm_openshift-multus(981403b4-25e5-4ea0-9267-9584719853d6)\"" pod="openshift-multus/multus-lqkxm" podUID="981403b4-25e5-4ea0-9267-9584719853d6" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.053212 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p2zdp_83ee6659-cfde-4844-bb92-ccb9b26451ea/ovnkube-controller/3.log" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.059317 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p2zdp_83ee6659-cfde-4844-bb92-ccb9b26451ea/ovnkube-controller/3.log" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.059679 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p2zdp_83ee6659-cfde-4844-bb92-ccb9b26451ea/ovn-acl-logging/0.log" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.060388 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p2zdp_83ee6659-cfde-4844-bb92-ccb9b26451ea/ovn-controller/0.log" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.061312 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p2zdp_83ee6659-cfde-4844-bb92-ccb9b26451ea/ovn-acl-logging/0.log" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.061339 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062005 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p2zdp_83ee6659-cfde-4844-bb92-ccb9b26451ea/ovn-controller/0.log" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062442 4950 generic.go:334] "Generic (PLEG): container finished" podID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerID="c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d" exitCode=0 Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062474 4950 generic.go:334] "Generic (PLEG): container finished" podID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerID="50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc" exitCode=0 Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062485 4950 generic.go:334] "Generic (PLEG): container finished" podID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerID="ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8" exitCode=0 Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062497 4950 generic.go:334] "Generic (PLEG): container finished" podID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerID="59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f" exitCode=0 Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062509 4950 generic.go:334] "Generic (PLEG): container finished" podID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerID="38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8" exitCode=0 Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062521 4950 generic.go:334] "Generic (PLEG): container finished" podID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerID="6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1" exitCode=0 Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062507 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerDied","Data":"c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062578 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerDied","Data":"50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062596 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerDied","Data":"ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062534 4950 generic.go:334] "Generic (PLEG): container finished" podID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerID="8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61" exitCode=143 Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062626 4950 generic.go:334] "Generic (PLEG): container finished" podID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerID="6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2" exitCode=143 Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062616 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerDied","Data":"59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062684 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerDied","Data":"38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062708 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerDied","Data":"6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062725 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062741 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062749 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062756 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062763 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062772 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062779 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062787 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062794 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062802 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062812 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerDied","Data":"8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062824 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062833 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062841 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062848 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062855 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062862 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062869 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062876 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062883 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062890 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062900 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerDied","Data":"6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062912 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062920 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062927 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062934 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062941 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062948 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062956 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062964 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062971 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062978 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062987 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" event={"ID":"83ee6659-cfde-4844-bb92-ccb9b26451ea","Type":"ContainerDied","Data":"64ef4fd94ed84bce6bf75deb94b086c35f39f0a5c352ac110845beb6153ade3c"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.062997 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.063007 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.063014 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.063021 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.063029 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.063036 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.063043 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.063050 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.063057 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.063064 4950 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309"} Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.092117 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-9gqqw" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.096859 4950 scope.go:117] "RemoveContainer" containerID="c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.126676 4950 scope.go:117] "RemoveContainer" containerID="544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.153850 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bnfvb"] Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.154119 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="kubecfg-setup" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154142 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="kubecfg-setup" Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.154153 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovnkube-controller" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154160 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovnkube-controller" Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.154167 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovnkube-controller" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154173 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovnkube-controller" Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.154183 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="kube-rbac-proxy-ovn-metrics" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154188 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="kube-rbac-proxy-ovn-metrics" Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.154196 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="nbdb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154203 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="nbdb" Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.154212 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="northd" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154219 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="northd" Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.154231 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovnkube-controller" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154238 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovnkube-controller" Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.154247 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="kube-rbac-proxy-node" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154254 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="kube-rbac-proxy-node" Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.154261 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovnkube-controller" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154266 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovnkube-controller" Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.154275 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovn-controller" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154282 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovn-controller" Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.154293 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="sbdb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154304 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="sbdb" Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.154315 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovn-acl-logging" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154323 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovn-acl-logging" Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.154334 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovnkube-controller" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154342 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovnkube-controller" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154475 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="nbdb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154488 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="sbdb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154496 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovnkube-controller" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154506 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="northd" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154518 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="kube-rbac-proxy-ovn-metrics" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154531 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovnkube-controller" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154539 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovn-controller" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154549 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="kube-rbac-proxy-node" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154560 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovnkube-controller" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154568 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovn-acl-logging" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154779 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovnkube-controller" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.154791 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" containerName="ovnkube-controller" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.163893 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.169318 4950 scope.go:117] "RemoveContainer" containerID="50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.182629 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-cni-netd\") pod \"83ee6659-cfde-4844-bb92-ccb9b26451ea\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.182718 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/83ee6659-cfde-4844-bb92-ccb9b26451ea-ovnkube-script-lib\") pod \"83ee6659-cfde-4844-bb92-ccb9b26451ea\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.182749 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/83ee6659-cfde-4844-bb92-ccb9b26451ea-ovn-node-metrics-cert\") pod \"83ee6659-cfde-4844-bb92-ccb9b26451ea\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.182771 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-etc-openvswitch\") pod \"83ee6659-cfde-4844-bb92-ccb9b26451ea\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.182797 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-cni-bin\") pod \"83ee6659-cfde-4844-bb92-ccb9b26451ea\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.182823 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-var-lib-openvswitch\") pod \"83ee6659-cfde-4844-bb92-ccb9b26451ea\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.182854 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-kubelet\") pod \"83ee6659-cfde-4844-bb92-ccb9b26451ea\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.182877 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-run-netns\") pod \"83ee6659-cfde-4844-bb92-ccb9b26451ea\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.182898 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-run-systemd\") pod \"83ee6659-cfde-4844-bb92-ccb9b26451ea\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.182925 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-slash\") pod \"83ee6659-cfde-4844-bb92-ccb9b26451ea\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.182944 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-var-lib-cni-networks-ovn-kubernetes\") pod \"83ee6659-cfde-4844-bb92-ccb9b26451ea\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.182976 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/83ee6659-cfde-4844-bb92-ccb9b26451ea-ovnkube-config\") pod \"83ee6659-cfde-4844-bb92-ccb9b26451ea\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.182994 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-log-socket\") pod \"83ee6659-cfde-4844-bb92-ccb9b26451ea\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.183025 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-node-log\") pod \"83ee6659-cfde-4844-bb92-ccb9b26451ea\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.183048 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/83ee6659-cfde-4844-bb92-ccb9b26451ea-env-overrides\") pod \"83ee6659-cfde-4844-bb92-ccb9b26451ea\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.183089 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8qvm\" (UniqueName: \"kubernetes.io/projected/83ee6659-cfde-4844-bb92-ccb9b26451ea-kube-api-access-l8qvm\") pod \"83ee6659-cfde-4844-bb92-ccb9b26451ea\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.183116 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-run-ovn\") pod \"83ee6659-cfde-4844-bb92-ccb9b26451ea\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.183136 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-systemd-units\") pod \"83ee6659-cfde-4844-bb92-ccb9b26451ea\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.183182 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-run-ovn-kubernetes\") pod \"83ee6659-cfde-4844-bb92-ccb9b26451ea\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.183203 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-run-openvswitch\") pod \"83ee6659-cfde-4844-bb92-ccb9b26451ea\" (UID: \"83ee6659-cfde-4844-bb92-ccb9b26451ea\") " Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.183607 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "83ee6659-cfde-4844-bb92-ccb9b26451ea" (UID: "83ee6659-cfde-4844-bb92-ccb9b26451ea"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.183692 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "83ee6659-cfde-4844-bb92-ccb9b26451ea" (UID: "83ee6659-cfde-4844-bb92-ccb9b26451ea"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.184321 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83ee6659-cfde-4844-bb92-ccb9b26451ea-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "83ee6659-cfde-4844-bb92-ccb9b26451ea" (UID: "83ee6659-cfde-4844-bb92-ccb9b26451ea"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.184340 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "83ee6659-cfde-4844-bb92-ccb9b26451ea" (UID: "83ee6659-cfde-4844-bb92-ccb9b26451ea"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.184344 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83ee6659-cfde-4844-bb92-ccb9b26451ea-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "83ee6659-cfde-4844-bb92-ccb9b26451ea" (UID: "83ee6659-cfde-4844-bb92-ccb9b26451ea"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.184373 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "83ee6659-cfde-4844-bb92-ccb9b26451ea" (UID: "83ee6659-cfde-4844-bb92-ccb9b26451ea"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.184392 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-log-socket" (OuterVolumeSpecName: "log-socket") pod "83ee6659-cfde-4844-bb92-ccb9b26451ea" (UID: "83ee6659-cfde-4844-bb92-ccb9b26451ea"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.184468 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-node-log" (OuterVolumeSpecName: "node-log") pod "83ee6659-cfde-4844-bb92-ccb9b26451ea" (UID: "83ee6659-cfde-4844-bb92-ccb9b26451ea"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.184494 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "83ee6659-cfde-4844-bb92-ccb9b26451ea" (UID: "83ee6659-cfde-4844-bb92-ccb9b26451ea"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.184529 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "83ee6659-cfde-4844-bb92-ccb9b26451ea" (UID: "83ee6659-cfde-4844-bb92-ccb9b26451ea"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.184942 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "83ee6659-cfde-4844-bb92-ccb9b26451ea" (UID: "83ee6659-cfde-4844-bb92-ccb9b26451ea"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.184971 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "83ee6659-cfde-4844-bb92-ccb9b26451ea" (UID: "83ee6659-cfde-4844-bb92-ccb9b26451ea"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.184995 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "83ee6659-cfde-4844-bb92-ccb9b26451ea" (UID: "83ee6659-cfde-4844-bb92-ccb9b26451ea"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.185020 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "83ee6659-cfde-4844-bb92-ccb9b26451ea" (UID: "83ee6659-cfde-4844-bb92-ccb9b26451ea"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.186531 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "83ee6659-cfde-4844-bb92-ccb9b26451ea" (UID: "83ee6659-cfde-4844-bb92-ccb9b26451ea"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.186610 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-slash" (OuterVolumeSpecName: "host-slash") pod "83ee6659-cfde-4844-bb92-ccb9b26451ea" (UID: "83ee6659-cfde-4844-bb92-ccb9b26451ea"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.186847 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83ee6659-cfde-4844-bb92-ccb9b26451ea-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "83ee6659-cfde-4844-bb92-ccb9b26451ea" (UID: "83ee6659-cfde-4844-bb92-ccb9b26451ea"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.187954 4950 scope.go:117] "RemoveContainer" containerID="ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.192588 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83ee6659-cfde-4844-bb92-ccb9b26451ea-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "83ee6659-cfde-4844-bb92-ccb9b26451ea" (UID: "83ee6659-cfde-4844-bb92-ccb9b26451ea"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.197364 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83ee6659-cfde-4844-bb92-ccb9b26451ea-kube-api-access-l8qvm" (OuterVolumeSpecName: "kube-api-access-l8qvm") pod "83ee6659-cfde-4844-bb92-ccb9b26451ea" (UID: "83ee6659-cfde-4844-bb92-ccb9b26451ea"). InnerVolumeSpecName "kube-api-access-l8qvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.212116 4950 scope.go:117] "RemoveContainer" containerID="59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.216143 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "83ee6659-cfde-4844-bb92-ccb9b26451ea" (UID: "83ee6659-cfde-4844-bb92-ccb9b26451ea"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.226368 4950 scope.go:117] "RemoveContainer" containerID="38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.239167 4950 scope.go:117] "RemoveContainer" containerID="6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.251914 4950 scope.go:117] "RemoveContainer" containerID="8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.264199 4950 scope.go:117] "RemoveContainer" containerID="6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.277544 4950 scope.go:117] "RemoveContainer" containerID="046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286200 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-host-kubelet\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286261 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-var-lib-openvswitch\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286329 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-log-socket\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286365 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-host-run-ovn-kubernetes\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286404 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-host-slash\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286452 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ef26e798-ed73-449a-8c6e-26c518f1972d-ovnkube-config\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286479 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-systemd-units\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286510 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-host-cni-bin\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286537 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef26e798-ed73-449a-8c6e-26c518f1972d-env-overrides\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286576 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-run-systemd\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286604 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ef26e798-ed73-449a-8c6e-26c518f1972d-ovn-node-metrics-cert\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286630 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-node-log\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286657 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286692 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-host-cni-netd\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286746 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ef26e798-ed73-449a-8c6e-26c518f1972d-ovnkube-script-lib\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286773 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-host-run-netns\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286798 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbsxz\" (UniqueName: \"kubernetes.io/projected/ef26e798-ed73-449a-8c6e-26c518f1972d-kube-api-access-qbsxz\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286823 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-run-openvswitch\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286845 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-run-ovn\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286865 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-etc-openvswitch\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286919 4950 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286934 4950 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286945 4950 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286956 4950 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-slash\") on node \"crc\" DevicePath \"\"" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286969 4950 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286982 4950 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/83ee6659-cfde-4844-bb92-ccb9b26451ea-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.286996 4950 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-log-socket\") on node \"crc\" DevicePath \"\"" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.287008 4950 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-node-log\") on node \"crc\" DevicePath \"\"" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.287019 4950 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/83ee6659-cfde-4844-bb92-ccb9b26451ea-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.287031 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8qvm\" (UniqueName: \"kubernetes.io/projected/83ee6659-cfde-4844-bb92-ccb9b26451ea-kube-api-access-l8qvm\") on node \"crc\" DevicePath \"\"" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.287043 4950 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.287053 4950 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.287066 4950 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.287077 4950 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.287089 4950 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.287104 4950 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/83ee6659-cfde-4844-bb92-ccb9b26451ea-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.287116 4950 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/83ee6659-cfde-4844-bb92-ccb9b26451ea-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.287127 4950 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.287138 4950 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.287149 4950 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/83ee6659-cfde-4844-bb92-ccb9b26451ea-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.293311 4950 scope.go:117] "RemoveContainer" containerID="c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d" Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.294360 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d\": container with ID starting with c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d not found: ID does not exist" containerID="c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.294435 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d"} err="failed to get container status \"c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d\": rpc error: code = NotFound desc = could not find container \"c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d\": container with ID starting with c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.294473 4950 scope.go:117] "RemoveContainer" containerID="544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4" Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.294888 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4\": container with ID starting with 544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4 not found: ID does not exist" containerID="544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.295019 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4"} err="failed to get container status \"544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4\": rpc error: code = NotFound desc = could not find container \"544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4\": container with ID starting with 544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.295134 4950 scope.go:117] "RemoveContainer" containerID="50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc" Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.295586 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\": container with ID starting with 50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc not found: ID does not exist" containerID="50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.295622 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc"} err="failed to get container status \"50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\": rpc error: code = NotFound desc = could not find container \"50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\": container with ID starting with 50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.295645 4950 scope.go:117] "RemoveContainer" containerID="ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8" Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.295918 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\": container with ID starting with ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8 not found: ID does not exist" containerID="ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.296084 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8"} err="failed to get container status \"ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\": rpc error: code = NotFound desc = could not find container \"ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\": container with ID starting with ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.296190 4950 scope.go:117] "RemoveContainer" containerID="59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f" Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.296658 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\": container with ID starting with 59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f not found: ID does not exist" containerID="59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.296776 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f"} err="failed to get container status \"59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\": rpc error: code = NotFound desc = could not find container \"59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\": container with ID starting with 59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.296879 4950 scope.go:117] "RemoveContainer" containerID="38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8" Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.297222 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\": container with ID starting with 38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8 not found: ID does not exist" containerID="38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.297261 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8"} err="failed to get container status \"38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\": rpc error: code = NotFound desc = could not find container \"38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\": container with ID starting with 38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.297287 4950 scope.go:117] "RemoveContainer" containerID="6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1" Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.297797 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\": container with ID starting with 6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1 not found: ID does not exist" containerID="6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.297890 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1"} err="failed to get container status \"6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\": rpc error: code = NotFound desc = could not find container \"6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\": container with ID starting with 6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.297937 4950 scope.go:117] "RemoveContainer" containerID="8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61" Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.298309 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\": container with ID starting with 8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61 not found: ID does not exist" containerID="8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.298437 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61"} err="failed to get container status \"8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\": rpc error: code = NotFound desc = could not find container \"8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\": container with ID starting with 8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.298548 4950 scope.go:117] "RemoveContainer" containerID="6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2" Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.298987 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\": container with ID starting with 6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2 not found: ID does not exist" containerID="6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.299015 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2"} err="failed to get container status \"6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\": rpc error: code = NotFound desc = could not find container \"6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\": container with ID starting with 6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.299121 4950 scope.go:117] "RemoveContainer" containerID="046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309" Mar 18 20:18:56 crc kubenswrapper[4950]: E0318 20:18:56.299728 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\": container with ID starting with 046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309 not found: ID does not exist" containerID="046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.299755 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309"} err="failed to get container status \"046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\": rpc error: code = NotFound desc = could not find container \"046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\": container with ID starting with 046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.299772 4950 scope.go:117] "RemoveContainer" containerID="c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.300133 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d"} err="failed to get container status \"c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d\": rpc error: code = NotFound desc = could not find container \"c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d\": container with ID starting with c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.300391 4950 scope.go:117] "RemoveContainer" containerID="544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.300853 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4"} err="failed to get container status \"544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4\": rpc error: code = NotFound desc = could not find container \"544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4\": container with ID starting with 544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.300877 4950 scope.go:117] "RemoveContainer" containerID="50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.301180 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc"} err="failed to get container status \"50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\": rpc error: code = NotFound desc = could not find container \"50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\": container with ID starting with 50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.301318 4950 scope.go:117] "RemoveContainer" containerID="ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.301735 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8"} err="failed to get container status \"ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\": rpc error: code = NotFound desc = could not find container \"ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\": container with ID starting with ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.301764 4950 scope.go:117] "RemoveContainer" containerID="59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.302020 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f"} err="failed to get container status \"59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\": rpc error: code = NotFound desc = could not find container \"59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\": container with ID starting with 59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.302128 4950 scope.go:117] "RemoveContainer" containerID="38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.302569 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8"} err="failed to get container status \"38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\": rpc error: code = NotFound desc = could not find container \"38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\": container with ID starting with 38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.302593 4950 scope.go:117] "RemoveContainer" containerID="6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.302857 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1"} err="failed to get container status \"6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\": rpc error: code = NotFound desc = could not find container \"6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\": container with ID starting with 6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.302969 4950 scope.go:117] "RemoveContainer" containerID="8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.303458 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61"} err="failed to get container status \"8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\": rpc error: code = NotFound desc = could not find container \"8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\": container with ID starting with 8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.303484 4950 scope.go:117] "RemoveContainer" containerID="6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.303750 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2"} err="failed to get container status \"6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\": rpc error: code = NotFound desc = could not find container \"6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\": container with ID starting with 6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.303956 4950 scope.go:117] "RemoveContainer" containerID="046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.304290 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309"} err="failed to get container status \"046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\": rpc error: code = NotFound desc = could not find container \"046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\": container with ID starting with 046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.304317 4950 scope.go:117] "RemoveContainer" containerID="c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.304694 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d"} err="failed to get container status \"c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d\": rpc error: code = NotFound desc = could not find container \"c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d\": container with ID starting with c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.304830 4950 scope.go:117] "RemoveContainer" containerID="544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.305138 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4"} err="failed to get container status \"544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4\": rpc error: code = NotFound desc = could not find container \"544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4\": container with ID starting with 544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.305168 4950 scope.go:117] "RemoveContainer" containerID="50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.305536 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc"} err="failed to get container status \"50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\": rpc error: code = NotFound desc = could not find container \"50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\": container with ID starting with 50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.305574 4950 scope.go:117] "RemoveContainer" containerID="ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.305953 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8"} err="failed to get container status \"ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\": rpc error: code = NotFound desc = could not find container \"ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\": container with ID starting with ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.306135 4950 scope.go:117] "RemoveContainer" containerID="59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.306554 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f"} err="failed to get container status \"59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\": rpc error: code = NotFound desc = could not find container \"59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\": container with ID starting with 59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.306600 4950 scope.go:117] "RemoveContainer" containerID="38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.307052 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8"} err="failed to get container status \"38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\": rpc error: code = NotFound desc = could not find container \"38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\": container with ID starting with 38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.307102 4950 scope.go:117] "RemoveContainer" containerID="6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.307573 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1"} err="failed to get container status \"6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\": rpc error: code = NotFound desc = could not find container \"6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\": container with ID starting with 6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.307725 4950 scope.go:117] "RemoveContainer" containerID="8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.308199 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61"} err="failed to get container status \"8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\": rpc error: code = NotFound desc = could not find container \"8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\": container with ID starting with 8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.308232 4950 scope.go:117] "RemoveContainer" containerID="6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.308786 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2"} err="failed to get container status \"6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\": rpc error: code = NotFound desc = could not find container \"6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\": container with ID starting with 6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.308917 4950 scope.go:117] "RemoveContainer" containerID="046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.309303 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309"} err="failed to get container status \"046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\": rpc error: code = NotFound desc = could not find container \"046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\": container with ID starting with 046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.309329 4950 scope.go:117] "RemoveContainer" containerID="c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.309791 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d"} err="failed to get container status \"c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d\": rpc error: code = NotFound desc = could not find container \"c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d\": container with ID starting with c79e63827dd79e3b61d11748bfd9e38535d06be37b46201e737958f4c792399d not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.309974 4950 scope.go:117] "RemoveContainer" containerID="544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.310400 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4"} err="failed to get container status \"544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4\": rpc error: code = NotFound desc = could not find container \"544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4\": container with ID starting with 544cfbf2312645dc96798b07511768b995a9b4f5d1743a08106e3de611d181f4 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.310451 4950 scope.go:117] "RemoveContainer" containerID="50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.310830 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc"} err="failed to get container status \"50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\": rpc error: code = NotFound desc = could not find container \"50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc\": container with ID starting with 50f5ae20a7835d3cb7ca4b67604fe21de4c97587df957d765c62e4e56d47bcbc not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.310862 4950 scope.go:117] "RemoveContainer" containerID="ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.311295 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8"} err="failed to get container status \"ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\": rpc error: code = NotFound desc = could not find container \"ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8\": container with ID starting with ff3f1ee1c6252adb03649fc01926a0051f2e9133ce77054a4b0864d6e664a4e8 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.311320 4950 scope.go:117] "RemoveContainer" containerID="59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.311726 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f"} err="failed to get container status \"59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\": rpc error: code = NotFound desc = could not find container \"59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f\": container with ID starting with 59b0491518acf72db4525d3270751e8225ca8ecfa49349c95ea1f0adf7cc0f3f not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.311913 4950 scope.go:117] "RemoveContainer" containerID="38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.312358 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8"} err="failed to get container status \"38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\": rpc error: code = NotFound desc = could not find container \"38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8\": container with ID starting with 38a06d49bc7faf3f3f3b77a18159f106d4001a2bb0e67ae4d8bce5a7c5a576a8 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.312391 4950 scope.go:117] "RemoveContainer" containerID="6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.312676 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1"} err="failed to get container status \"6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\": rpc error: code = NotFound desc = could not find container \"6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1\": container with ID starting with 6763183a354eab4538181906cb34250b922fdd92dd3a0bc60ec7c183711ec3c1 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.312709 4950 scope.go:117] "RemoveContainer" containerID="8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.312973 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61"} err="failed to get container status \"8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\": rpc error: code = NotFound desc = could not find container \"8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61\": container with ID starting with 8d1c2ad81b89bc63c96bfe4e52f401f912f48696acf93e2bd23feeb068117a61 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.313005 4950 scope.go:117] "RemoveContainer" containerID="6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.313429 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2"} err="failed to get container status \"6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\": rpc error: code = NotFound desc = could not find container \"6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2\": container with ID starting with 6e66fb5753c06eaad98e34d3591041fe1250102acd67ad970793e8f9672abbd2 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.313470 4950 scope.go:117] "RemoveContainer" containerID="046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.313777 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309"} err="failed to get container status \"046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\": rpc error: code = NotFound desc = could not find container \"046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309\": container with ID starting with 046ba205d04f5185680a005e7f33f33fc5693a1f1c84f526720eb6e64982f309 not found: ID does not exist" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.388736 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-run-systemd\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.388791 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ef26e798-ed73-449a-8c6e-26c518f1972d-ovn-node-metrics-cert\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.388815 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-node-log\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.388845 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.388879 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-host-cni-netd\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.388899 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ef26e798-ed73-449a-8c6e-26c518f1972d-ovnkube-script-lib\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.388923 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-host-run-netns\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.388948 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbsxz\" (UniqueName: \"kubernetes.io/projected/ef26e798-ed73-449a-8c6e-26c518f1972d-kube-api-access-qbsxz\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.388952 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.388972 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-run-openvswitch\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.388998 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-run-ovn\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.389025 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-etc-openvswitch\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.389057 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-host-kubelet\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.389083 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-var-lib-openvswitch\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.389293 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-log-socket\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.389332 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-host-run-ovn-kubernetes\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.389363 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-host-slash\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.389388 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ef26e798-ed73-449a-8c6e-26c518f1972d-ovnkube-config\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.389389 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-node-log\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.389440 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-systemd-units\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.389508 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-systemd-units\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.389553 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-host-cni-netd\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.388914 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-run-systemd\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.389989 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-host-slash\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.390059 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-host-cni-bin\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.390059 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-run-openvswitch\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.390133 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-run-ovn\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.390131 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-host-kubelet\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.390154 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-etc-openvswitch\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.390166 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-host-run-ovn-kubernetes\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.390186 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-log-socket\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.390208 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-host-run-netns\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.390188 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef26e798-ed73-449a-8c6e-26c518f1972d-env-overrides\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.390232 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-host-cni-bin\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.390571 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ef26e798-ed73-449a-8c6e-26c518f1972d-var-lib-openvswitch\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.390967 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef26e798-ed73-449a-8c6e-26c518f1972d-env-overrides\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.391184 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ef26e798-ed73-449a-8c6e-26c518f1972d-ovnkube-script-lib\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.391704 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ef26e798-ed73-449a-8c6e-26c518f1972d-ovnkube-config\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.394794 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ef26e798-ed73-449a-8c6e-26c518f1972d-ovn-node-metrics-cert\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.406339 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbsxz\" (UniqueName: \"kubernetes.io/projected/ef26e798-ed73-449a-8c6e-26c518f1972d-kube-api-access-qbsxz\") pod \"ovnkube-node-bnfvb\" (UID: \"ef26e798-ed73-449a-8c6e-26c518f1972d\") " pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:56 crc kubenswrapper[4950]: I0318 20:18:56.492775 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:18:57 crc kubenswrapper[4950]: I0318 20:18:57.071376 4950 generic.go:334] "Generic (PLEG): container finished" podID="ef26e798-ed73-449a-8c6e-26c518f1972d" containerID="4955012d0bf5e0a68d69c1b17556ab02ce942fd79ab52ff887f5963a08c318b2" exitCode=0 Mar 18 20:18:57 crc kubenswrapper[4950]: I0318 20:18:57.071453 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" event={"ID":"ef26e798-ed73-449a-8c6e-26c518f1972d","Type":"ContainerDied","Data":"4955012d0bf5e0a68d69c1b17556ab02ce942fd79ab52ff887f5963a08c318b2"} Mar 18 20:18:57 crc kubenswrapper[4950]: I0318 20:18:57.071502 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" event={"ID":"ef26e798-ed73-449a-8c6e-26c518f1972d","Type":"ContainerStarted","Data":"5568a5c0bcc8d0ea48f96ea860f9525fa5e9d49e1e0eb0e9285f64462e3c2a3d"} Mar 18 20:18:57 crc kubenswrapper[4950]: I0318 20:18:57.073788 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lqkxm_981403b4-25e5-4ea0-9267-9584719853d6/kube-multus/2.log" Mar 18 20:18:57 crc kubenswrapper[4950]: I0318 20:18:57.075401 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-p2zdp" Mar 18 20:18:57 crc kubenswrapper[4950]: I0318 20:18:57.158480 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-p2zdp"] Mar 18 20:18:57 crc kubenswrapper[4950]: I0318 20:18:57.167476 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-p2zdp"] Mar 18 20:18:57 crc kubenswrapper[4950]: I0318 20:18:57.494362 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83ee6659-cfde-4844-bb92-ccb9b26451ea" path="/var/lib/kubelet/pods/83ee6659-cfde-4844-bb92-ccb9b26451ea/volumes" Mar 18 20:18:58 crc kubenswrapper[4950]: I0318 20:18:58.083665 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" event={"ID":"ef26e798-ed73-449a-8c6e-26c518f1972d","Type":"ContainerStarted","Data":"457492e14e7d094244ac0ab86c19e09d3f268633192823a6bdc2edc187e20cab"} Mar 18 20:18:58 crc kubenswrapper[4950]: I0318 20:18:58.083962 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" event={"ID":"ef26e798-ed73-449a-8c6e-26c518f1972d","Type":"ContainerStarted","Data":"b3ea3da48c932b188f169b1090f75e146b8dcc5a161841fe7ab56ce4df9259f0"} Mar 18 20:18:58 crc kubenswrapper[4950]: I0318 20:18:58.083975 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" event={"ID":"ef26e798-ed73-449a-8c6e-26c518f1972d","Type":"ContainerStarted","Data":"15edbdc6f3e88ecc669e80e5c76965340f8d5beba59c24c2b734f7491c038738"} Mar 18 20:18:58 crc kubenswrapper[4950]: I0318 20:18:58.083984 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" event={"ID":"ef26e798-ed73-449a-8c6e-26c518f1972d","Type":"ContainerStarted","Data":"a340bdd929be86e085ca4d76e0e2741b5304cd2bea5dd7f12d99d288e92340a8"} Mar 18 20:18:58 crc kubenswrapper[4950]: I0318 20:18:58.083992 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" event={"ID":"ef26e798-ed73-449a-8c6e-26c518f1972d","Type":"ContainerStarted","Data":"1847a5ba4febc33f946d9db73d958d02335ca9d3157afc451e978ae83a631129"} Mar 18 20:18:58 crc kubenswrapper[4950]: I0318 20:18:58.084000 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" event={"ID":"ef26e798-ed73-449a-8c6e-26c518f1972d","Type":"ContainerStarted","Data":"b3e9147074a5c016f18869909a0cae0cf3c4855f009ce601f92d44b21bbdbd50"} Mar 18 20:19:01 crc kubenswrapper[4950]: I0318 20:19:01.107667 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" event={"ID":"ef26e798-ed73-449a-8c6e-26c518f1972d","Type":"ContainerStarted","Data":"8a2ee16e3d37bdf6a237becfbae24c0250a833b0e543acf75aef917bfadb24d2"} Mar 18 20:19:03 crc kubenswrapper[4950]: I0318 20:19:03.122761 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" event={"ID":"ef26e798-ed73-449a-8c6e-26c518f1972d","Type":"ContainerStarted","Data":"be1f1d3a0fc5d34a2175f70fb4659cba159f8f7c9b66feaa0c86b5cae0ed9d9f"} Mar 18 20:19:03 crc kubenswrapper[4950]: I0318 20:19:03.123174 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:19:03 crc kubenswrapper[4950]: I0318 20:19:03.123219 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:19:03 crc kubenswrapper[4950]: I0318 20:19:03.163348 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" podStartSLOduration=7.163322713 podStartE2EDuration="7.163322713s" podCreationTimestamp="2026-03-18 20:18:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:19:03.155399185 +0000 UTC m=+756.396241103" watchObservedRunningTime="2026-03-18 20:19:03.163322713 +0000 UTC m=+756.404164601" Mar 18 20:19:03 crc kubenswrapper[4950]: I0318 20:19:03.195215 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:19:04 crc kubenswrapper[4950]: I0318 20:19:04.131846 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:19:04 crc kubenswrapper[4950]: I0318 20:19:04.184695 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:19:10 crc kubenswrapper[4950]: I0318 20:19:10.480599 4950 scope.go:117] "RemoveContainer" containerID="25e6ec3b99e4fad5368ef138e2764c1c0dab606df101d76849a96e45014cf62c" Mar 18 20:19:10 crc kubenswrapper[4950]: E0318 20:19:10.481614 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-lqkxm_openshift-multus(981403b4-25e5-4ea0-9267-9584719853d6)\"" pod="openshift-multus/multus-lqkxm" podUID="981403b4-25e5-4ea0-9267-9584719853d6" Mar 18 20:19:22 crc kubenswrapper[4950]: I0318 20:19:22.482543 4950 scope.go:117] "RemoveContainer" containerID="25e6ec3b99e4fad5368ef138e2764c1c0dab606df101d76849a96e45014cf62c" Mar 18 20:19:23 crc kubenswrapper[4950]: I0318 20:19:23.265456 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lqkxm_981403b4-25e5-4ea0-9267-9584719853d6/kube-multus/2.log" Mar 18 20:19:23 crc kubenswrapper[4950]: I0318 20:19:23.265829 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lqkxm" event={"ID":"981403b4-25e5-4ea0-9267-9584719853d6","Type":"ContainerStarted","Data":"4dadee84bf3e3604ec609dc1c78a18675f1888e422959e135663ae670e8aebd2"} Mar 18 20:19:26 crc kubenswrapper[4950]: I0318 20:19:26.533635 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bnfvb" Mar 18 20:19:35 crc kubenswrapper[4950]: I0318 20:19:35.228865 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k"] Mar 18 20:19:35 crc kubenswrapper[4950]: I0318 20:19:35.231964 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k" Mar 18 20:19:35 crc kubenswrapper[4950]: I0318 20:19:35.239311 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 18 20:19:35 crc kubenswrapper[4950]: I0318 20:19:35.242984 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k"] Mar 18 20:19:35 crc kubenswrapper[4950]: I0318 20:19:35.294694 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/534d37ef-a537-4ea3-832e-4609580d9f47-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k\" (UID: \"534d37ef-a537-4ea3-832e-4609580d9f47\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k" Mar 18 20:19:35 crc kubenswrapper[4950]: I0318 20:19:35.294891 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/534d37ef-a537-4ea3-832e-4609580d9f47-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k\" (UID: \"534d37ef-a537-4ea3-832e-4609580d9f47\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k" Mar 18 20:19:35 crc kubenswrapper[4950]: I0318 20:19:35.294978 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2b4w\" (UniqueName: \"kubernetes.io/projected/534d37ef-a537-4ea3-832e-4609580d9f47-kube-api-access-p2b4w\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k\" (UID: \"534d37ef-a537-4ea3-832e-4609580d9f47\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k" Mar 18 20:19:35 crc kubenswrapper[4950]: I0318 20:19:35.396301 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/534d37ef-a537-4ea3-832e-4609580d9f47-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k\" (UID: \"534d37ef-a537-4ea3-832e-4609580d9f47\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k" Mar 18 20:19:35 crc kubenswrapper[4950]: I0318 20:19:35.396660 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2b4w\" (UniqueName: \"kubernetes.io/projected/534d37ef-a537-4ea3-832e-4609580d9f47-kube-api-access-p2b4w\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k\" (UID: \"534d37ef-a537-4ea3-832e-4609580d9f47\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k" Mar 18 20:19:35 crc kubenswrapper[4950]: I0318 20:19:35.396792 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/534d37ef-a537-4ea3-832e-4609580d9f47-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k\" (UID: \"534d37ef-a537-4ea3-832e-4609580d9f47\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k" Mar 18 20:19:35 crc kubenswrapper[4950]: I0318 20:19:35.396900 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/534d37ef-a537-4ea3-832e-4609580d9f47-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k\" (UID: \"534d37ef-a537-4ea3-832e-4609580d9f47\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k" Mar 18 20:19:35 crc kubenswrapper[4950]: I0318 20:19:35.397567 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/534d37ef-a537-4ea3-832e-4609580d9f47-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k\" (UID: \"534d37ef-a537-4ea3-832e-4609580d9f47\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k" Mar 18 20:19:35 crc kubenswrapper[4950]: I0318 20:19:35.429641 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2b4w\" (UniqueName: \"kubernetes.io/projected/534d37ef-a537-4ea3-832e-4609580d9f47-kube-api-access-p2b4w\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k\" (UID: \"534d37ef-a537-4ea3-832e-4609580d9f47\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k" Mar 18 20:19:35 crc kubenswrapper[4950]: I0318 20:19:35.553871 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k" Mar 18 20:19:36 crc kubenswrapper[4950]: I0318 20:19:36.010196 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k"] Mar 18 20:19:36 crc kubenswrapper[4950]: I0318 20:19:36.351815 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k" event={"ID":"534d37ef-a537-4ea3-832e-4609580d9f47","Type":"ContainerStarted","Data":"06b7482ff40f93f7086f3bd5a1d32471c85f60873651d0c97117fa4bd5495a12"} Mar 18 20:19:36 crc kubenswrapper[4950]: I0318 20:19:36.351893 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k" event={"ID":"534d37ef-a537-4ea3-832e-4609580d9f47","Type":"ContainerStarted","Data":"fc33a074a000070c30d2fce6832092dbb8a509600de9ddc701e38604bc3424cf"} Mar 18 20:19:37 crc kubenswrapper[4950]: I0318 20:19:37.362857 4950 generic.go:334] "Generic (PLEG): container finished" podID="534d37ef-a537-4ea3-832e-4609580d9f47" containerID="06b7482ff40f93f7086f3bd5a1d32471c85f60873651d0c97117fa4bd5495a12" exitCode=0 Mar 18 20:19:37 crc kubenswrapper[4950]: I0318 20:19:37.362927 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k" event={"ID":"534d37ef-a537-4ea3-832e-4609580d9f47","Type":"ContainerDied","Data":"06b7482ff40f93f7086f3bd5a1d32471c85f60873651d0c97117fa4bd5495a12"} Mar 18 20:19:39 crc kubenswrapper[4950]: I0318 20:19:39.381803 4950 generic.go:334] "Generic (PLEG): container finished" podID="534d37ef-a537-4ea3-832e-4609580d9f47" containerID="5d74cbf23548158b142b9a88a839482be1915e004893b65d047de77f425dda50" exitCode=0 Mar 18 20:19:39 crc kubenswrapper[4950]: I0318 20:19:39.381851 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k" event={"ID":"534d37ef-a537-4ea3-832e-4609580d9f47","Type":"ContainerDied","Data":"5d74cbf23548158b142b9a88a839482be1915e004893b65d047de77f425dda50"} Mar 18 20:19:40 crc kubenswrapper[4950]: I0318 20:19:40.392005 4950 generic.go:334] "Generic (PLEG): container finished" podID="534d37ef-a537-4ea3-832e-4609580d9f47" containerID="0892ba5d438ced4e83c0dc22fbc7ea3cbc13a33f9c6661ba14c0d86e589a5424" exitCode=0 Mar 18 20:19:40 crc kubenswrapper[4950]: I0318 20:19:40.392069 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k" event={"ID":"534d37ef-a537-4ea3-832e-4609580d9f47","Type":"ContainerDied","Data":"0892ba5d438ced4e83c0dc22fbc7ea3cbc13a33f9c6661ba14c0d86e589a5424"} Mar 18 20:19:41 crc kubenswrapper[4950]: I0318 20:19:41.665975 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k" Mar 18 20:19:41 crc kubenswrapper[4950]: I0318 20:19:41.786226 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/534d37ef-a537-4ea3-832e-4609580d9f47-bundle\") pod \"534d37ef-a537-4ea3-832e-4609580d9f47\" (UID: \"534d37ef-a537-4ea3-832e-4609580d9f47\") " Mar 18 20:19:41 crc kubenswrapper[4950]: I0318 20:19:41.786299 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/534d37ef-a537-4ea3-832e-4609580d9f47-util\") pod \"534d37ef-a537-4ea3-832e-4609580d9f47\" (UID: \"534d37ef-a537-4ea3-832e-4609580d9f47\") " Mar 18 20:19:41 crc kubenswrapper[4950]: I0318 20:19:41.786319 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2b4w\" (UniqueName: \"kubernetes.io/projected/534d37ef-a537-4ea3-832e-4609580d9f47-kube-api-access-p2b4w\") pod \"534d37ef-a537-4ea3-832e-4609580d9f47\" (UID: \"534d37ef-a537-4ea3-832e-4609580d9f47\") " Mar 18 20:19:41 crc kubenswrapper[4950]: I0318 20:19:41.787041 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/534d37ef-a537-4ea3-832e-4609580d9f47-bundle" (OuterVolumeSpecName: "bundle") pod "534d37ef-a537-4ea3-832e-4609580d9f47" (UID: "534d37ef-a537-4ea3-832e-4609580d9f47"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:19:41 crc kubenswrapper[4950]: I0318 20:19:41.796445 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/534d37ef-a537-4ea3-832e-4609580d9f47-kube-api-access-p2b4w" (OuterVolumeSpecName: "kube-api-access-p2b4w") pod "534d37ef-a537-4ea3-832e-4609580d9f47" (UID: "534d37ef-a537-4ea3-832e-4609580d9f47"). InnerVolumeSpecName "kube-api-access-p2b4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:19:41 crc kubenswrapper[4950]: I0318 20:19:41.856250 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/534d37ef-a537-4ea3-832e-4609580d9f47-util" (OuterVolumeSpecName: "util") pod "534d37ef-a537-4ea3-832e-4609580d9f47" (UID: "534d37ef-a537-4ea3-832e-4609580d9f47"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:19:41 crc kubenswrapper[4950]: I0318 20:19:41.888052 4950 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/534d37ef-a537-4ea3-832e-4609580d9f47-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:19:41 crc kubenswrapper[4950]: I0318 20:19:41.888083 4950 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/534d37ef-a537-4ea3-832e-4609580d9f47-util\") on node \"crc\" DevicePath \"\"" Mar 18 20:19:41 crc kubenswrapper[4950]: I0318 20:19:41.888092 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2b4w\" (UniqueName: \"kubernetes.io/projected/534d37ef-a537-4ea3-832e-4609580d9f47-kube-api-access-p2b4w\") on node \"crc\" DevicePath \"\"" Mar 18 20:19:42 crc kubenswrapper[4950]: I0318 20:19:42.410363 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k" event={"ID":"534d37ef-a537-4ea3-832e-4609580d9f47","Type":"ContainerDied","Data":"fc33a074a000070c30d2fce6832092dbb8a509600de9ddc701e38604bc3424cf"} Mar 18 20:19:42 crc kubenswrapper[4950]: I0318 20:19:42.410661 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc33a074a000070c30d2fce6832092dbb8a509600de9ddc701e38604bc3424cf" Mar 18 20:19:42 crc kubenswrapper[4950]: I0318 20:19:42.410479 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k" Mar 18 20:19:44 crc kubenswrapper[4950]: I0318 20:19:44.388604 4950 scope.go:117] "RemoveContainer" containerID="21071767903abe4cc9037d6899818f503456804578d981c8dcfc413f8ec77034" Mar 18 20:19:46 crc kubenswrapper[4950]: I0318 20:19:46.822321 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-phx52"] Mar 18 20:19:46 crc kubenswrapper[4950]: E0318 20:19:46.823631 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="534d37ef-a537-4ea3-832e-4609580d9f47" containerName="pull" Mar 18 20:19:46 crc kubenswrapper[4950]: I0318 20:19:46.823714 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="534d37ef-a537-4ea3-832e-4609580d9f47" containerName="pull" Mar 18 20:19:46 crc kubenswrapper[4950]: E0318 20:19:46.823773 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="534d37ef-a537-4ea3-832e-4609580d9f47" containerName="util" Mar 18 20:19:46 crc kubenswrapper[4950]: I0318 20:19:46.823831 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="534d37ef-a537-4ea3-832e-4609580d9f47" containerName="util" Mar 18 20:19:46 crc kubenswrapper[4950]: E0318 20:19:46.823904 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="534d37ef-a537-4ea3-832e-4609580d9f47" containerName="extract" Mar 18 20:19:46 crc kubenswrapper[4950]: I0318 20:19:46.823954 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="534d37ef-a537-4ea3-832e-4609580d9f47" containerName="extract" Mar 18 20:19:46 crc kubenswrapper[4950]: I0318 20:19:46.824096 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="534d37ef-a537-4ea3-832e-4609580d9f47" containerName="extract" Mar 18 20:19:46 crc kubenswrapper[4950]: I0318 20:19:46.824534 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-phx52" Mar 18 20:19:46 crc kubenswrapper[4950]: I0318 20:19:46.826719 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-zkgqf" Mar 18 20:19:46 crc kubenswrapper[4950]: I0318 20:19:46.826751 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Mar 18 20:19:46 crc kubenswrapper[4950]: I0318 20:19:46.827697 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Mar 18 20:19:46 crc kubenswrapper[4950]: I0318 20:19:46.839126 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-phx52"] Mar 18 20:19:46 crc kubenswrapper[4950]: I0318 20:19:46.956145 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2t24\" (UniqueName: \"kubernetes.io/projected/de65b778-0bca-479c-826c-e7d935a294c4-kube-api-access-k2t24\") pod \"nmstate-operator-796d4cfff4-phx52\" (UID: \"de65b778-0bca-479c-826c-e7d935a294c4\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-phx52" Mar 18 20:19:47 crc kubenswrapper[4950]: I0318 20:19:47.057533 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2t24\" (UniqueName: \"kubernetes.io/projected/de65b778-0bca-479c-826c-e7d935a294c4-kube-api-access-k2t24\") pod \"nmstate-operator-796d4cfff4-phx52\" (UID: \"de65b778-0bca-479c-826c-e7d935a294c4\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-phx52" Mar 18 20:19:47 crc kubenswrapper[4950]: I0318 20:19:47.088900 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2t24\" (UniqueName: \"kubernetes.io/projected/de65b778-0bca-479c-826c-e7d935a294c4-kube-api-access-k2t24\") pod \"nmstate-operator-796d4cfff4-phx52\" (UID: \"de65b778-0bca-479c-826c-e7d935a294c4\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-phx52" Mar 18 20:19:47 crc kubenswrapper[4950]: I0318 20:19:47.144938 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-phx52" Mar 18 20:19:47 crc kubenswrapper[4950]: I0318 20:19:47.396477 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-phx52"] Mar 18 20:19:47 crc kubenswrapper[4950]: I0318 20:19:47.451322 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-phx52" event={"ID":"de65b778-0bca-479c-826c-e7d935a294c4","Type":"ContainerStarted","Data":"2944215617fc38687ef25762a1f058855835abb9cfa2c5eec0ef4eff632fc20a"} Mar 18 20:19:50 crc kubenswrapper[4950]: I0318 20:19:50.483071 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-phx52" event={"ID":"de65b778-0bca-479c-826c-e7d935a294c4","Type":"ContainerStarted","Data":"499aa8c5d5c2f482d9e50212fc23ce5fe30be2a136ace1e19ec129d89cb6c4d8"} Mar 18 20:19:50 crc kubenswrapper[4950]: I0318 20:19:50.546760 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-796d4cfff4-phx52" podStartSLOduration=1.868871588 podStartE2EDuration="4.546741122s" podCreationTimestamp="2026-03-18 20:19:46 +0000 UTC" firstStartedPulling="2026-03-18 20:19:47.390610366 +0000 UTC m=+800.631452234" lastFinishedPulling="2026-03-18 20:19:50.06847986 +0000 UTC m=+803.309321768" observedRunningTime="2026-03-18 20:19:50.54364235 +0000 UTC m=+803.784484238" watchObservedRunningTime="2026-03-18 20:19:50.546741122 +0000 UTC m=+803.787583000" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.595920 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-gqrnm"] Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.597249 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-gqrnm" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.599953 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-8w4rc" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.605201 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-gqrnm"] Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.608519 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-64n89"] Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.609176 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-64n89" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.611257 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.621802 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-64n89"] Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.650127 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-6x6mb"] Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.650960 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-6x6mb" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.701514 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5xp9\" (UniqueName: \"kubernetes.io/projected/e1b75269-989a-489f-843b-94004e8cdf62-kube-api-access-p5xp9\") pod \"nmstate-metrics-9b8c8685d-gqrnm\" (UID: \"e1b75269-989a-489f-843b-94004e8cdf62\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-gqrnm" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.701561 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/452b41f3-a30b-4aec-a879-aff3872199d3-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-64n89\" (UID: \"452b41f3-a30b-4aec-a879-aff3872199d3\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-64n89" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.701599 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kjx4\" (UniqueName: \"kubernetes.io/projected/452b41f3-a30b-4aec-a879-aff3872199d3-kube-api-access-7kjx4\") pod \"nmstate-webhook-5f558f5558-64n89\" (UID: \"452b41f3-a30b-4aec-a879-aff3872199d3\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-64n89" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.763046 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-jnvkz"] Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.763843 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jnvkz" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.768486 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.768599 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-dqs27" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.768731 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.775772 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-jnvkz"] Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.803235 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/61d14d6f-1408-4309-90c9-2971b45d85dc-nmstate-lock\") pod \"nmstate-handler-6x6mb\" (UID: \"61d14d6f-1408-4309-90c9-2971b45d85dc\") " pod="openshift-nmstate/nmstate-handler-6x6mb" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.803295 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kjx4\" (UniqueName: \"kubernetes.io/projected/452b41f3-a30b-4aec-a879-aff3872199d3-kube-api-access-7kjx4\") pod \"nmstate-webhook-5f558f5558-64n89\" (UID: \"452b41f3-a30b-4aec-a879-aff3872199d3\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-64n89" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.803375 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/61d14d6f-1408-4309-90c9-2971b45d85dc-dbus-socket\") pod \"nmstate-handler-6x6mb\" (UID: \"61d14d6f-1408-4309-90c9-2971b45d85dc\") " pod="openshift-nmstate/nmstate-handler-6x6mb" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.803423 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjnkg\" (UniqueName: \"kubernetes.io/projected/61d14d6f-1408-4309-90c9-2971b45d85dc-kube-api-access-fjnkg\") pod \"nmstate-handler-6x6mb\" (UID: \"61d14d6f-1408-4309-90c9-2971b45d85dc\") " pod="openshift-nmstate/nmstate-handler-6x6mb" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.803605 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/61d14d6f-1408-4309-90c9-2971b45d85dc-ovs-socket\") pod \"nmstate-handler-6x6mb\" (UID: \"61d14d6f-1408-4309-90c9-2971b45d85dc\") " pod="openshift-nmstate/nmstate-handler-6x6mb" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.803696 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5xp9\" (UniqueName: \"kubernetes.io/projected/e1b75269-989a-489f-843b-94004e8cdf62-kube-api-access-p5xp9\") pod \"nmstate-metrics-9b8c8685d-gqrnm\" (UID: \"e1b75269-989a-489f-843b-94004e8cdf62\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-gqrnm" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.803732 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/452b41f3-a30b-4aec-a879-aff3872199d3-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-64n89\" (UID: \"452b41f3-a30b-4aec-a879-aff3872199d3\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-64n89" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.823484 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/452b41f3-a30b-4aec-a879-aff3872199d3-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-64n89\" (UID: \"452b41f3-a30b-4aec-a879-aff3872199d3\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-64n89" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.827186 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kjx4\" (UniqueName: \"kubernetes.io/projected/452b41f3-a30b-4aec-a879-aff3872199d3-kube-api-access-7kjx4\") pod \"nmstate-webhook-5f558f5558-64n89\" (UID: \"452b41f3-a30b-4aec-a879-aff3872199d3\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-64n89" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.828050 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5xp9\" (UniqueName: \"kubernetes.io/projected/e1b75269-989a-489f-843b-94004e8cdf62-kube-api-access-p5xp9\") pod \"nmstate-metrics-9b8c8685d-gqrnm\" (UID: \"e1b75269-989a-489f-843b-94004e8cdf62\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-gqrnm" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.904510 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/61d14d6f-1408-4309-90c9-2971b45d85dc-nmstate-lock\") pod \"nmstate-handler-6x6mb\" (UID: \"61d14d6f-1408-4309-90c9-2971b45d85dc\") " pod="openshift-nmstate/nmstate-handler-6x6mb" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.904561 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjg78\" (UniqueName: \"kubernetes.io/projected/0db507f0-dd6f-405c-ae17-4a9e9553cc50-kube-api-access-jjg78\") pod \"nmstate-console-plugin-86f58fcf4-jnvkz\" (UID: \"0db507f0-dd6f-405c-ae17-4a9e9553cc50\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jnvkz" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.904607 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/61d14d6f-1408-4309-90c9-2971b45d85dc-dbus-socket\") pod \"nmstate-handler-6x6mb\" (UID: \"61d14d6f-1408-4309-90c9-2971b45d85dc\") " pod="openshift-nmstate/nmstate-handler-6x6mb" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.904626 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjnkg\" (UniqueName: \"kubernetes.io/projected/61d14d6f-1408-4309-90c9-2971b45d85dc-kube-api-access-fjnkg\") pod \"nmstate-handler-6x6mb\" (UID: \"61d14d6f-1408-4309-90c9-2971b45d85dc\") " pod="openshift-nmstate/nmstate-handler-6x6mb" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.904668 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/61d14d6f-1408-4309-90c9-2971b45d85dc-nmstate-lock\") pod \"nmstate-handler-6x6mb\" (UID: \"61d14d6f-1408-4309-90c9-2971b45d85dc\") " pod="openshift-nmstate/nmstate-handler-6x6mb" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.904937 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/61d14d6f-1408-4309-90c9-2971b45d85dc-dbus-socket\") pod \"nmstate-handler-6x6mb\" (UID: \"61d14d6f-1408-4309-90c9-2971b45d85dc\") " pod="openshift-nmstate/nmstate-handler-6x6mb" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.904979 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0db507f0-dd6f-405c-ae17-4a9e9553cc50-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-jnvkz\" (UID: \"0db507f0-dd6f-405c-ae17-4a9e9553cc50\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jnvkz" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.905007 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0db507f0-dd6f-405c-ae17-4a9e9553cc50-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-jnvkz\" (UID: \"0db507f0-dd6f-405c-ae17-4a9e9553cc50\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jnvkz" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.905054 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/61d14d6f-1408-4309-90c9-2971b45d85dc-ovs-socket\") pod \"nmstate-handler-6x6mb\" (UID: \"61d14d6f-1408-4309-90c9-2971b45d85dc\") " pod="openshift-nmstate/nmstate-handler-6x6mb" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.905116 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/61d14d6f-1408-4309-90c9-2971b45d85dc-ovs-socket\") pod \"nmstate-handler-6x6mb\" (UID: \"61d14d6f-1408-4309-90c9-2971b45d85dc\") " pod="openshift-nmstate/nmstate-handler-6x6mb" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.912240 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-gqrnm" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.922490 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-64n89" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.934511 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjnkg\" (UniqueName: \"kubernetes.io/projected/61d14d6f-1408-4309-90c9-2971b45d85dc-kube-api-access-fjnkg\") pod \"nmstate-handler-6x6mb\" (UID: \"61d14d6f-1408-4309-90c9-2971b45d85dc\") " pod="openshift-nmstate/nmstate-handler-6x6mb" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.963723 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-6x6mb" Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.996273 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5c6fcdb7d9-7s82h"] Mar 18 20:19:55 crc kubenswrapper[4950]: I0318 20:19:55.997015 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.005947 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjg78\" (UniqueName: \"kubernetes.io/projected/0db507f0-dd6f-405c-ae17-4a9e9553cc50-kube-api-access-jjg78\") pod \"nmstate-console-plugin-86f58fcf4-jnvkz\" (UID: \"0db507f0-dd6f-405c-ae17-4a9e9553cc50\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jnvkz" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.006019 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0db507f0-dd6f-405c-ae17-4a9e9553cc50-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-jnvkz\" (UID: \"0db507f0-dd6f-405c-ae17-4a9e9553cc50\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jnvkz" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.006047 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0db507f0-dd6f-405c-ae17-4a9e9553cc50-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-jnvkz\" (UID: \"0db507f0-dd6f-405c-ae17-4a9e9553cc50\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jnvkz" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.007466 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0db507f0-dd6f-405c-ae17-4a9e9553cc50-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-jnvkz\" (UID: \"0db507f0-dd6f-405c-ae17-4a9e9553cc50\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jnvkz" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.010133 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0db507f0-dd6f-405c-ae17-4a9e9553cc50-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-jnvkz\" (UID: \"0db507f0-dd6f-405c-ae17-4a9e9553cc50\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jnvkz" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.016703 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5c6fcdb7d9-7s82h"] Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.026993 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjg78\" (UniqueName: \"kubernetes.io/projected/0db507f0-dd6f-405c-ae17-4a9e9553cc50-kube-api-access-jjg78\") pod \"nmstate-console-plugin-86f58fcf4-jnvkz\" (UID: \"0db507f0-dd6f-405c-ae17-4a9e9553cc50\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jnvkz" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.077680 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jnvkz" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.106973 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cfb3736b-4668-4bf1-b895-120f3159bcf8-oauth-serving-cert\") pod \"console-5c6fcdb7d9-7s82h\" (UID: \"cfb3736b-4668-4bf1-b895-120f3159bcf8\") " pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.107507 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cfb3736b-4668-4bf1-b895-120f3159bcf8-console-config\") pod \"console-5c6fcdb7d9-7s82h\" (UID: \"cfb3736b-4668-4bf1-b895-120f3159bcf8\") " pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.107539 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cfb3736b-4668-4bf1-b895-120f3159bcf8-console-serving-cert\") pod \"console-5c6fcdb7d9-7s82h\" (UID: \"cfb3736b-4668-4bf1-b895-120f3159bcf8\") " pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.107568 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cfb3736b-4668-4bf1-b895-120f3159bcf8-service-ca\") pod \"console-5c6fcdb7d9-7s82h\" (UID: \"cfb3736b-4668-4bf1-b895-120f3159bcf8\") " pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.107588 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cfb3736b-4668-4bf1-b895-120f3159bcf8-trusted-ca-bundle\") pod \"console-5c6fcdb7d9-7s82h\" (UID: \"cfb3736b-4668-4bf1-b895-120f3159bcf8\") " pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.107609 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55t6r\" (UniqueName: \"kubernetes.io/projected/cfb3736b-4668-4bf1-b895-120f3159bcf8-kube-api-access-55t6r\") pod \"console-5c6fcdb7d9-7s82h\" (UID: \"cfb3736b-4668-4bf1-b895-120f3159bcf8\") " pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.107628 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cfb3736b-4668-4bf1-b895-120f3159bcf8-console-oauth-config\") pod \"console-5c6fcdb7d9-7s82h\" (UID: \"cfb3736b-4668-4bf1-b895-120f3159bcf8\") " pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.175077 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-64n89"] Mar 18 20:19:56 crc kubenswrapper[4950]: W0318 20:19:56.185280 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod452b41f3_a30b_4aec_a879_aff3872199d3.slice/crio-6801b1ea6e94598fb8636e0dba03412c931da9c35b12aed9594cbbc5584485dc WatchSource:0}: Error finding container 6801b1ea6e94598fb8636e0dba03412c931da9c35b12aed9594cbbc5584485dc: Status 404 returned error can't find the container with id 6801b1ea6e94598fb8636e0dba03412c931da9c35b12aed9594cbbc5584485dc Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.209855 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cfb3736b-4668-4bf1-b895-120f3159bcf8-service-ca\") pod \"console-5c6fcdb7d9-7s82h\" (UID: \"cfb3736b-4668-4bf1-b895-120f3159bcf8\") " pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.209903 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cfb3736b-4668-4bf1-b895-120f3159bcf8-trusted-ca-bundle\") pod \"console-5c6fcdb7d9-7s82h\" (UID: \"cfb3736b-4668-4bf1-b895-120f3159bcf8\") " pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.209928 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55t6r\" (UniqueName: \"kubernetes.io/projected/cfb3736b-4668-4bf1-b895-120f3159bcf8-kube-api-access-55t6r\") pod \"console-5c6fcdb7d9-7s82h\" (UID: \"cfb3736b-4668-4bf1-b895-120f3159bcf8\") " pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.209963 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cfb3736b-4668-4bf1-b895-120f3159bcf8-console-oauth-config\") pod \"console-5c6fcdb7d9-7s82h\" (UID: \"cfb3736b-4668-4bf1-b895-120f3159bcf8\") " pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.209990 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cfb3736b-4668-4bf1-b895-120f3159bcf8-oauth-serving-cert\") pod \"console-5c6fcdb7d9-7s82h\" (UID: \"cfb3736b-4668-4bf1-b895-120f3159bcf8\") " pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.210112 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cfb3736b-4668-4bf1-b895-120f3159bcf8-console-config\") pod \"console-5c6fcdb7d9-7s82h\" (UID: \"cfb3736b-4668-4bf1-b895-120f3159bcf8\") " pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.210132 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cfb3736b-4668-4bf1-b895-120f3159bcf8-console-serving-cert\") pod \"console-5c6fcdb7d9-7s82h\" (UID: \"cfb3736b-4668-4bf1-b895-120f3159bcf8\") " pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.213920 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cfb3736b-4668-4bf1-b895-120f3159bcf8-console-serving-cert\") pod \"console-5c6fcdb7d9-7s82h\" (UID: \"cfb3736b-4668-4bf1-b895-120f3159bcf8\") " pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.214130 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cfb3736b-4668-4bf1-b895-120f3159bcf8-service-ca\") pod \"console-5c6fcdb7d9-7s82h\" (UID: \"cfb3736b-4668-4bf1-b895-120f3159bcf8\") " pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.214644 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cfb3736b-4668-4bf1-b895-120f3159bcf8-oauth-serving-cert\") pod \"console-5c6fcdb7d9-7s82h\" (UID: \"cfb3736b-4668-4bf1-b895-120f3159bcf8\") " pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.216519 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cfb3736b-4668-4bf1-b895-120f3159bcf8-console-config\") pod \"console-5c6fcdb7d9-7s82h\" (UID: \"cfb3736b-4668-4bf1-b895-120f3159bcf8\") " pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.217847 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cfb3736b-4668-4bf1-b895-120f3159bcf8-trusted-ca-bundle\") pod \"console-5c6fcdb7d9-7s82h\" (UID: \"cfb3736b-4668-4bf1-b895-120f3159bcf8\") " pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.217885 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cfb3736b-4668-4bf1-b895-120f3159bcf8-console-oauth-config\") pod \"console-5c6fcdb7d9-7s82h\" (UID: \"cfb3736b-4668-4bf1-b895-120f3159bcf8\") " pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.217925 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-gqrnm"] Mar 18 20:19:56 crc kubenswrapper[4950]: W0318 20:19:56.224400 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1b75269_989a_489f_843b_94004e8cdf62.slice/crio-7bd65e3d60107a2d446627d246d5ff4c9c08b38cd4df6d941ad2ea5d2909c643 WatchSource:0}: Error finding container 7bd65e3d60107a2d446627d246d5ff4c9c08b38cd4df6d941ad2ea5d2909c643: Status 404 returned error can't find the container with id 7bd65e3d60107a2d446627d246d5ff4c9c08b38cd4df6d941ad2ea5d2909c643 Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.232224 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55t6r\" (UniqueName: \"kubernetes.io/projected/cfb3736b-4668-4bf1-b895-120f3159bcf8-kube-api-access-55t6r\") pod \"console-5c6fcdb7d9-7s82h\" (UID: \"cfb3736b-4668-4bf1-b895-120f3159bcf8\") " pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.311738 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.499917 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-jnvkz"] Mar 18 20:19:56 crc kubenswrapper[4950]: W0318 20:19:56.507514 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0db507f0_dd6f_405c_ae17_4a9e9553cc50.slice/crio-a93429c6499b7a3b71b975242f68cb481e5920a1f1b1297873b0b6c768f8089c WatchSource:0}: Error finding container a93429c6499b7a3b71b975242f68cb481e5920a1f1b1297873b0b6c768f8089c: Status 404 returned error can't find the container with id a93429c6499b7a3b71b975242f68cb481e5920a1f1b1297873b0b6c768f8089c Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.514755 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-gqrnm" event={"ID":"e1b75269-989a-489f-843b-94004e8cdf62","Type":"ContainerStarted","Data":"7bd65e3d60107a2d446627d246d5ff4c9c08b38cd4df6d941ad2ea5d2909c643"} Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.515985 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-6x6mb" event={"ID":"61d14d6f-1408-4309-90c9-2971b45d85dc","Type":"ContainerStarted","Data":"3e655f1e49bb5bc2392594763c4ce81c297340d5b83bf7425c1a27d81bedde08"} Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.516697 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jnvkz" event={"ID":"0db507f0-dd6f-405c-ae17-4a9e9553cc50","Type":"ContainerStarted","Data":"a93429c6499b7a3b71b975242f68cb481e5920a1f1b1297873b0b6c768f8089c"} Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.520028 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-64n89" event={"ID":"452b41f3-a30b-4aec-a879-aff3872199d3","Type":"ContainerStarted","Data":"6801b1ea6e94598fb8636e0dba03412c931da9c35b12aed9594cbbc5584485dc"} Mar 18 20:19:56 crc kubenswrapper[4950]: I0318 20:19:56.688731 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5c6fcdb7d9-7s82h"] Mar 18 20:19:56 crc kubenswrapper[4950]: W0318 20:19:56.694317 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfb3736b_4668_4bf1_b895_120f3159bcf8.slice/crio-14f4a6a6945270e996a60266389b3da49bf14f2649c19ec41a2db842db07deb5 WatchSource:0}: Error finding container 14f4a6a6945270e996a60266389b3da49bf14f2649c19ec41a2db842db07deb5: Status 404 returned error can't find the container with id 14f4a6a6945270e996a60266389b3da49bf14f2649c19ec41a2db842db07deb5 Mar 18 20:19:57 crc kubenswrapper[4950]: I0318 20:19:57.529175 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5c6fcdb7d9-7s82h" event={"ID":"cfb3736b-4668-4bf1-b895-120f3159bcf8","Type":"ContainerStarted","Data":"6431f0ae7d30b7a4d1e74d6eac92569fc1f8de3889702674e09a055cfcf877b7"} Mar 18 20:19:57 crc kubenswrapper[4950]: I0318 20:19:57.529929 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5c6fcdb7d9-7s82h" event={"ID":"cfb3736b-4668-4bf1-b895-120f3159bcf8","Type":"ContainerStarted","Data":"14f4a6a6945270e996a60266389b3da49bf14f2649c19ec41a2db842db07deb5"} Mar 18 20:19:57 crc kubenswrapper[4950]: I0318 20:19:57.623161 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5c6fcdb7d9-7s82h" podStartSLOduration=2.623138474 podStartE2EDuration="2.623138474s" podCreationTimestamp="2026-03-18 20:19:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:19:57.615976946 +0000 UTC m=+810.856818844" watchObservedRunningTime="2026-03-18 20:19:57.623138474 +0000 UTC m=+810.863980352" Mar 18 20:20:00 crc kubenswrapper[4950]: I0318 20:20:00.138784 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564420-54tbm"] Mar 18 20:20:00 crc kubenswrapper[4950]: I0318 20:20:00.140379 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564420-54tbm" Mar 18 20:20:00 crc kubenswrapper[4950]: I0318 20:20:00.142568 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:20:00 crc kubenswrapper[4950]: I0318 20:20:00.143431 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:20:00 crc kubenswrapper[4950]: I0318 20:20:00.144656 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:20:00 crc kubenswrapper[4950]: I0318 20:20:00.148551 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564420-54tbm"] Mar 18 20:20:00 crc kubenswrapper[4950]: I0318 20:20:00.262906 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h29dv\" (UniqueName: \"kubernetes.io/projected/3bad5973-c35e-4a25-9ba4-50958354a198-kube-api-access-h29dv\") pod \"auto-csr-approver-29564420-54tbm\" (UID: \"3bad5973-c35e-4a25-9ba4-50958354a198\") " pod="openshift-infra/auto-csr-approver-29564420-54tbm" Mar 18 20:20:00 crc kubenswrapper[4950]: I0318 20:20:00.363715 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h29dv\" (UniqueName: \"kubernetes.io/projected/3bad5973-c35e-4a25-9ba4-50958354a198-kube-api-access-h29dv\") pod \"auto-csr-approver-29564420-54tbm\" (UID: \"3bad5973-c35e-4a25-9ba4-50958354a198\") " pod="openshift-infra/auto-csr-approver-29564420-54tbm" Mar 18 20:20:00 crc kubenswrapper[4950]: I0318 20:20:00.383580 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h29dv\" (UniqueName: \"kubernetes.io/projected/3bad5973-c35e-4a25-9ba4-50958354a198-kube-api-access-h29dv\") pod \"auto-csr-approver-29564420-54tbm\" (UID: \"3bad5973-c35e-4a25-9ba4-50958354a198\") " pod="openshift-infra/auto-csr-approver-29564420-54tbm" Mar 18 20:20:00 crc kubenswrapper[4950]: I0318 20:20:00.454108 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564420-54tbm" Mar 18 20:20:00 crc kubenswrapper[4950]: I0318 20:20:00.551807 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jnvkz" event={"ID":"0db507f0-dd6f-405c-ae17-4a9e9553cc50","Type":"ContainerStarted","Data":"51f1b0eeecf99d00fc9613283ae10fe39a998bd4f438cc08c4897e4afe802ee8"} Mar 18 20:20:00 crc kubenswrapper[4950]: I0318 20:20:00.557224 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-64n89" event={"ID":"452b41f3-a30b-4aec-a879-aff3872199d3","Type":"ContainerStarted","Data":"9b6ad0a23f10661d1d87682f72ad2fd46c41f75406df8b85a1763d165eccf8f1"} Mar 18 20:20:00 crc kubenswrapper[4950]: I0318 20:20:00.558113 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f558f5558-64n89" Mar 18 20:20:00 crc kubenswrapper[4950]: I0318 20:20:00.575706 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-jnvkz" podStartSLOduration=2.699912692 podStartE2EDuration="5.575692124s" podCreationTimestamp="2026-03-18 20:19:55 +0000 UTC" firstStartedPulling="2026-03-18 20:19:56.50968706 +0000 UTC m=+809.750528928" lastFinishedPulling="2026-03-18 20:19:59.385466502 +0000 UTC m=+812.626308360" observedRunningTime="2026-03-18 20:20:00.572017458 +0000 UTC m=+813.812859326" watchObservedRunningTime="2026-03-18 20:20:00.575692124 +0000 UTC m=+813.816533992" Mar 18 20:20:00 crc kubenswrapper[4950]: I0318 20:20:00.584591 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-gqrnm" event={"ID":"e1b75269-989a-489f-843b-94004e8cdf62","Type":"ContainerStarted","Data":"c3e9ce35fd540d28f22aec87ec5f16cb396774a9cbb3de74bc9a8278990335e9"} Mar 18 20:20:00 crc kubenswrapper[4950]: I0318 20:20:00.593440 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f558f5558-64n89" podStartSLOduration=2.388333768 podStartE2EDuration="5.59342565s" podCreationTimestamp="2026-03-18 20:19:55 +0000 UTC" firstStartedPulling="2026-03-18 20:19:56.188097783 +0000 UTC m=+809.428939651" lastFinishedPulling="2026-03-18 20:19:59.393189665 +0000 UTC m=+812.634031533" observedRunningTime="2026-03-18 20:20:00.589958189 +0000 UTC m=+813.830800057" watchObservedRunningTime="2026-03-18 20:20:00.59342565 +0000 UTC m=+813.834267518" Mar 18 20:20:00 crc kubenswrapper[4950]: I0318 20:20:00.593503 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-6x6mb" event={"ID":"61d14d6f-1408-4309-90c9-2971b45d85dc","Type":"ContainerStarted","Data":"1af94c34646699094cbd5252382825e13b61f4431ded57570d31705872f9817e"} Mar 18 20:20:00 crc kubenswrapper[4950]: I0318 20:20:00.594393 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-6x6mb" Mar 18 20:20:00 crc kubenswrapper[4950]: I0318 20:20:00.619110 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-6x6mb" podStartSLOduration=2.256937326 podStartE2EDuration="5.619095814s" podCreationTimestamp="2026-03-18 20:19:55 +0000 UTC" firstStartedPulling="2026-03-18 20:19:56.02732204 +0000 UTC m=+809.268163908" lastFinishedPulling="2026-03-18 20:19:59.389480528 +0000 UTC m=+812.630322396" observedRunningTime="2026-03-18 20:20:00.617680327 +0000 UTC m=+813.858522195" watchObservedRunningTime="2026-03-18 20:20:00.619095814 +0000 UTC m=+813.859937682" Mar 18 20:20:00 crc kubenswrapper[4950]: I0318 20:20:00.891193 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564420-54tbm"] Mar 18 20:20:00 crc kubenswrapper[4950]: W0318 20:20:00.894761 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3bad5973_c35e_4a25_9ba4_50958354a198.slice/crio-88fd0ab9b71cace4951238aae6071593cd868f02d04890117cc2aa0a3697ad87 WatchSource:0}: Error finding container 88fd0ab9b71cace4951238aae6071593cd868f02d04890117cc2aa0a3697ad87: Status 404 returned error can't find the container with id 88fd0ab9b71cace4951238aae6071593cd868f02d04890117cc2aa0a3697ad87 Mar 18 20:20:01 crc kubenswrapper[4950]: I0318 20:20:01.602477 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564420-54tbm" event={"ID":"3bad5973-c35e-4a25-9ba4-50958354a198","Type":"ContainerStarted","Data":"88fd0ab9b71cace4951238aae6071593cd868f02d04890117cc2aa0a3697ad87"} Mar 18 20:20:02 crc kubenswrapper[4950]: I0318 20:20:02.610385 4950 generic.go:334] "Generic (PLEG): container finished" podID="3bad5973-c35e-4a25-9ba4-50958354a198" containerID="54d11f46c8dc572a33f1b53b29578543919f5927e4093db85e04577390e80f45" exitCode=0 Mar 18 20:20:02 crc kubenswrapper[4950]: I0318 20:20:02.610458 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564420-54tbm" event={"ID":"3bad5973-c35e-4a25-9ba4-50958354a198","Type":"ContainerDied","Data":"54d11f46c8dc572a33f1b53b29578543919f5927e4093db85e04577390e80f45"} Mar 18 20:20:02 crc kubenswrapper[4950]: I0318 20:20:02.612925 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-gqrnm" event={"ID":"e1b75269-989a-489f-843b-94004e8cdf62","Type":"ContainerStarted","Data":"5941e8986702b7130bab6299474f9333a3f54839455abf96cf721dcaebcaa981"} Mar 18 20:20:02 crc kubenswrapper[4950]: I0318 20:20:02.643778 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-gqrnm" podStartSLOduration=1.805394217 podStartE2EDuration="7.643752212s" podCreationTimestamp="2026-03-18 20:19:55 +0000 UTC" firstStartedPulling="2026-03-18 20:19:56.225786353 +0000 UTC m=+809.466628221" lastFinishedPulling="2026-03-18 20:20:02.064144348 +0000 UTC m=+815.304986216" observedRunningTime="2026-03-18 20:20:02.640233569 +0000 UTC m=+815.881075457" watchObservedRunningTime="2026-03-18 20:20:02.643752212 +0000 UTC m=+815.884594100" Mar 18 20:20:03 crc kubenswrapper[4950]: I0318 20:20:03.925364 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564420-54tbm" Mar 18 20:20:04 crc kubenswrapper[4950]: I0318 20:20:04.018143 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h29dv\" (UniqueName: \"kubernetes.io/projected/3bad5973-c35e-4a25-9ba4-50958354a198-kube-api-access-h29dv\") pod \"3bad5973-c35e-4a25-9ba4-50958354a198\" (UID: \"3bad5973-c35e-4a25-9ba4-50958354a198\") " Mar 18 20:20:04 crc kubenswrapper[4950]: I0318 20:20:04.026839 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bad5973-c35e-4a25-9ba4-50958354a198-kube-api-access-h29dv" (OuterVolumeSpecName: "kube-api-access-h29dv") pod "3bad5973-c35e-4a25-9ba4-50958354a198" (UID: "3bad5973-c35e-4a25-9ba4-50958354a198"). InnerVolumeSpecName "kube-api-access-h29dv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:20:04 crc kubenswrapper[4950]: I0318 20:20:04.120486 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h29dv\" (UniqueName: \"kubernetes.io/projected/3bad5973-c35e-4a25-9ba4-50958354a198-kube-api-access-h29dv\") on node \"crc\" DevicePath \"\"" Mar 18 20:20:05 crc kubenswrapper[4950]: I0318 20:20:05.026326 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564420-54tbm" event={"ID":"3bad5973-c35e-4a25-9ba4-50958354a198","Type":"ContainerDied","Data":"88fd0ab9b71cace4951238aae6071593cd868f02d04890117cc2aa0a3697ad87"} Mar 18 20:20:05 crc kubenswrapper[4950]: I0318 20:20:05.026360 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88fd0ab9b71cace4951238aae6071593cd868f02d04890117cc2aa0a3697ad87" Mar 18 20:20:05 crc kubenswrapper[4950]: I0318 20:20:05.026454 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564420-54tbm" Mar 18 20:20:06 crc kubenswrapper[4950]: I0318 20:20:06.007696 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-6x6mb" Mar 18 20:20:06 crc kubenswrapper[4950]: I0318 20:20:06.022089 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564414-nf9v5"] Mar 18 20:20:06 crc kubenswrapper[4950]: I0318 20:20:06.031746 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564414-nf9v5"] Mar 18 20:20:06 crc kubenswrapper[4950]: I0318 20:20:06.311859 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:20:06 crc kubenswrapper[4950]: I0318 20:20:06.312094 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:20:06 crc kubenswrapper[4950]: I0318 20:20:06.316211 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:20:07 crc kubenswrapper[4950]: I0318 20:20:07.057824 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5c6fcdb7d9-7s82h" Mar 18 20:20:07 crc kubenswrapper[4950]: I0318 20:20:07.142827 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-46hjn"] Mar 18 20:20:07 crc kubenswrapper[4950]: I0318 20:20:07.493472 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88c51f5c-d48e-4779-879a-f7169556d9bd" path="/var/lib/kubelet/pods/88c51f5c-d48e-4779-879a-f7169556d9bd/volumes" Mar 18 20:20:15 crc kubenswrapper[4950]: I0318 20:20:15.934287 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f558f5558-64n89" Mar 18 20:20:23 crc kubenswrapper[4950]: I0318 20:20:23.559116 4950 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 18 20:20:30 crc kubenswrapper[4950]: I0318 20:20:30.311996 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk"] Mar 18 20:20:30 crc kubenswrapper[4950]: E0318 20:20:30.313256 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bad5973-c35e-4a25-9ba4-50958354a198" containerName="oc" Mar 18 20:20:30 crc kubenswrapper[4950]: I0318 20:20:30.313288 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bad5973-c35e-4a25-9ba4-50958354a198" containerName="oc" Mar 18 20:20:30 crc kubenswrapper[4950]: I0318 20:20:30.313547 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bad5973-c35e-4a25-9ba4-50958354a198" containerName="oc" Mar 18 20:20:30 crc kubenswrapper[4950]: I0318 20:20:30.315214 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk" Mar 18 20:20:30 crc kubenswrapper[4950]: I0318 20:20:30.320233 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 18 20:20:30 crc kubenswrapper[4950]: I0318 20:20:30.326396 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk"] Mar 18 20:20:30 crc kubenswrapper[4950]: I0318 20:20:30.425736 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c955c23a-4fa9-4745-bb00-05a54ae897f4-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk\" (UID: \"c955c23a-4fa9-4745-bb00-05a54ae897f4\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk" Mar 18 20:20:30 crc kubenswrapper[4950]: I0318 20:20:30.425835 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c955c23a-4fa9-4745-bb00-05a54ae897f4-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk\" (UID: \"c955c23a-4fa9-4745-bb00-05a54ae897f4\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk" Mar 18 20:20:30 crc kubenswrapper[4950]: I0318 20:20:30.425916 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2fdj\" (UniqueName: \"kubernetes.io/projected/c955c23a-4fa9-4745-bb00-05a54ae897f4-kube-api-access-d2fdj\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk\" (UID: \"c955c23a-4fa9-4745-bb00-05a54ae897f4\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk" Mar 18 20:20:30 crc kubenswrapper[4950]: I0318 20:20:30.527082 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c955c23a-4fa9-4745-bb00-05a54ae897f4-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk\" (UID: \"c955c23a-4fa9-4745-bb00-05a54ae897f4\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk" Mar 18 20:20:30 crc kubenswrapper[4950]: I0318 20:20:30.528591 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c955c23a-4fa9-4745-bb00-05a54ae897f4-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk\" (UID: \"c955c23a-4fa9-4745-bb00-05a54ae897f4\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk" Mar 18 20:20:30 crc kubenswrapper[4950]: I0318 20:20:30.528691 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c955c23a-4fa9-4745-bb00-05a54ae897f4-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk\" (UID: \"c955c23a-4fa9-4745-bb00-05a54ae897f4\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk" Mar 18 20:20:30 crc kubenswrapper[4950]: I0318 20:20:30.528898 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2fdj\" (UniqueName: \"kubernetes.io/projected/c955c23a-4fa9-4745-bb00-05a54ae897f4-kube-api-access-d2fdj\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk\" (UID: \"c955c23a-4fa9-4745-bb00-05a54ae897f4\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk" Mar 18 20:20:30 crc kubenswrapper[4950]: I0318 20:20:30.529190 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c955c23a-4fa9-4745-bb00-05a54ae897f4-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk\" (UID: \"c955c23a-4fa9-4745-bb00-05a54ae897f4\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk" Mar 18 20:20:30 crc kubenswrapper[4950]: I0318 20:20:30.557463 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2fdj\" (UniqueName: \"kubernetes.io/projected/c955c23a-4fa9-4745-bb00-05a54ae897f4-kube-api-access-d2fdj\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk\" (UID: \"c955c23a-4fa9-4745-bb00-05a54ae897f4\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk" Mar 18 20:20:30 crc kubenswrapper[4950]: I0318 20:20:30.643927 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk" Mar 18 20:20:30 crc kubenswrapper[4950]: I0318 20:20:30.866262 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk"] Mar 18 20:20:31 crc kubenswrapper[4950]: I0318 20:20:31.245687 4950 generic.go:334] "Generic (PLEG): container finished" podID="c955c23a-4fa9-4745-bb00-05a54ae897f4" containerID="43ef934d7c8525faaca28f2c98fa1e5eb3af8a032b844ecf04dd784bdc8f3bb7" exitCode=0 Mar 18 20:20:31 crc kubenswrapper[4950]: I0318 20:20:31.245846 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk" event={"ID":"c955c23a-4fa9-4745-bb00-05a54ae897f4","Type":"ContainerDied","Data":"43ef934d7c8525faaca28f2c98fa1e5eb3af8a032b844ecf04dd784bdc8f3bb7"} Mar 18 20:20:31 crc kubenswrapper[4950]: I0318 20:20:31.246092 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk" event={"ID":"c955c23a-4fa9-4745-bb00-05a54ae897f4","Type":"ContainerStarted","Data":"f7ec611c609c84b2934e72ec77d67e369c736d990728138317f5d3058ba2d919"} Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.191323 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-46hjn" podUID="1cec2292-f508-42b2-916f-9a5808045626" containerName="console" containerID="cri-o://60eb1f53d56060d604da66326a97602acb70000d70709e4e1a6cc167a6f57a65" gracePeriod=15 Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.527756 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-46hjn_1cec2292-f508-42b2-916f-9a5808045626/console/0.log" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.527835 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.559330 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1cec2292-f508-42b2-916f-9a5808045626-console-serving-cert\") pod \"1cec2292-f508-42b2-916f-9a5808045626\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.559440 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-oauth-serving-cert\") pod \"1cec2292-f508-42b2-916f-9a5808045626\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.559518 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1cec2292-f508-42b2-916f-9a5808045626-console-oauth-config\") pod \"1cec2292-f508-42b2-916f-9a5808045626\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.559571 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k2z5\" (UniqueName: \"kubernetes.io/projected/1cec2292-f508-42b2-916f-9a5808045626-kube-api-access-2k2z5\") pod \"1cec2292-f508-42b2-916f-9a5808045626\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.559612 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-console-config\") pod \"1cec2292-f508-42b2-916f-9a5808045626\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.559649 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-trusted-ca-bundle\") pod \"1cec2292-f508-42b2-916f-9a5808045626\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.559683 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-service-ca\") pod \"1cec2292-f508-42b2-916f-9a5808045626\" (UID: \"1cec2292-f508-42b2-916f-9a5808045626\") " Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.560748 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-console-config" (OuterVolumeSpecName: "console-config") pod "1cec2292-f508-42b2-916f-9a5808045626" (UID: "1cec2292-f508-42b2-916f-9a5808045626"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.560798 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "1cec2292-f508-42b2-916f-9a5808045626" (UID: "1cec2292-f508-42b2-916f-9a5808045626"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.561204 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-service-ca" (OuterVolumeSpecName: "service-ca") pod "1cec2292-f508-42b2-916f-9a5808045626" (UID: "1cec2292-f508-42b2-916f-9a5808045626"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.561318 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1cec2292-f508-42b2-916f-9a5808045626" (UID: "1cec2292-f508-42b2-916f-9a5808045626"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.565752 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cec2292-f508-42b2-916f-9a5808045626-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "1cec2292-f508-42b2-916f-9a5808045626" (UID: "1cec2292-f508-42b2-916f-9a5808045626"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.565836 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cec2292-f508-42b2-916f-9a5808045626-kube-api-access-2k2z5" (OuterVolumeSpecName: "kube-api-access-2k2z5") pod "1cec2292-f508-42b2-916f-9a5808045626" (UID: "1cec2292-f508-42b2-916f-9a5808045626"). InnerVolumeSpecName "kube-api-access-2k2z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.567887 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cec2292-f508-42b2-916f-9a5808045626-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "1cec2292-f508-42b2-916f-9a5808045626" (UID: "1cec2292-f508-42b2-916f-9a5808045626"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.643499 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7tfkw"] Mar 18 20:20:32 crc kubenswrapper[4950]: E0318 20:20:32.644282 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cec2292-f508-42b2-916f-9a5808045626" containerName="console" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.644306 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cec2292-f508-42b2-916f-9a5808045626" containerName="console" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.644526 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cec2292-f508-42b2-916f-9a5808045626" containerName="console" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.645251 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tfkw" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.653672 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7tfkw"] Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.661143 4950 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1cec2292-f508-42b2-916f-9a5808045626-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.661193 4950 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.661206 4950 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1cec2292-f508-42b2-916f-9a5808045626-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.661220 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k2z5\" (UniqueName: \"kubernetes.io/projected/1cec2292-f508-42b2-916f-9a5808045626-kube-api-access-2k2z5\") on node \"crc\" DevicePath \"\"" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.661233 4950 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-console-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.661245 4950 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.661257 4950 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1cec2292-f508-42b2-916f-9a5808045626-service-ca\") on node \"crc\" DevicePath \"\"" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.762219 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxg5g\" (UniqueName: \"kubernetes.io/projected/ca3038fa-120b-412e-80b2-24aca2dd19e8-kube-api-access-pxg5g\") pod \"redhat-operators-7tfkw\" (UID: \"ca3038fa-120b-412e-80b2-24aca2dd19e8\") " pod="openshift-marketplace/redhat-operators-7tfkw" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.762360 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca3038fa-120b-412e-80b2-24aca2dd19e8-catalog-content\") pod \"redhat-operators-7tfkw\" (UID: \"ca3038fa-120b-412e-80b2-24aca2dd19e8\") " pod="openshift-marketplace/redhat-operators-7tfkw" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.762400 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca3038fa-120b-412e-80b2-24aca2dd19e8-utilities\") pod \"redhat-operators-7tfkw\" (UID: \"ca3038fa-120b-412e-80b2-24aca2dd19e8\") " pod="openshift-marketplace/redhat-operators-7tfkw" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.863708 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxg5g\" (UniqueName: \"kubernetes.io/projected/ca3038fa-120b-412e-80b2-24aca2dd19e8-kube-api-access-pxg5g\") pod \"redhat-operators-7tfkw\" (UID: \"ca3038fa-120b-412e-80b2-24aca2dd19e8\") " pod="openshift-marketplace/redhat-operators-7tfkw" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.863785 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca3038fa-120b-412e-80b2-24aca2dd19e8-catalog-content\") pod \"redhat-operators-7tfkw\" (UID: \"ca3038fa-120b-412e-80b2-24aca2dd19e8\") " pod="openshift-marketplace/redhat-operators-7tfkw" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.863812 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca3038fa-120b-412e-80b2-24aca2dd19e8-utilities\") pod \"redhat-operators-7tfkw\" (UID: \"ca3038fa-120b-412e-80b2-24aca2dd19e8\") " pod="openshift-marketplace/redhat-operators-7tfkw" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.864169 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca3038fa-120b-412e-80b2-24aca2dd19e8-utilities\") pod \"redhat-operators-7tfkw\" (UID: \"ca3038fa-120b-412e-80b2-24aca2dd19e8\") " pod="openshift-marketplace/redhat-operators-7tfkw" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.864605 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca3038fa-120b-412e-80b2-24aca2dd19e8-catalog-content\") pod \"redhat-operators-7tfkw\" (UID: \"ca3038fa-120b-412e-80b2-24aca2dd19e8\") " pod="openshift-marketplace/redhat-operators-7tfkw" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.883002 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxg5g\" (UniqueName: \"kubernetes.io/projected/ca3038fa-120b-412e-80b2-24aca2dd19e8-kube-api-access-pxg5g\") pod \"redhat-operators-7tfkw\" (UID: \"ca3038fa-120b-412e-80b2-24aca2dd19e8\") " pod="openshift-marketplace/redhat-operators-7tfkw" Mar 18 20:20:32 crc kubenswrapper[4950]: I0318 20:20:32.972048 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tfkw" Mar 18 20:20:33 crc kubenswrapper[4950]: I0318 20:20:33.261552 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-46hjn_1cec2292-f508-42b2-916f-9a5808045626/console/0.log" Mar 18 20:20:33 crc kubenswrapper[4950]: I0318 20:20:33.261890 4950 generic.go:334] "Generic (PLEG): container finished" podID="1cec2292-f508-42b2-916f-9a5808045626" containerID="60eb1f53d56060d604da66326a97602acb70000d70709e4e1a6cc167a6f57a65" exitCode=2 Mar 18 20:20:33 crc kubenswrapper[4950]: I0318 20:20:33.261938 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-46hjn" event={"ID":"1cec2292-f508-42b2-916f-9a5808045626","Type":"ContainerDied","Data":"60eb1f53d56060d604da66326a97602acb70000d70709e4e1a6cc167a6f57a65"} Mar 18 20:20:33 crc kubenswrapper[4950]: I0318 20:20:33.261960 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-46hjn" Mar 18 20:20:33 crc kubenswrapper[4950]: I0318 20:20:33.261986 4950 scope.go:117] "RemoveContainer" containerID="60eb1f53d56060d604da66326a97602acb70000d70709e4e1a6cc167a6f57a65" Mar 18 20:20:33 crc kubenswrapper[4950]: I0318 20:20:33.261974 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-46hjn" event={"ID":"1cec2292-f508-42b2-916f-9a5808045626","Type":"ContainerDied","Data":"b44d263b6e57afdd082bb8aef49f2502668371c453f03cf85053218b0197ca42"} Mar 18 20:20:33 crc kubenswrapper[4950]: I0318 20:20:33.265781 4950 generic.go:334] "Generic (PLEG): container finished" podID="c955c23a-4fa9-4745-bb00-05a54ae897f4" containerID="a0fec433e817bc7b82f8f6f8588a6f5534e4d98078f5167740084ac3f0f596a6" exitCode=0 Mar 18 20:20:33 crc kubenswrapper[4950]: I0318 20:20:33.265814 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk" event={"ID":"c955c23a-4fa9-4745-bb00-05a54ae897f4","Type":"ContainerDied","Data":"a0fec433e817bc7b82f8f6f8588a6f5534e4d98078f5167740084ac3f0f596a6"} Mar 18 20:20:33 crc kubenswrapper[4950]: I0318 20:20:33.303091 4950 scope.go:117] "RemoveContainer" containerID="60eb1f53d56060d604da66326a97602acb70000d70709e4e1a6cc167a6f57a65" Mar 18 20:20:33 crc kubenswrapper[4950]: E0318 20:20:33.304078 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60eb1f53d56060d604da66326a97602acb70000d70709e4e1a6cc167a6f57a65\": container with ID starting with 60eb1f53d56060d604da66326a97602acb70000d70709e4e1a6cc167a6f57a65 not found: ID does not exist" containerID="60eb1f53d56060d604da66326a97602acb70000d70709e4e1a6cc167a6f57a65" Mar 18 20:20:33 crc kubenswrapper[4950]: I0318 20:20:33.304108 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60eb1f53d56060d604da66326a97602acb70000d70709e4e1a6cc167a6f57a65"} err="failed to get container status \"60eb1f53d56060d604da66326a97602acb70000d70709e4e1a6cc167a6f57a65\": rpc error: code = NotFound desc = could not find container \"60eb1f53d56060d604da66326a97602acb70000d70709e4e1a6cc167a6f57a65\": container with ID starting with 60eb1f53d56060d604da66326a97602acb70000d70709e4e1a6cc167a6f57a65 not found: ID does not exist" Mar 18 20:20:33 crc kubenswrapper[4950]: I0318 20:20:33.307397 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-46hjn"] Mar 18 20:20:33 crc kubenswrapper[4950]: I0318 20:20:33.310845 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-46hjn"] Mar 18 20:20:33 crc kubenswrapper[4950]: I0318 20:20:33.386332 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7tfkw"] Mar 18 20:20:33 crc kubenswrapper[4950]: I0318 20:20:33.485992 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cec2292-f508-42b2-916f-9a5808045626" path="/var/lib/kubelet/pods/1cec2292-f508-42b2-916f-9a5808045626/volumes" Mar 18 20:20:34 crc kubenswrapper[4950]: I0318 20:20:34.275949 4950 generic.go:334] "Generic (PLEG): container finished" podID="ca3038fa-120b-412e-80b2-24aca2dd19e8" containerID="8af013eb61caabc1a7a29cb9ebcbaa4e824cd87b0b96803895878d5100156b8f" exitCode=0 Mar 18 20:20:34 crc kubenswrapper[4950]: I0318 20:20:34.276058 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tfkw" event={"ID":"ca3038fa-120b-412e-80b2-24aca2dd19e8","Type":"ContainerDied","Data":"8af013eb61caabc1a7a29cb9ebcbaa4e824cd87b0b96803895878d5100156b8f"} Mar 18 20:20:34 crc kubenswrapper[4950]: I0318 20:20:34.276094 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tfkw" event={"ID":"ca3038fa-120b-412e-80b2-24aca2dd19e8","Type":"ContainerStarted","Data":"504f87f073a1d77b20ced1f43a36ef60a28fb0ef005c3206892a6973651e9591"} Mar 18 20:20:34 crc kubenswrapper[4950]: I0318 20:20:34.279687 4950 generic.go:334] "Generic (PLEG): container finished" podID="c955c23a-4fa9-4745-bb00-05a54ae897f4" containerID="381caa9582977e811f30389f7ed46753b0cc549fcd1f519dc15a643267eb0bd2" exitCode=0 Mar 18 20:20:34 crc kubenswrapper[4950]: I0318 20:20:34.279731 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk" event={"ID":"c955c23a-4fa9-4745-bb00-05a54ae897f4","Type":"ContainerDied","Data":"381caa9582977e811f30389f7ed46753b0cc549fcd1f519dc15a643267eb0bd2"} Mar 18 20:20:35 crc kubenswrapper[4950]: I0318 20:20:35.290283 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tfkw" event={"ID":"ca3038fa-120b-412e-80b2-24aca2dd19e8","Type":"ContainerStarted","Data":"5f1752e256cd8af351e0e47ae0db4e862d1d884a40e79ca93919c7c562a6f404"} Mar 18 20:20:35 crc kubenswrapper[4950]: I0318 20:20:35.588796 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk" Mar 18 20:20:35 crc kubenswrapper[4950]: I0318 20:20:35.701558 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2fdj\" (UniqueName: \"kubernetes.io/projected/c955c23a-4fa9-4745-bb00-05a54ae897f4-kube-api-access-d2fdj\") pod \"c955c23a-4fa9-4745-bb00-05a54ae897f4\" (UID: \"c955c23a-4fa9-4745-bb00-05a54ae897f4\") " Mar 18 20:20:35 crc kubenswrapper[4950]: I0318 20:20:35.701623 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c955c23a-4fa9-4745-bb00-05a54ae897f4-bundle\") pod \"c955c23a-4fa9-4745-bb00-05a54ae897f4\" (UID: \"c955c23a-4fa9-4745-bb00-05a54ae897f4\") " Mar 18 20:20:35 crc kubenswrapper[4950]: I0318 20:20:35.701656 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c955c23a-4fa9-4745-bb00-05a54ae897f4-util\") pod \"c955c23a-4fa9-4745-bb00-05a54ae897f4\" (UID: \"c955c23a-4fa9-4745-bb00-05a54ae897f4\") " Mar 18 20:20:35 crc kubenswrapper[4950]: I0318 20:20:35.702732 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c955c23a-4fa9-4745-bb00-05a54ae897f4-bundle" (OuterVolumeSpecName: "bundle") pod "c955c23a-4fa9-4745-bb00-05a54ae897f4" (UID: "c955c23a-4fa9-4745-bb00-05a54ae897f4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:20:35 crc kubenswrapper[4950]: I0318 20:20:35.707155 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c955c23a-4fa9-4745-bb00-05a54ae897f4-kube-api-access-d2fdj" (OuterVolumeSpecName: "kube-api-access-d2fdj") pod "c955c23a-4fa9-4745-bb00-05a54ae897f4" (UID: "c955c23a-4fa9-4745-bb00-05a54ae897f4"). InnerVolumeSpecName "kube-api-access-d2fdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:20:35 crc kubenswrapper[4950]: I0318 20:20:35.731862 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c955c23a-4fa9-4745-bb00-05a54ae897f4-util" (OuterVolumeSpecName: "util") pod "c955c23a-4fa9-4745-bb00-05a54ae897f4" (UID: "c955c23a-4fa9-4745-bb00-05a54ae897f4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:20:35 crc kubenswrapper[4950]: I0318 20:20:35.802885 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2fdj\" (UniqueName: \"kubernetes.io/projected/c955c23a-4fa9-4745-bb00-05a54ae897f4-kube-api-access-d2fdj\") on node \"crc\" DevicePath \"\"" Mar 18 20:20:35 crc kubenswrapper[4950]: I0318 20:20:35.802943 4950 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c955c23a-4fa9-4745-bb00-05a54ae897f4-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:20:35 crc kubenswrapper[4950]: I0318 20:20:35.803010 4950 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c955c23a-4fa9-4745-bb00-05a54ae897f4-util\") on node \"crc\" DevicePath \"\"" Mar 18 20:20:36 crc kubenswrapper[4950]: I0318 20:20:36.299109 4950 generic.go:334] "Generic (PLEG): container finished" podID="ca3038fa-120b-412e-80b2-24aca2dd19e8" containerID="5f1752e256cd8af351e0e47ae0db4e862d1d884a40e79ca93919c7c562a6f404" exitCode=0 Mar 18 20:20:36 crc kubenswrapper[4950]: I0318 20:20:36.299174 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tfkw" event={"ID":"ca3038fa-120b-412e-80b2-24aca2dd19e8","Type":"ContainerDied","Data":"5f1752e256cd8af351e0e47ae0db4e862d1d884a40e79ca93919c7c562a6f404"} Mar 18 20:20:36 crc kubenswrapper[4950]: I0318 20:20:36.302610 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk" event={"ID":"c955c23a-4fa9-4745-bb00-05a54ae897f4","Type":"ContainerDied","Data":"f7ec611c609c84b2934e72ec77d67e369c736d990728138317f5d3058ba2d919"} Mar 18 20:20:36 crc kubenswrapper[4950]: I0318 20:20:36.302641 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7ec611c609c84b2934e72ec77d67e369c736d990728138317f5d3058ba2d919" Mar 18 20:20:36 crc kubenswrapper[4950]: I0318 20:20:36.302705 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk" Mar 18 20:20:37 crc kubenswrapper[4950]: I0318 20:20:37.312010 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tfkw" event={"ID":"ca3038fa-120b-412e-80b2-24aca2dd19e8","Type":"ContainerStarted","Data":"95cdad0f9ff05b1ff2cacc5b7342267ecc72202391456ee0b088de53e2cac6a0"} Mar 18 20:20:37 crc kubenswrapper[4950]: E0318 20:20:37.566438 4950 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/NetworkManager-dispatcher.service\": RecentStats: unable to find data in memory cache]" Mar 18 20:20:42 crc kubenswrapper[4950]: I0318 20:20:42.973388 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7tfkw" Mar 18 20:20:42 crc kubenswrapper[4950]: I0318 20:20:42.974500 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7tfkw" Mar 18 20:20:44 crc kubenswrapper[4950]: I0318 20:20:44.018807 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7tfkw" podUID="ca3038fa-120b-412e-80b2-24aca2dd19e8" containerName="registry-server" probeResult="failure" output=< Mar 18 20:20:44 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 20:20:44 crc kubenswrapper[4950]: > Mar 18 20:20:44 crc kubenswrapper[4950]: I0318 20:20:44.480960 4950 scope.go:117] "RemoveContainer" containerID="606b0b7a879af47936c0166efea78e8d4cb57aa6954e2812f7432305854e6981" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.391014 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7tfkw" podStartSLOduration=12.945822675 podStartE2EDuration="15.390999573s" podCreationTimestamp="2026-03-18 20:20:32 +0000 UTC" firstStartedPulling="2026-03-18 20:20:34.279606881 +0000 UTC m=+847.520448749" lastFinishedPulling="2026-03-18 20:20:36.724783769 +0000 UTC m=+849.965625647" observedRunningTime="2026-03-18 20:20:37.332320822 +0000 UTC m=+850.573162700" watchObservedRunningTime="2026-03-18 20:20:47.390999573 +0000 UTC m=+860.631841441" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.397171 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-648849f9d4-9zr5p"] Mar 18 20:20:47 crc kubenswrapper[4950]: E0318 20:20:47.397590 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c955c23a-4fa9-4745-bb00-05a54ae897f4" containerName="extract" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.397605 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="c955c23a-4fa9-4745-bb00-05a54ae897f4" containerName="extract" Mar 18 20:20:47 crc kubenswrapper[4950]: E0318 20:20:47.397617 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c955c23a-4fa9-4745-bb00-05a54ae897f4" containerName="pull" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.397623 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="c955c23a-4fa9-4745-bb00-05a54ae897f4" containerName="pull" Mar 18 20:20:47 crc kubenswrapper[4950]: E0318 20:20:47.397639 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c955c23a-4fa9-4745-bb00-05a54ae897f4" containerName="util" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.397649 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="c955c23a-4fa9-4745-bb00-05a54ae897f4" containerName="util" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.397863 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="c955c23a-4fa9-4745-bb00-05a54ae897f4" containerName="extract" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.398420 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-648849f9d4-9zr5p" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.402429 4950 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.402575 4950 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-nlk52" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.402606 4950 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.402653 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.402744 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.447735 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm6wp\" (UniqueName: \"kubernetes.io/projected/871672d1-30f1-4221-a434-2a8d60173f46-kube-api-access-fm6wp\") pod \"metallb-operator-controller-manager-648849f9d4-9zr5p\" (UID: \"871672d1-30f1-4221-a434-2a8d60173f46\") " pod="metallb-system/metallb-operator-controller-manager-648849f9d4-9zr5p" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.447783 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/871672d1-30f1-4221-a434-2a8d60173f46-apiservice-cert\") pod \"metallb-operator-controller-manager-648849f9d4-9zr5p\" (UID: \"871672d1-30f1-4221-a434-2a8d60173f46\") " pod="metallb-system/metallb-operator-controller-manager-648849f9d4-9zr5p" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.447826 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/871672d1-30f1-4221-a434-2a8d60173f46-webhook-cert\") pod \"metallb-operator-controller-manager-648849f9d4-9zr5p\" (UID: \"871672d1-30f1-4221-a434-2a8d60173f46\") " pod="metallb-system/metallb-operator-controller-manager-648849f9d4-9zr5p" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.468362 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-648849f9d4-9zr5p"] Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.548197 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm6wp\" (UniqueName: \"kubernetes.io/projected/871672d1-30f1-4221-a434-2a8d60173f46-kube-api-access-fm6wp\") pod \"metallb-operator-controller-manager-648849f9d4-9zr5p\" (UID: \"871672d1-30f1-4221-a434-2a8d60173f46\") " pod="metallb-system/metallb-operator-controller-manager-648849f9d4-9zr5p" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.548245 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/871672d1-30f1-4221-a434-2a8d60173f46-apiservice-cert\") pod \"metallb-operator-controller-manager-648849f9d4-9zr5p\" (UID: \"871672d1-30f1-4221-a434-2a8d60173f46\") " pod="metallb-system/metallb-operator-controller-manager-648849f9d4-9zr5p" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.548293 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/871672d1-30f1-4221-a434-2a8d60173f46-webhook-cert\") pod \"metallb-operator-controller-manager-648849f9d4-9zr5p\" (UID: \"871672d1-30f1-4221-a434-2a8d60173f46\") " pod="metallb-system/metallb-operator-controller-manager-648849f9d4-9zr5p" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.553942 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/871672d1-30f1-4221-a434-2a8d60173f46-apiservice-cert\") pod \"metallb-operator-controller-manager-648849f9d4-9zr5p\" (UID: \"871672d1-30f1-4221-a434-2a8d60173f46\") " pod="metallb-system/metallb-operator-controller-manager-648849f9d4-9zr5p" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.558714 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/871672d1-30f1-4221-a434-2a8d60173f46-webhook-cert\") pod \"metallb-operator-controller-manager-648849f9d4-9zr5p\" (UID: \"871672d1-30f1-4221-a434-2a8d60173f46\") " pod="metallb-system/metallb-operator-controller-manager-648849f9d4-9zr5p" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.568045 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm6wp\" (UniqueName: \"kubernetes.io/projected/871672d1-30f1-4221-a434-2a8d60173f46-kube-api-access-fm6wp\") pod \"metallb-operator-controller-manager-648849f9d4-9zr5p\" (UID: \"871672d1-30f1-4221-a434-2a8d60173f46\") " pod="metallb-system/metallb-operator-controller-manager-648849f9d4-9zr5p" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.721250 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-648849f9d4-9zr5p" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.752651 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-bf6cfb5cf-tllp8"] Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.753345 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-bf6cfb5cf-tllp8" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.756555 4950 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.756826 4950 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-98tbq" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.758295 4950 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.827923 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-bf6cfb5cf-tllp8"] Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.974700 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/210bdf50-0f58-4342-af2c-e0df41ac3e91-apiservice-cert\") pod \"metallb-operator-webhook-server-bf6cfb5cf-tllp8\" (UID: \"210bdf50-0f58-4342-af2c-e0df41ac3e91\") " pod="metallb-system/metallb-operator-webhook-server-bf6cfb5cf-tllp8" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.975444 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/210bdf50-0f58-4342-af2c-e0df41ac3e91-webhook-cert\") pod \"metallb-operator-webhook-server-bf6cfb5cf-tllp8\" (UID: \"210bdf50-0f58-4342-af2c-e0df41ac3e91\") " pod="metallb-system/metallb-operator-webhook-server-bf6cfb5cf-tllp8" Mar 18 20:20:47 crc kubenswrapper[4950]: I0318 20:20:47.975602 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h84c\" (UniqueName: \"kubernetes.io/projected/210bdf50-0f58-4342-af2c-e0df41ac3e91-kube-api-access-8h84c\") pod \"metallb-operator-webhook-server-bf6cfb5cf-tllp8\" (UID: \"210bdf50-0f58-4342-af2c-e0df41ac3e91\") " pod="metallb-system/metallb-operator-webhook-server-bf6cfb5cf-tllp8" Mar 18 20:20:48 crc kubenswrapper[4950]: I0318 20:20:48.076901 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/210bdf50-0f58-4342-af2c-e0df41ac3e91-apiservice-cert\") pod \"metallb-operator-webhook-server-bf6cfb5cf-tllp8\" (UID: \"210bdf50-0f58-4342-af2c-e0df41ac3e91\") " pod="metallb-system/metallb-operator-webhook-server-bf6cfb5cf-tllp8" Mar 18 20:20:48 crc kubenswrapper[4950]: I0318 20:20:48.076941 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/210bdf50-0f58-4342-af2c-e0df41ac3e91-webhook-cert\") pod \"metallb-operator-webhook-server-bf6cfb5cf-tllp8\" (UID: \"210bdf50-0f58-4342-af2c-e0df41ac3e91\") " pod="metallb-system/metallb-operator-webhook-server-bf6cfb5cf-tllp8" Mar 18 20:20:48 crc kubenswrapper[4950]: I0318 20:20:48.076981 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h84c\" (UniqueName: \"kubernetes.io/projected/210bdf50-0f58-4342-af2c-e0df41ac3e91-kube-api-access-8h84c\") pod \"metallb-operator-webhook-server-bf6cfb5cf-tllp8\" (UID: \"210bdf50-0f58-4342-af2c-e0df41ac3e91\") " pod="metallb-system/metallb-operator-webhook-server-bf6cfb5cf-tllp8" Mar 18 20:20:48 crc kubenswrapper[4950]: I0318 20:20:48.084218 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/210bdf50-0f58-4342-af2c-e0df41ac3e91-webhook-cert\") pod \"metallb-operator-webhook-server-bf6cfb5cf-tllp8\" (UID: \"210bdf50-0f58-4342-af2c-e0df41ac3e91\") " pod="metallb-system/metallb-operator-webhook-server-bf6cfb5cf-tllp8" Mar 18 20:20:48 crc kubenswrapper[4950]: I0318 20:20:48.087341 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/210bdf50-0f58-4342-af2c-e0df41ac3e91-apiservice-cert\") pod \"metallb-operator-webhook-server-bf6cfb5cf-tllp8\" (UID: \"210bdf50-0f58-4342-af2c-e0df41ac3e91\") " pod="metallb-system/metallb-operator-webhook-server-bf6cfb5cf-tllp8" Mar 18 20:20:48 crc kubenswrapper[4950]: I0318 20:20:48.108906 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-648849f9d4-9zr5p"] Mar 18 20:20:48 crc kubenswrapper[4950]: I0318 20:20:48.112896 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h84c\" (UniqueName: \"kubernetes.io/projected/210bdf50-0f58-4342-af2c-e0df41ac3e91-kube-api-access-8h84c\") pod \"metallb-operator-webhook-server-bf6cfb5cf-tllp8\" (UID: \"210bdf50-0f58-4342-af2c-e0df41ac3e91\") " pod="metallb-system/metallb-operator-webhook-server-bf6cfb5cf-tllp8" Mar 18 20:20:48 crc kubenswrapper[4950]: I0318 20:20:48.368810 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-648849f9d4-9zr5p" event={"ID":"871672d1-30f1-4221-a434-2a8d60173f46","Type":"ContainerStarted","Data":"ff32ebc92b2be093438b6b10d4a69af1769e0c68ea79e7dd9652d22a30ee5e55"} Mar 18 20:20:48 crc kubenswrapper[4950]: I0318 20:20:48.398181 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-bf6cfb5cf-tllp8" Mar 18 20:20:48 crc kubenswrapper[4950]: I0318 20:20:48.649162 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-bf6cfb5cf-tllp8"] Mar 18 20:20:49 crc kubenswrapper[4950]: I0318 20:20:49.377570 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-bf6cfb5cf-tllp8" event={"ID":"210bdf50-0f58-4342-af2c-e0df41ac3e91","Type":"ContainerStarted","Data":"bc9e077f9cc2afb61c100a3bcf815d3b69386765050078471fa704888a41be8b"} Mar 18 20:20:52 crc kubenswrapper[4950]: I0318 20:20:52.421384 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-648849f9d4-9zr5p" event={"ID":"871672d1-30f1-4221-a434-2a8d60173f46","Type":"ContainerStarted","Data":"64c38ebc3db2e9c0b9f7b04c92fd9fcde556d1fbd5982e91d25695ea13e43bf2"} Mar 18 20:20:52 crc kubenswrapper[4950]: I0318 20:20:52.422592 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-648849f9d4-9zr5p" Mar 18 20:20:52 crc kubenswrapper[4950]: I0318 20:20:52.443809 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-648849f9d4-9zr5p" podStartSLOduration=1.685323474 podStartE2EDuration="5.443786561s" podCreationTimestamp="2026-03-18 20:20:47 +0000 UTC" firstStartedPulling="2026-03-18 20:20:48.119757824 +0000 UTC m=+861.360599692" lastFinishedPulling="2026-03-18 20:20:51.878220911 +0000 UTC m=+865.119062779" observedRunningTime="2026-03-18 20:20:52.439298053 +0000 UTC m=+865.680139921" watchObservedRunningTime="2026-03-18 20:20:52.443786561 +0000 UTC m=+865.684628439" Mar 18 20:20:53 crc kubenswrapper[4950]: I0318 20:20:53.013838 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7tfkw" Mar 18 20:20:53 crc kubenswrapper[4950]: I0318 20:20:53.061224 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7tfkw" Mar 18 20:20:54 crc kubenswrapper[4950]: I0318 20:20:54.435096 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-bf6cfb5cf-tllp8" event={"ID":"210bdf50-0f58-4342-af2c-e0df41ac3e91","Type":"ContainerStarted","Data":"f47290e0da202dcee5268a00e45673dfd99274ec13fc5231c6d62ee2ac299b9d"} Mar 18 20:20:54 crc kubenswrapper[4950]: I0318 20:20:54.463999 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-bf6cfb5cf-tllp8" podStartSLOduration=2.394209501 podStartE2EDuration="7.463981755s" podCreationTimestamp="2026-03-18 20:20:47 +0000 UTC" firstStartedPulling="2026-03-18 20:20:48.658866418 +0000 UTC m=+861.899708286" lastFinishedPulling="2026-03-18 20:20:53.728638662 +0000 UTC m=+866.969480540" observedRunningTime="2026-03-18 20:20:54.463063501 +0000 UTC m=+867.703905369" watchObservedRunningTime="2026-03-18 20:20:54.463981755 +0000 UTC m=+867.704823623" Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.026578 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7tfkw"] Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.026816 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7tfkw" podUID="ca3038fa-120b-412e-80b2-24aca2dd19e8" containerName="registry-server" containerID="cri-o://95cdad0f9ff05b1ff2cacc5b7342267ecc72202391456ee0b088de53e2cac6a0" gracePeriod=2 Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.459497 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tfkw" Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.467673 4950 generic.go:334] "Generic (PLEG): container finished" podID="ca3038fa-120b-412e-80b2-24aca2dd19e8" containerID="95cdad0f9ff05b1ff2cacc5b7342267ecc72202391456ee0b088de53e2cac6a0" exitCode=0 Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.468749 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tfkw" Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.468873 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tfkw" event={"ID":"ca3038fa-120b-412e-80b2-24aca2dd19e8","Type":"ContainerDied","Data":"95cdad0f9ff05b1ff2cacc5b7342267ecc72202391456ee0b088de53e2cac6a0"} Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.468905 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-bf6cfb5cf-tllp8" Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.468916 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tfkw" event={"ID":"ca3038fa-120b-412e-80b2-24aca2dd19e8","Type":"ContainerDied","Data":"504f87f073a1d77b20ced1f43a36ef60a28fb0ef005c3206892a6973651e9591"} Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.468934 4950 scope.go:117] "RemoveContainer" containerID="95cdad0f9ff05b1ff2cacc5b7342267ecc72202391456ee0b088de53e2cac6a0" Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.487233 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca3038fa-120b-412e-80b2-24aca2dd19e8-catalog-content\") pod \"ca3038fa-120b-412e-80b2-24aca2dd19e8\" (UID: \"ca3038fa-120b-412e-80b2-24aca2dd19e8\") " Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.487306 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxg5g\" (UniqueName: \"kubernetes.io/projected/ca3038fa-120b-412e-80b2-24aca2dd19e8-kube-api-access-pxg5g\") pod \"ca3038fa-120b-412e-80b2-24aca2dd19e8\" (UID: \"ca3038fa-120b-412e-80b2-24aca2dd19e8\") " Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.487347 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca3038fa-120b-412e-80b2-24aca2dd19e8-utilities\") pod \"ca3038fa-120b-412e-80b2-24aca2dd19e8\" (UID: \"ca3038fa-120b-412e-80b2-24aca2dd19e8\") " Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.488264 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca3038fa-120b-412e-80b2-24aca2dd19e8-utilities" (OuterVolumeSpecName: "utilities") pod "ca3038fa-120b-412e-80b2-24aca2dd19e8" (UID: "ca3038fa-120b-412e-80b2-24aca2dd19e8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.496990 4950 scope.go:117] "RemoveContainer" containerID="5f1752e256cd8af351e0e47ae0db4e862d1d884a40e79ca93919c7c562a6f404" Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.503074 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca3038fa-120b-412e-80b2-24aca2dd19e8-kube-api-access-pxg5g" (OuterVolumeSpecName: "kube-api-access-pxg5g") pod "ca3038fa-120b-412e-80b2-24aca2dd19e8" (UID: "ca3038fa-120b-412e-80b2-24aca2dd19e8"). InnerVolumeSpecName "kube-api-access-pxg5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.547954 4950 scope.go:117] "RemoveContainer" containerID="8af013eb61caabc1a7a29cb9ebcbaa4e824cd87b0b96803895878d5100156b8f" Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.574798 4950 scope.go:117] "RemoveContainer" containerID="95cdad0f9ff05b1ff2cacc5b7342267ecc72202391456ee0b088de53e2cac6a0" Mar 18 20:20:55 crc kubenswrapper[4950]: E0318 20:20:55.575359 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95cdad0f9ff05b1ff2cacc5b7342267ecc72202391456ee0b088de53e2cac6a0\": container with ID starting with 95cdad0f9ff05b1ff2cacc5b7342267ecc72202391456ee0b088de53e2cac6a0 not found: ID does not exist" containerID="95cdad0f9ff05b1ff2cacc5b7342267ecc72202391456ee0b088de53e2cac6a0" Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.575451 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95cdad0f9ff05b1ff2cacc5b7342267ecc72202391456ee0b088de53e2cac6a0"} err="failed to get container status \"95cdad0f9ff05b1ff2cacc5b7342267ecc72202391456ee0b088de53e2cac6a0\": rpc error: code = NotFound desc = could not find container \"95cdad0f9ff05b1ff2cacc5b7342267ecc72202391456ee0b088de53e2cac6a0\": container with ID starting with 95cdad0f9ff05b1ff2cacc5b7342267ecc72202391456ee0b088de53e2cac6a0 not found: ID does not exist" Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.575479 4950 scope.go:117] "RemoveContainer" containerID="5f1752e256cd8af351e0e47ae0db4e862d1d884a40e79ca93919c7c562a6f404" Mar 18 20:20:55 crc kubenswrapper[4950]: E0318 20:20:55.575867 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f1752e256cd8af351e0e47ae0db4e862d1d884a40e79ca93919c7c562a6f404\": container with ID starting with 5f1752e256cd8af351e0e47ae0db4e862d1d884a40e79ca93919c7c562a6f404 not found: ID does not exist" containerID="5f1752e256cd8af351e0e47ae0db4e862d1d884a40e79ca93919c7c562a6f404" Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.575923 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f1752e256cd8af351e0e47ae0db4e862d1d884a40e79ca93919c7c562a6f404"} err="failed to get container status \"5f1752e256cd8af351e0e47ae0db4e862d1d884a40e79ca93919c7c562a6f404\": rpc error: code = NotFound desc = could not find container \"5f1752e256cd8af351e0e47ae0db4e862d1d884a40e79ca93919c7c562a6f404\": container with ID starting with 5f1752e256cd8af351e0e47ae0db4e862d1d884a40e79ca93919c7c562a6f404 not found: ID does not exist" Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.575950 4950 scope.go:117] "RemoveContainer" containerID="8af013eb61caabc1a7a29cb9ebcbaa4e824cd87b0b96803895878d5100156b8f" Mar 18 20:20:55 crc kubenswrapper[4950]: E0318 20:20:55.576234 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8af013eb61caabc1a7a29cb9ebcbaa4e824cd87b0b96803895878d5100156b8f\": container with ID starting with 8af013eb61caabc1a7a29cb9ebcbaa4e824cd87b0b96803895878d5100156b8f not found: ID does not exist" containerID="8af013eb61caabc1a7a29cb9ebcbaa4e824cd87b0b96803895878d5100156b8f" Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.576264 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8af013eb61caabc1a7a29cb9ebcbaa4e824cd87b0b96803895878d5100156b8f"} err="failed to get container status \"8af013eb61caabc1a7a29cb9ebcbaa4e824cd87b0b96803895878d5100156b8f\": rpc error: code = NotFound desc = could not find container \"8af013eb61caabc1a7a29cb9ebcbaa4e824cd87b0b96803895878d5100156b8f\": container with ID starting with 8af013eb61caabc1a7a29cb9ebcbaa4e824cd87b0b96803895878d5100156b8f not found: ID does not exist" Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.589832 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca3038fa-120b-412e-80b2-24aca2dd19e8-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.589856 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxg5g\" (UniqueName: \"kubernetes.io/projected/ca3038fa-120b-412e-80b2-24aca2dd19e8-kube-api-access-pxg5g\") on node \"crc\" DevicePath \"\"" Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.654005 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca3038fa-120b-412e-80b2-24aca2dd19e8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ca3038fa-120b-412e-80b2-24aca2dd19e8" (UID: "ca3038fa-120b-412e-80b2-24aca2dd19e8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.691025 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca3038fa-120b-412e-80b2-24aca2dd19e8-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.803566 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7tfkw"] Mar 18 20:20:55 crc kubenswrapper[4950]: I0318 20:20:55.808364 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7tfkw"] Mar 18 20:20:57 crc kubenswrapper[4950]: I0318 20:20:57.491666 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca3038fa-120b-412e-80b2-24aca2dd19e8" path="/var/lib/kubelet/pods/ca3038fa-120b-412e-80b2-24aca2dd19e8/volumes" Mar 18 20:21:03 crc kubenswrapper[4950]: I0318 20:21:03.835938 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:21:03 crc kubenswrapper[4950]: I0318 20:21:03.836266 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:21:08 crc kubenswrapper[4950]: I0318 20:21:08.403830 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-bf6cfb5cf-tllp8" Mar 18 20:21:27 crc kubenswrapper[4950]: I0318 20:21:27.725033 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-648849f9d4-9zr5p" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.501232 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-mlhr4"] Mar 18 20:21:28 crc kubenswrapper[4950]: E0318 20:21:28.501803 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca3038fa-120b-412e-80b2-24aca2dd19e8" containerName="extract-content" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.501818 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca3038fa-120b-412e-80b2-24aca2dd19e8" containerName="extract-content" Mar 18 20:21:28 crc kubenswrapper[4950]: E0318 20:21:28.501829 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca3038fa-120b-412e-80b2-24aca2dd19e8" containerName="registry-server" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.501835 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca3038fa-120b-412e-80b2-24aca2dd19e8" containerName="registry-server" Mar 18 20:21:28 crc kubenswrapper[4950]: E0318 20:21:28.501841 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca3038fa-120b-412e-80b2-24aca2dd19e8" containerName="extract-utilities" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.501847 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca3038fa-120b-412e-80b2-24aca2dd19e8" containerName="extract-utilities" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.501941 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca3038fa-120b-412e-80b2-24aca2dd19e8" containerName="registry-server" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.503645 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.505045 4950 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.505613 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.505923 4950 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-2j9xk" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.508102 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-nqppx"] Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.508987 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nqppx" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.512718 4950 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.543404 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-nqppx"] Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.570144 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d221ef79-335c-46db-a491-b74bc07300c4-frr-conf\") pod \"frr-k8s-mlhr4\" (UID: \"d221ef79-335c-46db-a491-b74bc07300c4\") " pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.570208 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d221ef79-335c-46db-a491-b74bc07300c4-reloader\") pod \"frr-k8s-mlhr4\" (UID: \"d221ef79-335c-46db-a491-b74bc07300c4\") " pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.570241 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d221ef79-335c-46db-a491-b74bc07300c4-metrics\") pod \"frr-k8s-mlhr4\" (UID: \"d221ef79-335c-46db-a491-b74bc07300c4\") " pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.570274 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b4c0afa4-f131-44df-b2f1-9908b3b578e6-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-nqppx\" (UID: \"b4c0afa4-f131-44df-b2f1-9908b3b578e6\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nqppx" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.570299 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d221ef79-335c-46db-a491-b74bc07300c4-metrics-certs\") pod \"frr-k8s-mlhr4\" (UID: \"d221ef79-335c-46db-a491-b74bc07300c4\") " pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.570314 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pksr\" (UniqueName: \"kubernetes.io/projected/d221ef79-335c-46db-a491-b74bc07300c4-kube-api-access-5pksr\") pod \"frr-k8s-mlhr4\" (UID: \"d221ef79-335c-46db-a491-b74bc07300c4\") " pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.570349 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvjqv\" (UniqueName: \"kubernetes.io/projected/b4c0afa4-f131-44df-b2f1-9908b3b578e6-kube-api-access-vvjqv\") pod \"frr-k8s-webhook-server-bcc4b6f68-nqppx\" (UID: \"b4c0afa4-f131-44df-b2f1-9908b3b578e6\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nqppx" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.570394 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d221ef79-335c-46db-a491-b74bc07300c4-frr-sockets\") pod \"frr-k8s-mlhr4\" (UID: \"d221ef79-335c-46db-a491-b74bc07300c4\") " pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.570443 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d221ef79-335c-46db-a491-b74bc07300c4-frr-startup\") pod \"frr-k8s-mlhr4\" (UID: \"d221ef79-335c-46db-a491-b74bc07300c4\") " pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.621392 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-6k4zn"] Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.622280 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-6k4zn" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.626692 4950 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.626891 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.626965 4950 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-ml7hx" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.627039 4950 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.630159 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-7bb4cc7c98-mhk7w"] Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.631107 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-mhk7w" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.641798 4950 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.650020 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-mhk7w"] Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.678541 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d221ef79-335c-46db-a491-b74bc07300c4-frr-conf\") pod \"frr-k8s-mlhr4\" (UID: \"d221ef79-335c-46db-a491-b74bc07300c4\") " pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.678590 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d221ef79-335c-46db-a491-b74bc07300c4-reloader\") pod \"frr-k8s-mlhr4\" (UID: \"d221ef79-335c-46db-a491-b74bc07300c4\") " pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.678611 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d221ef79-335c-46db-a491-b74bc07300c4-metrics\") pod \"frr-k8s-mlhr4\" (UID: \"d221ef79-335c-46db-a491-b74bc07300c4\") " pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.678632 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b4c0afa4-f131-44df-b2f1-9908b3b578e6-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-nqppx\" (UID: \"b4c0afa4-f131-44df-b2f1-9908b3b578e6\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nqppx" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.678656 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d221ef79-335c-46db-a491-b74bc07300c4-metrics-certs\") pod \"frr-k8s-mlhr4\" (UID: \"d221ef79-335c-46db-a491-b74bc07300c4\") " pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.678681 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pksr\" (UniqueName: \"kubernetes.io/projected/d221ef79-335c-46db-a491-b74bc07300c4-kube-api-access-5pksr\") pod \"frr-k8s-mlhr4\" (UID: \"d221ef79-335c-46db-a491-b74bc07300c4\") " pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.678699 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvjqv\" (UniqueName: \"kubernetes.io/projected/b4c0afa4-f131-44df-b2f1-9908b3b578e6-kube-api-access-vvjqv\") pod \"frr-k8s-webhook-server-bcc4b6f68-nqppx\" (UID: \"b4c0afa4-f131-44df-b2f1-9908b3b578e6\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nqppx" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.678716 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d221ef79-335c-46db-a491-b74bc07300c4-frr-sockets\") pod \"frr-k8s-mlhr4\" (UID: \"d221ef79-335c-46db-a491-b74bc07300c4\") " pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.678732 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d221ef79-335c-46db-a491-b74bc07300c4-frr-startup\") pod \"frr-k8s-mlhr4\" (UID: \"d221ef79-335c-46db-a491-b74bc07300c4\") " pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:28 crc kubenswrapper[4950]: E0318 20:21:28.684507 4950 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Mar 18 20:21:28 crc kubenswrapper[4950]: E0318 20:21:28.684564 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b4c0afa4-f131-44df-b2f1-9908b3b578e6-cert podName:b4c0afa4-f131-44df-b2f1-9908b3b578e6 nodeName:}" failed. No retries permitted until 2026-03-18 20:21:29.184547998 +0000 UTC m=+902.425389866 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b4c0afa4-f131-44df-b2f1-9908b3b578e6-cert") pod "frr-k8s-webhook-server-bcc4b6f68-nqppx" (UID: "b4c0afa4-f131-44df-b2f1-9908b3b578e6") : secret "frr-k8s-webhook-server-cert" not found Mar 18 20:21:28 crc kubenswrapper[4950]: E0318 20:21:28.684756 4950 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Mar 18 20:21:28 crc kubenswrapper[4950]: E0318 20:21:28.684791 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d221ef79-335c-46db-a491-b74bc07300c4-metrics-certs podName:d221ef79-335c-46db-a491-b74bc07300c4 nodeName:}" failed. No retries permitted until 2026-03-18 20:21:29.184782034 +0000 UTC m=+902.425623902 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d221ef79-335c-46db-a491-b74bc07300c4-metrics-certs") pod "frr-k8s-mlhr4" (UID: "d221ef79-335c-46db-a491-b74bc07300c4") : secret "frr-k8s-certs-secret" not found Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.684835 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d221ef79-335c-46db-a491-b74bc07300c4-frr-startup\") pod \"frr-k8s-mlhr4\" (UID: \"d221ef79-335c-46db-a491-b74bc07300c4\") " pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.684867 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d221ef79-335c-46db-a491-b74bc07300c4-frr-conf\") pod \"frr-k8s-mlhr4\" (UID: \"d221ef79-335c-46db-a491-b74bc07300c4\") " pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.684971 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d221ef79-335c-46db-a491-b74bc07300c4-metrics\") pod \"frr-k8s-mlhr4\" (UID: \"d221ef79-335c-46db-a491-b74bc07300c4\") " pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.685139 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d221ef79-335c-46db-a491-b74bc07300c4-reloader\") pod \"frr-k8s-mlhr4\" (UID: \"d221ef79-335c-46db-a491-b74bc07300c4\") " pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.685156 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d221ef79-335c-46db-a491-b74bc07300c4-frr-sockets\") pod \"frr-k8s-mlhr4\" (UID: \"d221ef79-335c-46db-a491-b74bc07300c4\") " pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.732484 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvjqv\" (UniqueName: \"kubernetes.io/projected/b4c0afa4-f131-44df-b2f1-9908b3b578e6-kube-api-access-vvjqv\") pod \"frr-k8s-webhook-server-bcc4b6f68-nqppx\" (UID: \"b4c0afa4-f131-44df-b2f1-9908b3b578e6\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nqppx" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.741100 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pksr\" (UniqueName: \"kubernetes.io/projected/d221ef79-335c-46db-a491-b74bc07300c4-kube-api-access-5pksr\") pod \"frr-k8s-mlhr4\" (UID: \"d221ef79-335c-46db-a491-b74bc07300c4\") " pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.779594 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6f8ab895-b55a-43e6-a6fd-71f4ac42a50a-metrics-certs\") pod \"speaker-6k4zn\" (UID: \"6f8ab895-b55a-43e6-a6fd-71f4ac42a50a\") " pod="metallb-system/speaker-6k4zn" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.779663 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6f8ab895-b55a-43e6-a6fd-71f4ac42a50a-memberlist\") pod \"speaker-6k4zn\" (UID: \"6f8ab895-b55a-43e6-a6fd-71f4ac42a50a\") " pod="metallb-system/speaker-6k4zn" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.779707 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbx5x\" (UniqueName: \"kubernetes.io/projected/db65cfec-9483-4641-8794-5fbd50c4bc92-kube-api-access-wbx5x\") pod \"controller-7bb4cc7c98-mhk7w\" (UID: \"db65cfec-9483-4641-8794-5fbd50c4bc92\") " pod="metallb-system/controller-7bb4cc7c98-mhk7w" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.779724 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r596p\" (UniqueName: \"kubernetes.io/projected/6f8ab895-b55a-43e6-a6fd-71f4ac42a50a-kube-api-access-r596p\") pod \"speaker-6k4zn\" (UID: \"6f8ab895-b55a-43e6-a6fd-71f4ac42a50a\") " pod="metallb-system/speaker-6k4zn" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.779773 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6f8ab895-b55a-43e6-a6fd-71f4ac42a50a-metallb-excludel2\") pod \"speaker-6k4zn\" (UID: \"6f8ab895-b55a-43e6-a6fd-71f4ac42a50a\") " pod="metallb-system/speaker-6k4zn" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.779830 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db65cfec-9483-4641-8794-5fbd50c4bc92-metrics-certs\") pod \"controller-7bb4cc7c98-mhk7w\" (UID: \"db65cfec-9483-4641-8794-5fbd50c4bc92\") " pod="metallb-system/controller-7bb4cc7c98-mhk7w" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.779849 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db65cfec-9483-4641-8794-5fbd50c4bc92-cert\") pod \"controller-7bb4cc7c98-mhk7w\" (UID: \"db65cfec-9483-4641-8794-5fbd50c4bc92\") " pod="metallb-system/controller-7bb4cc7c98-mhk7w" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.880657 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6f8ab895-b55a-43e6-a6fd-71f4ac42a50a-memberlist\") pod \"speaker-6k4zn\" (UID: \"6f8ab895-b55a-43e6-a6fd-71f4ac42a50a\") " pod="metallb-system/speaker-6k4zn" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.880721 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbx5x\" (UniqueName: \"kubernetes.io/projected/db65cfec-9483-4641-8794-5fbd50c4bc92-kube-api-access-wbx5x\") pod \"controller-7bb4cc7c98-mhk7w\" (UID: \"db65cfec-9483-4641-8794-5fbd50c4bc92\") " pod="metallb-system/controller-7bb4cc7c98-mhk7w" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.880736 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r596p\" (UniqueName: \"kubernetes.io/projected/6f8ab895-b55a-43e6-a6fd-71f4ac42a50a-kube-api-access-r596p\") pod \"speaker-6k4zn\" (UID: \"6f8ab895-b55a-43e6-a6fd-71f4ac42a50a\") " pod="metallb-system/speaker-6k4zn" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.880751 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6f8ab895-b55a-43e6-a6fd-71f4ac42a50a-metallb-excludel2\") pod \"speaker-6k4zn\" (UID: \"6f8ab895-b55a-43e6-a6fd-71f4ac42a50a\") " pod="metallb-system/speaker-6k4zn" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.880788 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db65cfec-9483-4641-8794-5fbd50c4bc92-cert\") pod \"controller-7bb4cc7c98-mhk7w\" (UID: \"db65cfec-9483-4641-8794-5fbd50c4bc92\") " pod="metallb-system/controller-7bb4cc7c98-mhk7w" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.880802 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db65cfec-9483-4641-8794-5fbd50c4bc92-metrics-certs\") pod \"controller-7bb4cc7c98-mhk7w\" (UID: \"db65cfec-9483-4641-8794-5fbd50c4bc92\") " pod="metallb-system/controller-7bb4cc7c98-mhk7w" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.880830 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6f8ab895-b55a-43e6-a6fd-71f4ac42a50a-metrics-certs\") pod \"speaker-6k4zn\" (UID: \"6f8ab895-b55a-43e6-a6fd-71f4ac42a50a\") " pod="metallb-system/speaker-6k4zn" Mar 18 20:21:28 crc kubenswrapper[4950]: E0318 20:21:28.880934 4950 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Mar 18 20:21:28 crc kubenswrapper[4950]: E0318 20:21:28.880979 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f8ab895-b55a-43e6-a6fd-71f4ac42a50a-metrics-certs podName:6f8ab895-b55a-43e6-a6fd-71f4ac42a50a nodeName:}" failed. No retries permitted until 2026-03-18 20:21:29.380964992 +0000 UTC m=+902.621806860 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6f8ab895-b55a-43e6-a6fd-71f4ac42a50a-metrics-certs") pod "speaker-6k4zn" (UID: "6f8ab895-b55a-43e6-a6fd-71f4ac42a50a") : secret "speaker-certs-secret" not found Mar 18 20:21:28 crc kubenswrapper[4950]: E0318 20:21:28.881159 4950 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 18 20:21:28 crc kubenswrapper[4950]: E0318 20:21:28.881179 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f8ab895-b55a-43e6-a6fd-71f4ac42a50a-memberlist podName:6f8ab895-b55a-43e6-a6fd-71f4ac42a50a nodeName:}" failed. No retries permitted until 2026-03-18 20:21:29.381172958 +0000 UTC m=+902.622014826 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/6f8ab895-b55a-43e6-a6fd-71f4ac42a50a-memberlist") pod "speaker-6k4zn" (UID: "6f8ab895-b55a-43e6-a6fd-71f4ac42a50a") : secret "metallb-memberlist" not found Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.881960 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6f8ab895-b55a-43e6-a6fd-71f4ac42a50a-metallb-excludel2\") pod \"speaker-6k4zn\" (UID: \"6f8ab895-b55a-43e6-a6fd-71f4ac42a50a\") " pod="metallb-system/speaker-6k4zn" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.885269 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db65cfec-9483-4641-8794-5fbd50c4bc92-metrics-certs\") pod \"controller-7bb4cc7c98-mhk7w\" (UID: \"db65cfec-9483-4641-8794-5fbd50c4bc92\") " pod="metallb-system/controller-7bb4cc7c98-mhk7w" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.886591 4950 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.897567 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbx5x\" (UniqueName: \"kubernetes.io/projected/db65cfec-9483-4641-8794-5fbd50c4bc92-kube-api-access-wbx5x\") pod \"controller-7bb4cc7c98-mhk7w\" (UID: \"db65cfec-9483-4641-8794-5fbd50c4bc92\") " pod="metallb-system/controller-7bb4cc7c98-mhk7w" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.898872 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db65cfec-9483-4641-8794-5fbd50c4bc92-cert\") pod \"controller-7bb4cc7c98-mhk7w\" (UID: \"db65cfec-9483-4641-8794-5fbd50c4bc92\") " pod="metallb-system/controller-7bb4cc7c98-mhk7w" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.899078 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r596p\" (UniqueName: \"kubernetes.io/projected/6f8ab895-b55a-43e6-a6fd-71f4ac42a50a-kube-api-access-r596p\") pod \"speaker-6k4zn\" (UID: \"6f8ab895-b55a-43e6-a6fd-71f4ac42a50a\") " pod="metallb-system/speaker-6k4zn" Mar 18 20:21:28 crc kubenswrapper[4950]: I0318 20:21:28.959811 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-mhk7w" Mar 18 20:21:29 crc kubenswrapper[4950]: I0318 20:21:29.285239 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b4c0afa4-f131-44df-b2f1-9908b3b578e6-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-nqppx\" (UID: \"b4c0afa4-f131-44df-b2f1-9908b3b578e6\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nqppx" Mar 18 20:21:29 crc kubenswrapper[4950]: I0318 20:21:29.285521 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d221ef79-335c-46db-a491-b74bc07300c4-metrics-certs\") pod \"frr-k8s-mlhr4\" (UID: \"d221ef79-335c-46db-a491-b74bc07300c4\") " pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:29 crc kubenswrapper[4950]: I0318 20:21:29.288395 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d221ef79-335c-46db-a491-b74bc07300c4-metrics-certs\") pod \"frr-k8s-mlhr4\" (UID: \"d221ef79-335c-46db-a491-b74bc07300c4\") " pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:29 crc kubenswrapper[4950]: I0318 20:21:29.288764 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b4c0afa4-f131-44df-b2f1-9908b3b578e6-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-nqppx\" (UID: \"b4c0afa4-f131-44df-b2f1-9908b3b578e6\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nqppx" Mar 18 20:21:29 crc kubenswrapper[4950]: I0318 20:21:29.386975 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6f8ab895-b55a-43e6-a6fd-71f4ac42a50a-metrics-certs\") pod \"speaker-6k4zn\" (UID: \"6f8ab895-b55a-43e6-a6fd-71f4ac42a50a\") " pod="metallb-system/speaker-6k4zn" Mar 18 20:21:29 crc kubenswrapper[4950]: I0318 20:21:29.387111 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6f8ab895-b55a-43e6-a6fd-71f4ac42a50a-memberlist\") pod \"speaker-6k4zn\" (UID: \"6f8ab895-b55a-43e6-a6fd-71f4ac42a50a\") " pod="metallb-system/speaker-6k4zn" Mar 18 20:21:29 crc kubenswrapper[4950]: E0318 20:21:29.387516 4950 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 18 20:21:29 crc kubenswrapper[4950]: E0318 20:21:29.387623 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f8ab895-b55a-43e6-a6fd-71f4ac42a50a-memberlist podName:6f8ab895-b55a-43e6-a6fd-71f4ac42a50a nodeName:}" failed. No retries permitted until 2026-03-18 20:21:30.387592533 +0000 UTC m=+903.628434451 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/6f8ab895-b55a-43e6-a6fd-71f4ac42a50a-memberlist") pod "speaker-6k4zn" (UID: "6f8ab895-b55a-43e6-a6fd-71f4ac42a50a") : secret "metallb-memberlist" not found Mar 18 20:21:29 crc kubenswrapper[4950]: I0318 20:21:29.393147 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6f8ab895-b55a-43e6-a6fd-71f4ac42a50a-metrics-certs\") pod \"speaker-6k4zn\" (UID: \"6f8ab895-b55a-43e6-a6fd-71f4ac42a50a\") " pod="metallb-system/speaker-6k4zn" Mar 18 20:21:29 crc kubenswrapper[4950]: I0318 20:21:29.401640 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-mhk7w"] Mar 18 20:21:29 crc kubenswrapper[4950]: W0318 20:21:29.413640 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb65cfec_9483_4641_8794_5fbd50c4bc92.slice/crio-5e3af07b8d34a66ab9df05554de62d2d76be8849e397eef687daebaf71406371 WatchSource:0}: Error finding container 5e3af07b8d34a66ab9df05554de62d2d76be8849e397eef687daebaf71406371: Status 404 returned error can't find the container with id 5e3af07b8d34a66ab9df05554de62d2d76be8849e397eef687daebaf71406371 Mar 18 20:21:29 crc kubenswrapper[4950]: I0318 20:21:29.427499 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:29 crc kubenswrapper[4950]: I0318 20:21:29.452049 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nqppx" Mar 18 20:21:29 crc kubenswrapper[4950]: I0318 20:21:29.557035 4950 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 20:21:29 crc kubenswrapper[4950]: I0318 20:21:29.696097 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-nqppx"] Mar 18 20:21:29 crc kubenswrapper[4950]: W0318 20:21:29.704181 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4c0afa4_f131_44df_b2f1_9908b3b578e6.slice/crio-69ac267cd4b6e4276d5b0768fada4bef78ff2dd03edb0190f58755077e1bbfe0 WatchSource:0}: Error finding container 69ac267cd4b6e4276d5b0768fada4bef78ff2dd03edb0190f58755077e1bbfe0: Status 404 returned error can't find the container with id 69ac267cd4b6e4276d5b0768fada4bef78ff2dd03edb0190f58755077e1bbfe0 Mar 18 20:21:30 crc kubenswrapper[4950]: I0318 20:21:30.217597 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nqppx" event={"ID":"b4c0afa4-f131-44df-b2f1-9908b3b578e6","Type":"ContainerStarted","Data":"69ac267cd4b6e4276d5b0768fada4bef78ff2dd03edb0190f58755077e1bbfe0"} Mar 18 20:21:30 crc kubenswrapper[4950]: I0318 20:21:30.219510 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mlhr4" event={"ID":"d221ef79-335c-46db-a491-b74bc07300c4","Type":"ContainerStarted","Data":"61057bc0200a1d24db5c51bf837e24eb333c06394aeac9edc54c143b1926d085"} Mar 18 20:21:30 crc kubenswrapper[4950]: I0318 20:21:30.222673 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-mhk7w" event={"ID":"db65cfec-9483-4641-8794-5fbd50c4bc92","Type":"ContainerStarted","Data":"8f24efe866cd65f48aabd2d091cc0ef9a51ee556ff32835c055cd282e1afca4c"} Mar 18 20:21:30 crc kubenswrapper[4950]: I0318 20:21:30.222741 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-mhk7w" event={"ID":"db65cfec-9483-4641-8794-5fbd50c4bc92","Type":"ContainerStarted","Data":"832706d1da4eeafda17ecba008ba22458fe51a3aa639b23e1f99d69d81ffd827"} Mar 18 20:21:30 crc kubenswrapper[4950]: I0318 20:21:30.222756 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-mhk7w" event={"ID":"db65cfec-9483-4641-8794-5fbd50c4bc92","Type":"ContainerStarted","Data":"5e3af07b8d34a66ab9df05554de62d2d76be8849e397eef687daebaf71406371"} Mar 18 20:21:30 crc kubenswrapper[4950]: I0318 20:21:30.222883 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-7bb4cc7c98-mhk7w" Mar 18 20:21:30 crc kubenswrapper[4950]: I0318 20:21:30.253120 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-7bb4cc7c98-mhk7w" podStartSLOduration=2.253100418 podStartE2EDuration="2.253100418s" podCreationTimestamp="2026-03-18 20:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:21:30.251129366 +0000 UTC m=+903.491971234" watchObservedRunningTime="2026-03-18 20:21:30.253100418 +0000 UTC m=+903.493942296" Mar 18 20:21:30 crc kubenswrapper[4950]: I0318 20:21:30.408339 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6f8ab895-b55a-43e6-a6fd-71f4ac42a50a-memberlist\") pod \"speaker-6k4zn\" (UID: \"6f8ab895-b55a-43e6-a6fd-71f4ac42a50a\") " pod="metallb-system/speaker-6k4zn" Mar 18 20:21:30 crc kubenswrapper[4950]: I0318 20:21:30.413890 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6f8ab895-b55a-43e6-a6fd-71f4ac42a50a-memberlist\") pod \"speaker-6k4zn\" (UID: \"6f8ab895-b55a-43e6-a6fd-71f4ac42a50a\") " pod="metallb-system/speaker-6k4zn" Mar 18 20:21:30 crc kubenswrapper[4950]: I0318 20:21:30.442866 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-6k4zn" Mar 18 20:21:30 crc kubenswrapper[4950]: W0318 20:21:30.462907 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f8ab895_b55a_43e6_a6fd_71f4ac42a50a.slice/crio-308bedd8557377902873647286ec4b0d37552053df572a31207221e2f9ebe6ee WatchSource:0}: Error finding container 308bedd8557377902873647286ec4b0d37552053df572a31207221e2f9ebe6ee: Status 404 returned error can't find the container with id 308bedd8557377902873647286ec4b0d37552053df572a31207221e2f9ebe6ee Mar 18 20:21:31 crc kubenswrapper[4950]: I0318 20:21:31.229543 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6k4zn" event={"ID":"6f8ab895-b55a-43e6-a6fd-71f4ac42a50a","Type":"ContainerStarted","Data":"70ed88dde156089b334a7f0fbf411f0d50b534b5564aad9e7a9bcb893846d8e5"} Mar 18 20:21:31 crc kubenswrapper[4950]: I0318 20:21:31.229892 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6k4zn" event={"ID":"6f8ab895-b55a-43e6-a6fd-71f4ac42a50a","Type":"ContainerStarted","Data":"2a437023f5240841e652e6e9010df23f979ca36b4a8603dba8c4c60b5b2ce679"} Mar 18 20:21:31 crc kubenswrapper[4950]: I0318 20:21:31.229903 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6k4zn" event={"ID":"6f8ab895-b55a-43e6-a6fd-71f4ac42a50a","Type":"ContainerStarted","Data":"308bedd8557377902873647286ec4b0d37552053df572a31207221e2f9ebe6ee"} Mar 18 20:21:31 crc kubenswrapper[4950]: I0318 20:21:31.230053 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-6k4zn" Mar 18 20:21:31 crc kubenswrapper[4950]: I0318 20:21:31.256684 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-6k4zn" podStartSLOduration=3.256668314 podStartE2EDuration="3.256668314s" podCreationTimestamp="2026-03-18 20:21:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:21:31.253348687 +0000 UTC m=+904.494190545" watchObservedRunningTime="2026-03-18 20:21:31.256668314 +0000 UTC m=+904.497510182" Mar 18 20:21:33 crc kubenswrapper[4950]: I0318 20:21:33.837463 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:21:33 crc kubenswrapper[4950]: I0318 20:21:33.837831 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:21:37 crc kubenswrapper[4950]: I0318 20:21:37.268178 4950 generic.go:334] "Generic (PLEG): container finished" podID="d221ef79-335c-46db-a491-b74bc07300c4" containerID="ff0ce1e7bfd33acf28b79f01c5b677fe21c806e2c825ab78058a9122a0c9a89f" exitCode=0 Mar 18 20:21:37 crc kubenswrapper[4950]: I0318 20:21:37.268266 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mlhr4" event={"ID":"d221ef79-335c-46db-a491-b74bc07300c4","Type":"ContainerDied","Data":"ff0ce1e7bfd33acf28b79f01c5b677fe21c806e2c825ab78058a9122a0c9a89f"} Mar 18 20:21:38 crc kubenswrapper[4950]: I0318 20:21:38.278793 4950 generic.go:334] "Generic (PLEG): container finished" podID="d221ef79-335c-46db-a491-b74bc07300c4" containerID="e0c3879275290dd0c918b2bdc959c9d1c32336d6fdbeb727f444acb203315db2" exitCode=0 Mar 18 20:21:38 crc kubenswrapper[4950]: I0318 20:21:38.278877 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mlhr4" event={"ID":"d221ef79-335c-46db-a491-b74bc07300c4","Type":"ContainerDied","Data":"e0c3879275290dd0c918b2bdc959c9d1c32336d6fdbeb727f444acb203315db2"} Mar 18 20:21:39 crc kubenswrapper[4950]: I0318 20:21:39.286896 4950 generic.go:334] "Generic (PLEG): container finished" podID="d221ef79-335c-46db-a491-b74bc07300c4" containerID="2eba4e77a633330cc920e5f3ba495367f8e64c433d959fa86e897791d2c65049" exitCode=0 Mar 18 20:21:39 crc kubenswrapper[4950]: I0318 20:21:39.286945 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mlhr4" event={"ID":"d221ef79-335c-46db-a491-b74bc07300c4","Type":"ContainerDied","Data":"2eba4e77a633330cc920e5f3ba495367f8e64c433d959fa86e897791d2c65049"} Mar 18 20:21:40 crc kubenswrapper[4950]: I0318 20:21:40.298199 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mlhr4" event={"ID":"d221ef79-335c-46db-a491-b74bc07300c4","Type":"ContainerStarted","Data":"950588dd332d9af6408a61f322114090743c33a595fba076c9b9c684257d7921"} Mar 18 20:21:40 crc kubenswrapper[4950]: I0318 20:21:40.298502 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mlhr4" event={"ID":"d221ef79-335c-46db-a491-b74bc07300c4","Type":"ContainerStarted","Data":"555e10ebde8abc694e5994533ecc5f4439e0b86df8d40d65ee8bc3cccd5d8bf9"} Mar 18 20:21:40 crc kubenswrapper[4950]: I0318 20:21:40.298521 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:40 crc kubenswrapper[4950]: I0318 20:21:40.298532 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mlhr4" event={"ID":"d221ef79-335c-46db-a491-b74bc07300c4","Type":"ContainerStarted","Data":"9f4817af07c37ff3049773a4966c3253510dda08521734d071811b48ff57e507"} Mar 18 20:21:40 crc kubenswrapper[4950]: I0318 20:21:40.298544 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mlhr4" event={"ID":"d221ef79-335c-46db-a491-b74bc07300c4","Type":"ContainerStarted","Data":"1730f7a1c6941e29115a3248dff0a35bb62c9b399bc16c82256d8b2f0aca49d5"} Mar 18 20:21:40 crc kubenswrapper[4950]: I0318 20:21:40.298556 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mlhr4" event={"ID":"d221ef79-335c-46db-a491-b74bc07300c4","Type":"ContainerStarted","Data":"41a7fcb3017fcfc19ef754624fb280bd6e3cd217d59f49eda1032dee29cdfb60"} Mar 18 20:21:40 crc kubenswrapper[4950]: I0318 20:21:40.298566 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mlhr4" event={"ID":"d221ef79-335c-46db-a491-b74bc07300c4","Type":"ContainerStarted","Data":"db7c003cc3f2116567a65b3cd864bb4f4f00101bd330d8b3deb4f0236987c4fb"} Mar 18 20:21:40 crc kubenswrapper[4950]: I0318 20:21:40.300334 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nqppx" event={"ID":"b4c0afa4-f131-44df-b2f1-9908b3b578e6","Type":"ContainerStarted","Data":"fb0c024a907d80a46d1d6d52fbbce65a8ab73056ece7ecd63c28273af19e6944"} Mar 18 20:21:40 crc kubenswrapper[4950]: I0318 20:21:40.300498 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nqppx" Mar 18 20:21:40 crc kubenswrapper[4950]: I0318 20:21:40.333618 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-mlhr4" podStartSLOduration=5.530453786 podStartE2EDuration="12.333600824s" podCreationTimestamp="2026-03-18 20:21:28 +0000 UTC" firstStartedPulling="2026-03-18 20:21:29.55674862 +0000 UTC m=+902.797590488" lastFinishedPulling="2026-03-18 20:21:36.359895658 +0000 UTC m=+909.600737526" observedRunningTime="2026-03-18 20:21:40.327789021 +0000 UTC m=+913.568630899" watchObservedRunningTime="2026-03-18 20:21:40.333600824 +0000 UTC m=+913.574442702" Mar 18 20:21:40 crc kubenswrapper[4950]: I0318 20:21:40.355605 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nqppx" podStartSLOduration=1.908290054 podStartE2EDuration="12.355582782s" podCreationTimestamp="2026-03-18 20:21:28 +0000 UTC" firstStartedPulling="2026-03-18 20:21:29.706466387 +0000 UTC m=+902.947308255" lastFinishedPulling="2026-03-18 20:21:40.153759105 +0000 UTC m=+913.394600983" observedRunningTime="2026-03-18 20:21:40.347319544 +0000 UTC m=+913.588161422" watchObservedRunningTime="2026-03-18 20:21:40.355582782 +0000 UTC m=+913.596424670" Mar 18 20:21:40 crc kubenswrapper[4950]: I0318 20:21:40.450341 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-6k4zn" Mar 18 20:21:43 crc kubenswrapper[4950]: I0318 20:21:43.081170 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-qxppm"] Mar 18 20:21:43 crc kubenswrapper[4950]: I0318 20:21:43.082199 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qxppm" Mar 18 20:21:43 crc kubenswrapper[4950]: I0318 20:21:43.084723 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-zj7ng" Mar 18 20:21:43 crc kubenswrapper[4950]: I0318 20:21:43.085136 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Mar 18 20:21:43 crc kubenswrapper[4950]: I0318 20:21:43.088115 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Mar 18 20:21:43 crc kubenswrapper[4950]: I0318 20:21:43.098353 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qxppm"] Mar 18 20:21:43 crc kubenswrapper[4950]: I0318 20:21:43.098972 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvwzn\" (UniqueName: \"kubernetes.io/projected/78c9b258-edbc-44cb-8bb6-c56169a5e413-kube-api-access-vvwzn\") pod \"openstack-operator-index-qxppm\" (UID: \"78c9b258-edbc-44cb-8bb6-c56169a5e413\") " pod="openstack-operators/openstack-operator-index-qxppm" Mar 18 20:21:43 crc kubenswrapper[4950]: I0318 20:21:43.200778 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvwzn\" (UniqueName: \"kubernetes.io/projected/78c9b258-edbc-44cb-8bb6-c56169a5e413-kube-api-access-vvwzn\") pod \"openstack-operator-index-qxppm\" (UID: \"78c9b258-edbc-44cb-8bb6-c56169a5e413\") " pod="openstack-operators/openstack-operator-index-qxppm" Mar 18 20:21:43 crc kubenswrapper[4950]: I0318 20:21:43.217986 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvwzn\" (UniqueName: \"kubernetes.io/projected/78c9b258-edbc-44cb-8bb6-c56169a5e413-kube-api-access-vvwzn\") pod \"openstack-operator-index-qxppm\" (UID: \"78c9b258-edbc-44cb-8bb6-c56169a5e413\") " pod="openstack-operators/openstack-operator-index-qxppm" Mar 18 20:21:43 crc kubenswrapper[4950]: I0318 20:21:43.404195 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qxppm" Mar 18 20:21:43 crc kubenswrapper[4950]: I0318 20:21:43.749519 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qxppm"] Mar 18 20:21:43 crc kubenswrapper[4950]: W0318 20:21:43.757851 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod78c9b258_edbc_44cb_8bb6_c56169a5e413.slice/crio-15459a6cbb2f0ca703eb1dce707153e99662b3be8ce5490387b44237d4766bb3 WatchSource:0}: Error finding container 15459a6cbb2f0ca703eb1dce707153e99662b3be8ce5490387b44237d4766bb3: Status 404 returned error can't find the container with id 15459a6cbb2f0ca703eb1dce707153e99662b3be8ce5490387b44237d4766bb3 Mar 18 20:21:44 crc kubenswrapper[4950]: I0318 20:21:44.330099 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qxppm" event={"ID":"78c9b258-edbc-44cb-8bb6-c56169a5e413","Type":"ContainerStarted","Data":"15459a6cbb2f0ca703eb1dce707153e99662b3be8ce5490387b44237d4766bb3"} Mar 18 20:21:44 crc kubenswrapper[4950]: I0318 20:21:44.428063 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:44 crc kubenswrapper[4950]: I0318 20:21:44.483901 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:45 crc kubenswrapper[4950]: I0318 20:21:45.341101 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qxppm" event={"ID":"78c9b258-edbc-44cb-8bb6-c56169a5e413","Type":"ContainerStarted","Data":"6f052ce9ac8157da9ec93908e4e7658b1956a348ce2b8b2786d9bc300909e0cc"} Mar 18 20:21:45 crc kubenswrapper[4950]: I0318 20:21:45.401202 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-qxppm" podStartSLOduration=1.6073231080000001 podStartE2EDuration="2.401175099s" podCreationTimestamp="2026-03-18 20:21:43 +0000 UTC" firstStartedPulling="2026-03-18 20:21:43.760910314 +0000 UTC m=+917.001752182" lastFinishedPulling="2026-03-18 20:21:44.554762285 +0000 UTC m=+917.795604173" observedRunningTime="2026-03-18 20:21:45.397295947 +0000 UTC m=+918.638137825" watchObservedRunningTime="2026-03-18 20:21:45.401175099 +0000 UTC m=+918.642017007" Mar 18 20:21:46 crc kubenswrapper[4950]: I0318 20:21:46.461018 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-qxppm"] Mar 18 20:21:47 crc kubenswrapper[4950]: I0318 20:21:47.079119 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-29kf5"] Mar 18 20:21:47 crc kubenswrapper[4950]: I0318 20:21:47.080275 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-29kf5" Mar 18 20:21:47 crc kubenswrapper[4950]: I0318 20:21:47.096813 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-29kf5"] Mar 18 20:21:47 crc kubenswrapper[4950]: I0318 20:21:47.168188 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x29v9\" (UniqueName: \"kubernetes.io/projected/e6006ce5-8d5e-40d2-b9cb-a1a7f501269f-kube-api-access-x29v9\") pod \"openstack-operator-index-29kf5\" (UID: \"e6006ce5-8d5e-40d2-b9cb-a1a7f501269f\") " pod="openstack-operators/openstack-operator-index-29kf5" Mar 18 20:21:47 crc kubenswrapper[4950]: I0318 20:21:47.269803 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x29v9\" (UniqueName: \"kubernetes.io/projected/e6006ce5-8d5e-40d2-b9cb-a1a7f501269f-kube-api-access-x29v9\") pod \"openstack-operator-index-29kf5\" (UID: \"e6006ce5-8d5e-40d2-b9cb-a1a7f501269f\") " pod="openstack-operators/openstack-operator-index-29kf5" Mar 18 20:21:47 crc kubenswrapper[4950]: I0318 20:21:47.302979 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x29v9\" (UniqueName: \"kubernetes.io/projected/e6006ce5-8d5e-40d2-b9cb-a1a7f501269f-kube-api-access-x29v9\") pod \"openstack-operator-index-29kf5\" (UID: \"e6006ce5-8d5e-40d2-b9cb-a1a7f501269f\") " pod="openstack-operators/openstack-operator-index-29kf5" Mar 18 20:21:47 crc kubenswrapper[4950]: I0318 20:21:47.360073 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-qxppm" podUID="78c9b258-edbc-44cb-8bb6-c56169a5e413" containerName="registry-server" containerID="cri-o://6f052ce9ac8157da9ec93908e4e7658b1956a348ce2b8b2786d9bc300909e0cc" gracePeriod=2 Mar 18 20:21:47 crc kubenswrapper[4950]: I0318 20:21:47.434185 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-29kf5" Mar 18 20:21:47 crc kubenswrapper[4950]: I0318 20:21:47.786369 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qxppm" Mar 18 20:21:47 crc kubenswrapper[4950]: I0318 20:21:47.876827 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvwzn\" (UniqueName: \"kubernetes.io/projected/78c9b258-edbc-44cb-8bb6-c56169a5e413-kube-api-access-vvwzn\") pod \"78c9b258-edbc-44cb-8bb6-c56169a5e413\" (UID: \"78c9b258-edbc-44cb-8bb6-c56169a5e413\") " Mar 18 20:21:47 crc kubenswrapper[4950]: I0318 20:21:47.879692 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78c9b258-edbc-44cb-8bb6-c56169a5e413-kube-api-access-vvwzn" (OuterVolumeSpecName: "kube-api-access-vvwzn") pod "78c9b258-edbc-44cb-8bb6-c56169a5e413" (UID: "78c9b258-edbc-44cb-8bb6-c56169a5e413"). InnerVolumeSpecName "kube-api-access-vvwzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:21:47 crc kubenswrapper[4950]: I0318 20:21:47.880555 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-29kf5"] Mar 18 20:21:47 crc kubenswrapper[4950]: W0318 20:21:47.883541 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6006ce5_8d5e_40d2_b9cb_a1a7f501269f.slice/crio-121c191e6bd1b92f240933253be9295baccfb0a8ee1e0755de94a992d7ee3323 WatchSource:0}: Error finding container 121c191e6bd1b92f240933253be9295baccfb0a8ee1e0755de94a992d7ee3323: Status 404 returned error can't find the container with id 121c191e6bd1b92f240933253be9295baccfb0a8ee1e0755de94a992d7ee3323 Mar 18 20:21:47 crc kubenswrapper[4950]: I0318 20:21:47.979989 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvwzn\" (UniqueName: \"kubernetes.io/projected/78c9b258-edbc-44cb-8bb6-c56169a5e413-kube-api-access-vvwzn\") on node \"crc\" DevicePath \"\"" Mar 18 20:21:48 crc kubenswrapper[4950]: I0318 20:21:48.376695 4950 generic.go:334] "Generic (PLEG): container finished" podID="78c9b258-edbc-44cb-8bb6-c56169a5e413" containerID="6f052ce9ac8157da9ec93908e4e7658b1956a348ce2b8b2786d9bc300909e0cc" exitCode=0 Mar 18 20:21:48 crc kubenswrapper[4950]: I0318 20:21:48.377016 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qxppm" event={"ID":"78c9b258-edbc-44cb-8bb6-c56169a5e413","Type":"ContainerDied","Data":"6f052ce9ac8157da9ec93908e4e7658b1956a348ce2b8b2786d9bc300909e0cc"} Mar 18 20:21:48 crc kubenswrapper[4950]: I0318 20:21:48.377082 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qxppm" event={"ID":"78c9b258-edbc-44cb-8bb6-c56169a5e413","Type":"ContainerDied","Data":"15459a6cbb2f0ca703eb1dce707153e99662b3be8ce5490387b44237d4766bb3"} Mar 18 20:21:48 crc kubenswrapper[4950]: I0318 20:21:48.377082 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qxppm" Mar 18 20:21:48 crc kubenswrapper[4950]: I0318 20:21:48.377114 4950 scope.go:117] "RemoveContainer" containerID="6f052ce9ac8157da9ec93908e4e7658b1956a348ce2b8b2786d9bc300909e0cc" Mar 18 20:21:48 crc kubenswrapper[4950]: I0318 20:21:48.383579 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-29kf5" event={"ID":"e6006ce5-8d5e-40d2-b9cb-a1a7f501269f","Type":"ContainerStarted","Data":"121c191e6bd1b92f240933253be9295baccfb0a8ee1e0755de94a992d7ee3323"} Mar 18 20:21:48 crc kubenswrapper[4950]: I0318 20:21:48.423709 4950 scope.go:117] "RemoveContainer" containerID="6f052ce9ac8157da9ec93908e4e7658b1956a348ce2b8b2786d9bc300909e0cc" Mar 18 20:21:48 crc kubenswrapper[4950]: E0318 20:21:48.424776 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f052ce9ac8157da9ec93908e4e7658b1956a348ce2b8b2786d9bc300909e0cc\": container with ID starting with 6f052ce9ac8157da9ec93908e4e7658b1956a348ce2b8b2786d9bc300909e0cc not found: ID does not exist" containerID="6f052ce9ac8157da9ec93908e4e7658b1956a348ce2b8b2786d9bc300909e0cc" Mar 18 20:21:48 crc kubenswrapper[4950]: I0318 20:21:48.430696 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f052ce9ac8157da9ec93908e4e7658b1956a348ce2b8b2786d9bc300909e0cc"} err="failed to get container status \"6f052ce9ac8157da9ec93908e4e7658b1956a348ce2b8b2786d9bc300909e0cc\": rpc error: code = NotFound desc = could not find container \"6f052ce9ac8157da9ec93908e4e7658b1956a348ce2b8b2786d9bc300909e0cc\": container with ID starting with 6f052ce9ac8157da9ec93908e4e7658b1956a348ce2b8b2786d9bc300909e0cc not found: ID does not exist" Mar 18 20:21:48 crc kubenswrapper[4950]: I0318 20:21:48.443387 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-qxppm"] Mar 18 20:21:48 crc kubenswrapper[4950]: I0318 20:21:48.450912 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-qxppm"] Mar 18 20:21:48 crc kubenswrapper[4950]: I0318 20:21:48.965592 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-7bb4cc7c98-mhk7w" Mar 18 20:21:49 crc kubenswrapper[4950]: I0318 20:21:49.394280 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-29kf5" event={"ID":"e6006ce5-8d5e-40d2-b9cb-a1a7f501269f","Type":"ContainerStarted","Data":"b03711376bdaad226edd19db5726b3db1498c19a8313a4a051d5c486854d58bc"} Mar 18 20:21:49 crc kubenswrapper[4950]: I0318 20:21:49.412534 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-29kf5" podStartSLOduration=1.838222297 podStartE2EDuration="2.412515075s" podCreationTimestamp="2026-03-18 20:21:47 +0000 UTC" firstStartedPulling="2026-03-18 20:21:47.887930842 +0000 UTC m=+921.128772710" lastFinishedPulling="2026-03-18 20:21:48.46222361 +0000 UTC m=+921.703065488" observedRunningTime="2026-03-18 20:21:49.40967743 +0000 UTC m=+922.650519308" watchObservedRunningTime="2026-03-18 20:21:49.412515075 +0000 UTC m=+922.653356953" Mar 18 20:21:49 crc kubenswrapper[4950]: I0318 20:21:49.431612 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-mlhr4" Mar 18 20:21:49 crc kubenswrapper[4950]: I0318 20:21:49.488508 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78c9b258-edbc-44cb-8bb6-c56169a5e413" path="/var/lib/kubelet/pods/78c9b258-edbc-44cb-8bb6-c56169a5e413/volumes" Mar 18 20:21:57 crc kubenswrapper[4950]: I0318 20:21:57.434357 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-29kf5" Mar 18 20:21:57 crc kubenswrapper[4950]: I0318 20:21:57.434875 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-29kf5" Mar 18 20:21:57 crc kubenswrapper[4950]: I0318 20:21:57.476877 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-29kf5" Mar 18 20:21:57 crc kubenswrapper[4950]: I0318 20:21:57.518106 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-29kf5" Mar 18 20:21:59 crc kubenswrapper[4950]: I0318 20:21:59.129343 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm"] Mar 18 20:21:59 crc kubenswrapper[4950]: E0318 20:21:59.130147 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78c9b258-edbc-44cb-8bb6-c56169a5e413" containerName="registry-server" Mar 18 20:21:59 crc kubenswrapper[4950]: I0318 20:21:59.130168 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="78c9b258-edbc-44cb-8bb6-c56169a5e413" containerName="registry-server" Mar 18 20:21:59 crc kubenswrapper[4950]: I0318 20:21:59.130371 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="78c9b258-edbc-44cb-8bb6-c56169a5e413" containerName="registry-server" Mar 18 20:21:59 crc kubenswrapper[4950]: I0318 20:21:59.131832 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm" Mar 18 20:21:59 crc kubenswrapper[4950]: I0318 20:21:59.133784 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-cdt6d" Mar 18 20:21:59 crc kubenswrapper[4950]: I0318 20:21:59.150696 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d-util\") pod \"7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm\" (UID: \"0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d\") " pod="openstack-operators/7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm" Mar 18 20:21:59 crc kubenswrapper[4950]: I0318 20:21:59.150781 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d-bundle\") pod \"7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm\" (UID: \"0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d\") " pod="openstack-operators/7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm" Mar 18 20:21:59 crc kubenswrapper[4950]: I0318 20:21:59.150847 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr9p4\" (UniqueName: \"kubernetes.io/projected/0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d-kube-api-access-hr9p4\") pod \"7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm\" (UID: \"0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d\") " pod="openstack-operators/7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm" Mar 18 20:21:59 crc kubenswrapper[4950]: I0318 20:21:59.170764 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm"] Mar 18 20:21:59 crc kubenswrapper[4950]: I0318 20:21:59.251602 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d-util\") pod \"7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm\" (UID: \"0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d\") " pod="openstack-operators/7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm" Mar 18 20:21:59 crc kubenswrapper[4950]: I0318 20:21:59.251688 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d-bundle\") pod \"7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm\" (UID: \"0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d\") " pod="openstack-operators/7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm" Mar 18 20:21:59 crc kubenswrapper[4950]: I0318 20:21:59.251744 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr9p4\" (UniqueName: \"kubernetes.io/projected/0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d-kube-api-access-hr9p4\") pod \"7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm\" (UID: \"0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d\") " pod="openstack-operators/7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm" Mar 18 20:21:59 crc kubenswrapper[4950]: I0318 20:21:59.252212 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d-util\") pod \"7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm\" (UID: \"0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d\") " pod="openstack-operators/7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm" Mar 18 20:21:59 crc kubenswrapper[4950]: I0318 20:21:59.252786 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d-bundle\") pod \"7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm\" (UID: \"0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d\") " pod="openstack-operators/7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm" Mar 18 20:21:59 crc kubenswrapper[4950]: I0318 20:21:59.286562 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr9p4\" (UniqueName: \"kubernetes.io/projected/0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d-kube-api-access-hr9p4\") pod \"7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm\" (UID: \"0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d\") " pod="openstack-operators/7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm" Mar 18 20:21:59 crc kubenswrapper[4950]: I0318 20:21:59.454476 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm" Mar 18 20:21:59 crc kubenswrapper[4950]: I0318 20:21:59.472530 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-nqppx" Mar 18 20:21:59 crc kubenswrapper[4950]: I0318 20:21:59.675999 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm"] Mar 18 20:21:59 crc kubenswrapper[4950]: W0318 20:21:59.691563 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b6bceb7_5b53_4098_9b3b_d0c4ffa97a7d.slice/crio-02377a65e78143e835ec79366a39955377ef693f479f0ea7e01412a2c7d1ab67 WatchSource:0}: Error finding container 02377a65e78143e835ec79366a39955377ef693f479f0ea7e01412a2c7d1ab67: Status 404 returned error can't find the container with id 02377a65e78143e835ec79366a39955377ef693f479f0ea7e01412a2c7d1ab67 Mar 18 20:22:00 crc kubenswrapper[4950]: I0318 20:22:00.139768 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564422-glkgm"] Mar 18 20:22:00 crc kubenswrapper[4950]: I0318 20:22:00.140567 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564422-glkgm" Mar 18 20:22:00 crc kubenswrapper[4950]: I0318 20:22:00.142815 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:22:00 crc kubenswrapper[4950]: I0318 20:22:00.143345 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:22:00 crc kubenswrapper[4950]: I0318 20:22:00.147221 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:22:00 crc kubenswrapper[4950]: I0318 20:22:00.151746 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564422-glkgm"] Mar 18 20:22:00 crc kubenswrapper[4950]: I0318 20:22:00.264018 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4q4hw\" (UniqueName: \"kubernetes.io/projected/b60e86f9-3271-4263-bcaa-c4629cd1346d-kube-api-access-4q4hw\") pod \"auto-csr-approver-29564422-glkgm\" (UID: \"b60e86f9-3271-4263-bcaa-c4629cd1346d\") " pod="openshift-infra/auto-csr-approver-29564422-glkgm" Mar 18 20:22:00 crc kubenswrapper[4950]: I0318 20:22:00.365077 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4q4hw\" (UniqueName: \"kubernetes.io/projected/b60e86f9-3271-4263-bcaa-c4629cd1346d-kube-api-access-4q4hw\") pod \"auto-csr-approver-29564422-glkgm\" (UID: \"b60e86f9-3271-4263-bcaa-c4629cd1346d\") " pod="openshift-infra/auto-csr-approver-29564422-glkgm" Mar 18 20:22:00 crc kubenswrapper[4950]: I0318 20:22:00.395194 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4q4hw\" (UniqueName: \"kubernetes.io/projected/b60e86f9-3271-4263-bcaa-c4629cd1346d-kube-api-access-4q4hw\") pod \"auto-csr-approver-29564422-glkgm\" (UID: \"b60e86f9-3271-4263-bcaa-c4629cd1346d\") " pod="openshift-infra/auto-csr-approver-29564422-glkgm" Mar 18 20:22:00 crc kubenswrapper[4950]: I0318 20:22:00.459784 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564422-glkgm" Mar 18 20:22:00 crc kubenswrapper[4950]: I0318 20:22:00.482126 4950 generic.go:334] "Generic (PLEG): container finished" podID="0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d" containerID="63e92146793ee885100ea86792ee0655e412fba3f44cd81cba76e777a382a5e7" exitCode=0 Mar 18 20:22:00 crc kubenswrapper[4950]: I0318 20:22:00.482172 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm" event={"ID":"0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d","Type":"ContainerDied","Data":"63e92146793ee885100ea86792ee0655e412fba3f44cd81cba76e777a382a5e7"} Mar 18 20:22:00 crc kubenswrapper[4950]: I0318 20:22:00.482201 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm" event={"ID":"0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d","Type":"ContainerStarted","Data":"02377a65e78143e835ec79366a39955377ef693f479f0ea7e01412a2c7d1ab67"} Mar 18 20:22:00 crc kubenswrapper[4950]: I0318 20:22:00.909219 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564422-glkgm"] Mar 18 20:22:01 crc kubenswrapper[4950]: I0318 20:22:01.488307 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564422-glkgm" event={"ID":"b60e86f9-3271-4263-bcaa-c4629cd1346d","Type":"ContainerStarted","Data":"79ffc6076fdcb095448aac42c9037cfefc8bf99cd97602e1d9c08d2ecb6044a7"} Mar 18 20:22:01 crc kubenswrapper[4950]: I0318 20:22:01.489598 4950 generic.go:334] "Generic (PLEG): container finished" podID="0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d" containerID="0fda5a79a464ecfe74b4083d8a51ee133847ace751d39fb2ced4861d00f79291" exitCode=0 Mar 18 20:22:01 crc kubenswrapper[4950]: I0318 20:22:01.489636 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm" event={"ID":"0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d","Type":"ContainerDied","Data":"0fda5a79a464ecfe74b4083d8a51ee133847ace751d39fb2ced4861d00f79291"} Mar 18 20:22:02 crc kubenswrapper[4950]: I0318 20:22:02.496266 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564422-glkgm" event={"ID":"b60e86f9-3271-4263-bcaa-c4629cd1346d","Type":"ContainerStarted","Data":"7244a7cd7d6e9bf822acc6b28539c9730114bd6610cbc203b10c7c13f309f94a"} Mar 18 20:22:02 crc kubenswrapper[4950]: I0318 20:22:02.498742 4950 generic.go:334] "Generic (PLEG): container finished" podID="0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d" containerID="ec2d609ae97e796107fada2e8444e3385de10b904b2451f2602a9d1113ffde06" exitCode=0 Mar 18 20:22:02 crc kubenswrapper[4950]: I0318 20:22:02.498780 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm" event={"ID":"0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d","Type":"ContainerDied","Data":"ec2d609ae97e796107fada2e8444e3385de10b904b2451f2602a9d1113ffde06"} Mar 18 20:22:02 crc kubenswrapper[4950]: I0318 20:22:02.514816 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29564422-glkgm" podStartSLOduration=1.415155897 podStartE2EDuration="2.514800819s" podCreationTimestamp="2026-03-18 20:22:00 +0000 UTC" firstStartedPulling="2026-03-18 20:22:00.918978152 +0000 UTC m=+934.159820020" lastFinishedPulling="2026-03-18 20:22:02.018623034 +0000 UTC m=+935.259464942" observedRunningTime="2026-03-18 20:22:02.510770923 +0000 UTC m=+935.751612791" watchObservedRunningTime="2026-03-18 20:22:02.514800819 +0000 UTC m=+935.755642687" Mar 18 20:22:03 crc kubenswrapper[4950]: I0318 20:22:03.517236 4950 generic.go:334] "Generic (PLEG): container finished" podID="b60e86f9-3271-4263-bcaa-c4629cd1346d" containerID="7244a7cd7d6e9bf822acc6b28539c9730114bd6610cbc203b10c7c13f309f94a" exitCode=0 Mar 18 20:22:03 crc kubenswrapper[4950]: I0318 20:22:03.518353 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564422-glkgm" event={"ID":"b60e86f9-3271-4263-bcaa-c4629cd1346d","Type":"ContainerDied","Data":"7244a7cd7d6e9bf822acc6b28539c9730114bd6610cbc203b10c7c13f309f94a"} Mar 18 20:22:03 crc kubenswrapper[4950]: I0318 20:22:03.835847 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:22:03 crc kubenswrapper[4950]: I0318 20:22:03.835910 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:22:03 crc kubenswrapper[4950]: I0318 20:22:03.835956 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:22:03 crc kubenswrapper[4950]: I0318 20:22:03.836559 4950 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bef1deca2fab314ca277f2c06d049ce203ead40181a917f46a9164a0f3e058c3"} pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 20:22:03 crc kubenswrapper[4950]: I0318 20:22:03.836615 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" containerID="cri-o://bef1deca2fab314ca277f2c06d049ce203ead40181a917f46a9164a0f3e058c3" gracePeriod=600 Mar 18 20:22:03 crc kubenswrapper[4950]: I0318 20:22:03.840513 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm" Mar 18 20:22:03 crc kubenswrapper[4950]: I0318 20:22:03.918924 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d-util\") pod \"0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d\" (UID: \"0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d\") " Mar 18 20:22:03 crc kubenswrapper[4950]: I0318 20:22:03.918983 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d-bundle\") pod \"0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d\" (UID: \"0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d\") " Mar 18 20:22:03 crc kubenswrapper[4950]: I0318 20:22:03.919051 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hr9p4\" (UniqueName: \"kubernetes.io/projected/0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d-kube-api-access-hr9p4\") pod \"0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d\" (UID: \"0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d\") " Mar 18 20:22:03 crc kubenswrapper[4950]: I0318 20:22:03.921385 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d-bundle" (OuterVolumeSpecName: "bundle") pod "0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d" (UID: "0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:22:03 crc kubenswrapper[4950]: I0318 20:22:03.942906 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d-util" (OuterVolumeSpecName: "util") pod "0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d" (UID: "0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:22:04 crc kubenswrapper[4950]: I0318 20:22:04.020494 4950 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d-util\") on node \"crc\" DevicePath \"\"" Mar 18 20:22:04 crc kubenswrapper[4950]: I0318 20:22:04.020528 4950 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:22:04 crc kubenswrapper[4950]: I0318 20:22:04.072691 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d-kube-api-access-hr9p4" (OuterVolumeSpecName: "kube-api-access-hr9p4") pod "0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d" (UID: "0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d"). InnerVolumeSpecName "kube-api-access-hr9p4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:22:04 crc kubenswrapper[4950]: I0318 20:22:04.122036 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hr9p4\" (UniqueName: \"kubernetes.io/projected/0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d-kube-api-access-hr9p4\") on node \"crc\" DevicePath \"\"" Mar 18 20:22:04 crc kubenswrapper[4950]: I0318 20:22:04.529265 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm" event={"ID":"0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d","Type":"ContainerDied","Data":"02377a65e78143e835ec79366a39955377ef693f479f0ea7e01412a2c7d1ab67"} Mar 18 20:22:04 crc kubenswrapper[4950]: I0318 20:22:04.529639 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02377a65e78143e835ec79366a39955377ef693f479f0ea7e01412a2c7d1ab67" Mar 18 20:22:04 crc kubenswrapper[4950]: I0318 20:22:04.529298 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm" Mar 18 20:22:04 crc kubenswrapper[4950]: I0318 20:22:04.542335 4950 generic.go:334] "Generic (PLEG): container finished" podID="4048b439-3266-46e7-9de0-22377efacc46" containerID="bef1deca2fab314ca277f2c06d049ce203ead40181a917f46a9164a0f3e058c3" exitCode=0 Mar 18 20:22:04 crc kubenswrapper[4950]: I0318 20:22:04.542453 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerDied","Data":"bef1deca2fab314ca277f2c06d049ce203ead40181a917f46a9164a0f3e058c3"} Mar 18 20:22:04 crc kubenswrapper[4950]: I0318 20:22:04.542522 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerStarted","Data":"80bb5b9fd4b42a17689aca26b1eb183f4363b2ddacc1fdf73b829648cfa1e873"} Mar 18 20:22:04 crc kubenswrapper[4950]: I0318 20:22:04.542554 4950 scope.go:117] "RemoveContainer" containerID="7eafe5dcfcc22883e166beb04dbf1c20eecc93c473bcbfb2c29f6e0319e288f2" Mar 18 20:22:04 crc kubenswrapper[4950]: I0318 20:22:04.851071 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564422-glkgm" Mar 18 20:22:05 crc kubenswrapper[4950]: I0318 20:22:05.031510 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4q4hw\" (UniqueName: \"kubernetes.io/projected/b60e86f9-3271-4263-bcaa-c4629cd1346d-kube-api-access-4q4hw\") pod \"b60e86f9-3271-4263-bcaa-c4629cd1346d\" (UID: \"b60e86f9-3271-4263-bcaa-c4629cd1346d\") " Mar 18 20:22:05 crc kubenswrapper[4950]: I0318 20:22:05.037569 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b60e86f9-3271-4263-bcaa-c4629cd1346d-kube-api-access-4q4hw" (OuterVolumeSpecName: "kube-api-access-4q4hw") pod "b60e86f9-3271-4263-bcaa-c4629cd1346d" (UID: "b60e86f9-3271-4263-bcaa-c4629cd1346d"). InnerVolumeSpecName "kube-api-access-4q4hw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:22:05 crc kubenswrapper[4950]: I0318 20:22:05.133796 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4q4hw\" (UniqueName: \"kubernetes.io/projected/b60e86f9-3271-4263-bcaa-c4629cd1346d-kube-api-access-4q4hw\") on node \"crc\" DevicePath \"\"" Mar 18 20:22:05 crc kubenswrapper[4950]: I0318 20:22:05.555175 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564422-glkgm" Mar 18 20:22:05 crc kubenswrapper[4950]: I0318 20:22:05.555084 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564422-glkgm" event={"ID":"b60e86f9-3271-4263-bcaa-c4629cd1346d","Type":"ContainerDied","Data":"79ffc6076fdcb095448aac42c9037cfefc8bf99cd97602e1d9c08d2ecb6044a7"} Mar 18 20:22:05 crc kubenswrapper[4950]: I0318 20:22:05.556046 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79ffc6076fdcb095448aac42c9037cfefc8bf99cd97602e1d9c08d2ecb6044a7" Mar 18 20:22:05 crc kubenswrapper[4950]: I0318 20:22:05.578680 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564416-mcvh6"] Mar 18 20:22:05 crc kubenswrapper[4950]: I0318 20:22:05.586967 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564416-mcvh6"] Mar 18 20:22:06 crc kubenswrapper[4950]: I0318 20:22:06.834534 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ctmcv"] Mar 18 20:22:06 crc kubenswrapper[4950]: E0318 20:22:06.835282 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d" containerName="util" Mar 18 20:22:06 crc kubenswrapper[4950]: I0318 20:22:06.835300 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d" containerName="util" Mar 18 20:22:06 crc kubenswrapper[4950]: E0318 20:22:06.835322 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d" containerName="pull" Mar 18 20:22:06 crc kubenswrapper[4950]: I0318 20:22:06.835330 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d" containerName="pull" Mar 18 20:22:06 crc kubenswrapper[4950]: E0318 20:22:06.835371 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b60e86f9-3271-4263-bcaa-c4629cd1346d" containerName="oc" Mar 18 20:22:06 crc kubenswrapper[4950]: I0318 20:22:06.835381 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="b60e86f9-3271-4263-bcaa-c4629cd1346d" containerName="oc" Mar 18 20:22:06 crc kubenswrapper[4950]: E0318 20:22:06.835390 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d" containerName="extract" Mar 18 20:22:06 crc kubenswrapper[4950]: I0318 20:22:06.835397 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d" containerName="extract" Mar 18 20:22:06 crc kubenswrapper[4950]: I0318 20:22:06.835554 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d" containerName="extract" Mar 18 20:22:06 crc kubenswrapper[4950]: I0318 20:22:06.835571 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="b60e86f9-3271-4263-bcaa-c4629cd1346d" containerName="oc" Mar 18 20:22:06 crc kubenswrapper[4950]: I0318 20:22:06.836602 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ctmcv" Mar 18 20:22:06 crc kubenswrapper[4950]: I0318 20:22:06.850070 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ctmcv"] Mar 18 20:22:06 crc kubenswrapper[4950]: I0318 20:22:06.956328 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10835f63-f9b3-4dba-80a4-1af5ca39fb99-catalog-content\") pod \"redhat-marketplace-ctmcv\" (UID: \"10835f63-f9b3-4dba-80a4-1af5ca39fb99\") " pod="openshift-marketplace/redhat-marketplace-ctmcv" Mar 18 20:22:06 crc kubenswrapper[4950]: I0318 20:22:06.956445 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-css4l\" (UniqueName: \"kubernetes.io/projected/10835f63-f9b3-4dba-80a4-1af5ca39fb99-kube-api-access-css4l\") pod \"redhat-marketplace-ctmcv\" (UID: \"10835f63-f9b3-4dba-80a4-1af5ca39fb99\") " pod="openshift-marketplace/redhat-marketplace-ctmcv" Mar 18 20:22:06 crc kubenswrapper[4950]: I0318 20:22:06.956503 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10835f63-f9b3-4dba-80a4-1af5ca39fb99-utilities\") pod \"redhat-marketplace-ctmcv\" (UID: \"10835f63-f9b3-4dba-80a4-1af5ca39fb99\") " pod="openshift-marketplace/redhat-marketplace-ctmcv" Mar 18 20:22:07 crc kubenswrapper[4950]: I0318 20:22:07.057028 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-css4l\" (UniqueName: \"kubernetes.io/projected/10835f63-f9b3-4dba-80a4-1af5ca39fb99-kube-api-access-css4l\") pod \"redhat-marketplace-ctmcv\" (UID: \"10835f63-f9b3-4dba-80a4-1af5ca39fb99\") " pod="openshift-marketplace/redhat-marketplace-ctmcv" Mar 18 20:22:07 crc kubenswrapper[4950]: I0318 20:22:07.057082 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10835f63-f9b3-4dba-80a4-1af5ca39fb99-utilities\") pod \"redhat-marketplace-ctmcv\" (UID: \"10835f63-f9b3-4dba-80a4-1af5ca39fb99\") " pod="openshift-marketplace/redhat-marketplace-ctmcv" Mar 18 20:22:07 crc kubenswrapper[4950]: I0318 20:22:07.057128 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10835f63-f9b3-4dba-80a4-1af5ca39fb99-catalog-content\") pod \"redhat-marketplace-ctmcv\" (UID: \"10835f63-f9b3-4dba-80a4-1af5ca39fb99\") " pod="openshift-marketplace/redhat-marketplace-ctmcv" Mar 18 20:22:07 crc kubenswrapper[4950]: I0318 20:22:07.057642 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10835f63-f9b3-4dba-80a4-1af5ca39fb99-catalog-content\") pod \"redhat-marketplace-ctmcv\" (UID: \"10835f63-f9b3-4dba-80a4-1af5ca39fb99\") " pod="openshift-marketplace/redhat-marketplace-ctmcv" Mar 18 20:22:07 crc kubenswrapper[4950]: I0318 20:22:07.057661 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10835f63-f9b3-4dba-80a4-1af5ca39fb99-utilities\") pod \"redhat-marketplace-ctmcv\" (UID: \"10835f63-f9b3-4dba-80a4-1af5ca39fb99\") " pod="openshift-marketplace/redhat-marketplace-ctmcv" Mar 18 20:22:07 crc kubenswrapper[4950]: I0318 20:22:07.076398 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-css4l\" (UniqueName: \"kubernetes.io/projected/10835f63-f9b3-4dba-80a4-1af5ca39fb99-kube-api-access-css4l\") pod \"redhat-marketplace-ctmcv\" (UID: \"10835f63-f9b3-4dba-80a4-1af5ca39fb99\") " pod="openshift-marketplace/redhat-marketplace-ctmcv" Mar 18 20:22:07 crc kubenswrapper[4950]: I0318 20:22:07.153282 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ctmcv" Mar 18 20:22:07 crc kubenswrapper[4950]: I0318 20:22:07.490166 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13558485-efd4-47b9-a7ad-2a82ed08fe3b" path="/var/lib/kubelet/pods/13558485-efd4-47b9-a7ad-2a82ed08fe3b/volumes" Mar 18 20:22:07 crc kubenswrapper[4950]: I0318 20:22:07.635582 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ctmcv"] Mar 18 20:22:07 crc kubenswrapper[4950]: W0318 20:22:07.639998 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10835f63_f9b3_4dba_80a4_1af5ca39fb99.slice/crio-02109309e034dcc0313440ee44ef187b3501e9bdbdbc09f980e1d0e530b67612 WatchSource:0}: Error finding container 02109309e034dcc0313440ee44ef187b3501e9bdbdbc09f980e1d0e530b67612: Status 404 returned error can't find the container with id 02109309e034dcc0313440ee44ef187b3501e9bdbdbc09f980e1d0e530b67612 Mar 18 20:22:08 crc kubenswrapper[4950]: I0318 20:22:08.574676 4950 generic.go:334] "Generic (PLEG): container finished" podID="10835f63-f9b3-4dba-80a4-1af5ca39fb99" containerID="496c67be195eafd70e76f49f971c898c6cec7ce5c39c3df3adfee15685d61ac9" exitCode=0 Mar 18 20:22:08 crc kubenswrapper[4950]: I0318 20:22:08.574903 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ctmcv" event={"ID":"10835f63-f9b3-4dba-80a4-1af5ca39fb99","Type":"ContainerDied","Data":"496c67be195eafd70e76f49f971c898c6cec7ce5c39c3df3adfee15685d61ac9"} Mar 18 20:22:08 crc kubenswrapper[4950]: I0318 20:22:08.575272 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ctmcv" event={"ID":"10835f63-f9b3-4dba-80a4-1af5ca39fb99","Type":"ContainerStarted","Data":"02109309e034dcc0313440ee44ef187b3501e9bdbdbc09f980e1d0e530b67612"} Mar 18 20:22:09 crc kubenswrapper[4950]: I0318 20:22:09.581793 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ctmcv" event={"ID":"10835f63-f9b3-4dba-80a4-1af5ca39fb99","Type":"ContainerStarted","Data":"987d393dafd6ddcade06d315ec14edaa8519ad1e94b51d56111f84ed031a3853"} Mar 18 20:22:10 crc kubenswrapper[4950]: I0318 20:22:10.392014 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-b85c4d696-w5ws4"] Mar 18 20:22:10 crc kubenswrapper[4950]: I0318 20:22:10.405560 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-b85c4d696-w5ws4" Mar 18 20:22:10 crc kubenswrapper[4950]: I0318 20:22:10.408068 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-mkgvm" Mar 18 20:22:10 crc kubenswrapper[4950]: I0318 20:22:10.413992 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-b85c4d696-w5ws4"] Mar 18 20:22:10 crc kubenswrapper[4950]: I0318 20:22:10.549730 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24nkj\" (UniqueName: \"kubernetes.io/projected/040697e2-a9b5-4ec0-be8d-f935227c4a0a-kube-api-access-24nkj\") pod \"openstack-operator-controller-init-b85c4d696-w5ws4\" (UID: \"040697e2-a9b5-4ec0-be8d-f935227c4a0a\") " pod="openstack-operators/openstack-operator-controller-init-b85c4d696-w5ws4" Mar 18 20:22:10 crc kubenswrapper[4950]: I0318 20:22:10.591178 4950 generic.go:334] "Generic (PLEG): container finished" podID="10835f63-f9b3-4dba-80a4-1af5ca39fb99" containerID="987d393dafd6ddcade06d315ec14edaa8519ad1e94b51d56111f84ed031a3853" exitCode=0 Mar 18 20:22:10 crc kubenswrapper[4950]: I0318 20:22:10.593394 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ctmcv" event={"ID":"10835f63-f9b3-4dba-80a4-1af5ca39fb99","Type":"ContainerDied","Data":"987d393dafd6ddcade06d315ec14edaa8519ad1e94b51d56111f84ed031a3853"} Mar 18 20:22:10 crc kubenswrapper[4950]: I0318 20:22:10.651202 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24nkj\" (UniqueName: \"kubernetes.io/projected/040697e2-a9b5-4ec0-be8d-f935227c4a0a-kube-api-access-24nkj\") pod \"openstack-operator-controller-init-b85c4d696-w5ws4\" (UID: \"040697e2-a9b5-4ec0-be8d-f935227c4a0a\") " pod="openstack-operators/openstack-operator-controller-init-b85c4d696-w5ws4" Mar 18 20:22:10 crc kubenswrapper[4950]: I0318 20:22:10.670196 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24nkj\" (UniqueName: \"kubernetes.io/projected/040697e2-a9b5-4ec0-be8d-f935227c4a0a-kube-api-access-24nkj\") pod \"openstack-operator-controller-init-b85c4d696-w5ws4\" (UID: \"040697e2-a9b5-4ec0-be8d-f935227c4a0a\") " pod="openstack-operators/openstack-operator-controller-init-b85c4d696-w5ws4" Mar 18 20:22:10 crc kubenswrapper[4950]: I0318 20:22:10.727265 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-b85c4d696-w5ws4" Mar 18 20:22:11 crc kubenswrapper[4950]: W0318 20:22:11.211920 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod040697e2_a9b5_4ec0_be8d_f935227c4a0a.slice/crio-2cbd0a1d0450074b16edfc689fcd63c31e715d3ce01bf0540ff03191a4cbbeba WatchSource:0}: Error finding container 2cbd0a1d0450074b16edfc689fcd63c31e715d3ce01bf0540ff03191a4cbbeba: Status 404 returned error can't find the container with id 2cbd0a1d0450074b16edfc689fcd63c31e715d3ce01bf0540ff03191a4cbbeba Mar 18 20:22:11 crc kubenswrapper[4950]: I0318 20:22:11.220789 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-b85c4d696-w5ws4"] Mar 18 20:22:11 crc kubenswrapper[4950]: I0318 20:22:11.602217 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ctmcv" event={"ID":"10835f63-f9b3-4dba-80a4-1af5ca39fb99","Type":"ContainerStarted","Data":"a595aa3392b5fab8dce5e99fd528a8f1583c8da65adf046b1194f581d401db3d"} Mar 18 20:22:11 crc kubenswrapper[4950]: I0318 20:22:11.603932 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-b85c4d696-w5ws4" event={"ID":"040697e2-a9b5-4ec0-be8d-f935227c4a0a","Type":"ContainerStarted","Data":"2cbd0a1d0450074b16edfc689fcd63c31e715d3ce01bf0540ff03191a4cbbeba"} Mar 18 20:22:11 crc kubenswrapper[4950]: I0318 20:22:11.634816 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ctmcv" podStartSLOduration=3.172669256 podStartE2EDuration="5.63480128s" podCreationTimestamp="2026-03-18 20:22:06 +0000 UTC" firstStartedPulling="2026-03-18 20:22:08.578201536 +0000 UTC m=+941.819043414" lastFinishedPulling="2026-03-18 20:22:11.04033357 +0000 UTC m=+944.281175438" observedRunningTime="2026-03-18 20:22:11.630842996 +0000 UTC m=+944.871684864" watchObservedRunningTime="2026-03-18 20:22:11.63480128 +0000 UTC m=+944.875643148" Mar 18 20:22:11 crc kubenswrapper[4950]: I0318 20:22:11.814432 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dmmkn"] Mar 18 20:22:11 crc kubenswrapper[4950]: I0318 20:22:11.815805 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dmmkn" Mar 18 20:22:11 crc kubenswrapper[4950]: I0318 20:22:11.839243 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dmmkn"] Mar 18 20:22:11 crc kubenswrapper[4950]: I0318 20:22:11.873089 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz2ht\" (UniqueName: \"kubernetes.io/projected/82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d-kube-api-access-pz2ht\") pod \"certified-operators-dmmkn\" (UID: \"82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d\") " pod="openshift-marketplace/certified-operators-dmmkn" Mar 18 20:22:11 crc kubenswrapper[4950]: I0318 20:22:11.873371 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d-catalog-content\") pod \"certified-operators-dmmkn\" (UID: \"82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d\") " pod="openshift-marketplace/certified-operators-dmmkn" Mar 18 20:22:11 crc kubenswrapper[4950]: I0318 20:22:11.873493 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d-utilities\") pod \"certified-operators-dmmkn\" (UID: \"82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d\") " pod="openshift-marketplace/certified-operators-dmmkn" Mar 18 20:22:11 crc kubenswrapper[4950]: I0318 20:22:11.975084 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d-utilities\") pod \"certified-operators-dmmkn\" (UID: \"82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d\") " pod="openshift-marketplace/certified-operators-dmmkn" Mar 18 20:22:11 crc kubenswrapper[4950]: I0318 20:22:11.975147 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz2ht\" (UniqueName: \"kubernetes.io/projected/82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d-kube-api-access-pz2ht\") pod \"certified-operators-dmmkn\" (UID: \"82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d\") " pod="openshift-marketplace/certified-operators-dmmkn" Mar 18 20:22:11 crc kubenswrapper[4950]: I0318 20:22:11.975206 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d-catalog-content\") pod \"certified-operators-dmmkn\" (UID: \"82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d\") " pod="openshift-marketplace/certified-operators-dmmkn" Mar 18 20:22:11 crc kubenswrapper[4950]: I0318 20:22:11.975833 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d-catalog-content\") pod \"certified-operators-dmmkn\" (UID: \"82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d\") " pod="openshift-marketplace/certified-operators-dmmkn" Mar 18 20:22:11 crc kubenswrapper[4950]: I0318 20:22:11.976043 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d-utilities\") pod \"certified-operators-dmmkn\" (UID: \"82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d\") " pod="openshift-marketplace/certified-operators-dmmkn" Mar 18 20:22:12 crc kubenswrapper[4950]: I0318 20:22:12.018007 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pz2ht\" (UniqueName: \"kubernetes.io/projected/82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d-kube-api-access-pz2ht\") pod \"certified-operators-dmmkn\" (UID: \"82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d\") " pod="openshift-marketplace/certified-operators-dmmkn" Mar 18 20:22:12 crc kubenswrapper[4950]: I0318 20:22:12.131924 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dmmkn" Mar 18 20:22:12 crc kubenswrapper[4950]: I0318 20:22:12.628890 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dmmkn"] Mar 18 20:22:13 crc kubenswrapper[4950]: I0318 20:22:13.622924 4950 generic.go:334] "Generic (PLEG): container finished" podID="82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d" containerID="ddf3477edafdf4ac8c40fa4d0d930dfd4244c11d24998b162fbc5caa7801b95f" exitCode=0 Mar 18 20:22:13 crc kubenswrapper[4950]: I0318 20:22:13.623455 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dmmkn" event={"ID":"82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d","Type":"ContainerDied","Data":"ddf3477edafdf4ac8c40fa4d0d930dfd4244c11d24998b162fbc5caa7801b95f"} Mar 18 20:22:13 crc kubenswrapper[4950]: I0318 20:22:13.623480 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dmmkn" event={"ID":"82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d","Type":"ContainerStarted","Data":"c2dbe3e8d062c7f7f436a8c3b189edb74fa06c217913d0e421084de685dc5c0b"} Mar 18 20:22:16 crc kubenswrapper[4950]: I0318 20:22:16.653926 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-b85c4d696-w5ws4" event={"ID":"040697e2-a9b5-4ec0-be8d-f935227c4a0a","Type":"ContainerStarted","Data":"b1a88614448e7741dfccde854e2d62d50197f005c68406360bbed482f739aef0"} Mar 18 20:22:16 crc kubenswrapper[4950]: I0318 20:22:16.654251 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-b85c4d696-w5ws4" Mar 18 20:22:16 crc kubenswrapper[4950]: I0318 20:22:16.656621 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dmmkn" event={"ID":"82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d","Type":"ContainerStarted","Data":"7da5cd9a0b38d3ab44d5427cc03dd4ff7561b4f14528eded7d717e1f7f0ccc68"} Mar 18 20:22:16 crc kubenswrapper[4950]: I0318 20:22:16.684755 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-b85c4d696-w5ws4" podStartSLOduration=1.458947515 podStartE2EDuration="6.684738051s" podCreationTimestamp="2026-03-18 20:22:10 +0000 UTC" firstStartedPulling="2026-03-18 20:22:11.216206214 +0000 UTC m=+944.457048072" lastFinishedPulling="2026-03-18 20:22:16.44199674 +0000 UTC m=+949.682838608" observedRunningTime="2026-03-18 20:22:16.67897388 +0000 UTC m=+949.919815748" watchObservedRunningTime="2026-03-18 20:22:16.684738051 +0000 UTC m=+949.925579919" Mar 18 20:22:17 crc kubenswrapper[4950]: I0318 20:22:17.153828 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ctmcv" Mar 18 20:22:17 crc kubenswrapper[4950]: I0318 20:22:17.154819 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ctmcv" Mar 18 20:22:17 crc kubenswrapper[4950]: I0318 20:22:17.202851 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ctmcv" Mar 18 20:22:17 crc kubenswrapper[4950]: I0318 20:22:17.668285 4950 generic.go:334] "Generic (PLEG): container finished" podID="82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d" containerID="7da5cd9a0b38d3ab44d5427cc03dd4ff7561b4f14528eded7d717e1f7f0ccc68" exitCode=0 Mar 18 20:22:17 crc kubenswrapper[4950]: I0318 20:22:17.668488 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dmmkn" event={"ID":"82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d","Type":"ContainerDied","Data":"7da5cd9a0b38d3ab44d5427cc03dd4ff7561b4f14528eded7d717e1f7f0ccc68"} Mar 18 20:22:17 crc kubenswrapper[4950]: I0318 20:22:17.756684 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ctmcv" Mar 18 20:22:18 crc kubenswrapper[4950]: I0318 20:22:18.676242 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dmmkn" event={"ID":"82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d","Type":"ContainerStarted","Data":"22f56e6d93a3bb19b2e3a2652c04ed5b5cf974b5a93e54d08019b5c8cdcfb7b6"} Mar 18 20:22:20 crc kubenswrapper[4950]: I0318 20:22:20.796233 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dmmkn" podStartSLOduration=5.329159163 podStartE2EDuration="9.79620563s" podCreationTimestamp="2026-03-18 20:22:11 +0000 UTC" firstStartedPulling="2026-03-18 20:22:13.625083128 +0000 UTC m=+946.865924996" lastFinishedPulling="2026-03-18 20:22:18.092129585 +0000 UTC m=+951.332971463" observedRunningTime="2026-03-18 20:22:18.698128788 +0000 UTC m=+951.938970656" watchObservedRunningTime="2026-03-18 20:22:20.79620563 +0000 UTC m=+954.037047528" Mar 18 20:22:20 crc kubenswrapper[4950]: I0318 20:22:20.802213 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ctmcv"] Mar 18 20:22:20 crc kubenswrapper[4950]: I0318 20:22:20.802680 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ctmcv" podUID="10835f63-f9b3-4dba-80a4-1af5ca39fb99" containerName="registry-server" containerID="cri-o://a595aa3392b5fab8dce5e99fd528a8f1583c8da65adf046b1194f581d401db3d" gracePeriod=2 Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.167869 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ctmcv" Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.313955 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-css4l\" (UniqueName: \"kubernetes.io/projected/10835f63-f9b3-4dba-80a4-1af5ca39fb99-kube-api-access-css4l\") pod \"10835f63-f9b3-4dba-80a4-1af5ca39fb99\" (UID: \"10835f63-f9b3-4dba-80a4-1af5ca39fb99\") " Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.315779 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10835f63-f9b3-4dba-80a4-1af5ca39fb99-utilities\") pod \"10835f63-f9b3-4dba-80a4-1af5ca39fb99\" (UID: \"10835f63-f9b3-4dba-80a4-1af5ca39fb99\") " Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.315864 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10835f63-f9b3-4dba-80a4-1af5ca39fb99-catalog-content\") pod \"10835f63-f9b3-4dba-80a4-1af5ca39fb99\" (UID: \"10835f63-f9b3-4dba-80a4-1af5ca39fb99\") " Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.317057 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10835f63-f9b3-4dba-80a4-1af5ca39fb99-utilities" (OuterVolumeSpecName: "utilities") pod "10835f63-f9b3-4dba-80a4-1af5ca39fb99" (UID: "10835f63-f9b3-4dba-80a4-1af5ca39fb99"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.321873 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10835f63-f9b3-4dba-80a4-1af5ca39fb99-kube-api-access-css4l" (OuterVolumeSpecName: "kube-api-access-css4l") pod "10835f63-f9b3-4dba-80a4-1af5ca39fb99" (UID: "10835f63-f9b3-4dba-80a4-1af5ca39fb99"). InnerVolumeSpecName "kube-api-access-css4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.344256 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10835f63-f9b3-4dba-80a4-1af5ca39fb99-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "10835f63-f9b3-4dba-80a4-1af5ca39fb99" (UID: "10835f63-f9b3-4dba-80a4-1af5ca39fb99"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.417873 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10835f63-f9b3-4dba-80a4-1af5ca39fb99-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.417923 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10835f63-f9b3-4dba-80a4-1af5ca39fb99-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.417943 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-css4l\" (UniqueName: \"kubernetes.io/projected/10835f63-f9b3-4dba-80a4-1af5ca39fb99-kube-api-access-css4l\") on node \"crc\" DevicePath \"\"" Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.698712 4950 generic.go:334] "Generic (PLEG): container finished" podID="10835f63-f9b3-4dba-80a4-1af5ca39fb99" containerID="a595aa3392b5fab8dce5e99fd528a8f1583c8da65adf046b1194f581d401db3d" exitCode=0 Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.698985 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ctmcv" event={"ID":"10835f63-f9b3-4dba-80a4-1af5ca39fb99","Type":"ContainerDied","Data":"a595aa3392b5fab8dce5e99fd528a8f1583c8da65adf046b1194f581d401db3d"} Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.699105 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ctmcv" event={"ID":"10835f63-f9b3-4dba-80a4-1af5ca39fb99","Type":"ContainerDied","Data":"02109309e034dcc0313440ee44ef187b3501e9bdbdbc09f980e1d0e530b67612"} Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.699146 4950 scope.go:117] "RemoveContainer" containerID="a595aa3392b5fab8dce5e99fd528a8f1583c8da65adf046b1194f581d401db3d" Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.699029 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ctmcv" Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.723953 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ctmcv"] Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.736952 4950 scope.go:117] "RemoveContainer" containerID="987d393dafd6ddcade06d315ec14edaa8519ad1e94b51d56111f84ed031a3853" Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.741294 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ctmcv"] Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.769424 4950 scope.go:117] "RemoveContainer" containerID="496c67be195eafd70e76f49f971c898c6cec7ce5c39c3df3adfee15685d61ac9" Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.785674 4950 scope.go:117] "RemoveContainer" containerID="a595aa3392b5fab8dce5e99fd528a8f1583c8da65adf046b1194f581d401db3d" Mar 18 20:22:21 crc kubenswrapper[4950]: E0318 20:22:21.786223 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a595aa3392b5fab8dce5e99fd528a8f1583c8da65adf046b1194f581d401db3d\": container with ID starting with a595aa3392b5fab8dce5e99fd528a8f1583c8da65adf046b1194f581d401db3d not found: ID does not exist" containerID="a595aa3392b5fab8dce5e99fd528a8f1583c8da65adf046b1194f581d401db3d" Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.786256 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a595aa3392b5fab8dce5e99fd528a8f1583c8da65adf046b1194f581d401db3d"} err="failed to get container status \"a595aa3392b5fab8dce5e99fd528a8f1583c8da65adf046b1194f581d401db3d\": rpc error: code = NotFound desc = could not find container \"a595aa3392b5fab8dce5e99fd528a8f1583c8da65adf046b1194f581d401db3d\": container with ID starting with a595aa3392b5fab8dce5e99fd528a8f1583c8da65adf046b1194f581d401db3d not found: ID does not exist" Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.786287 4950 scope.go:117] "RemoveContainer" containerID="987d393dafd6ddcade06d315ec14edaa8519ad1e94b51d56111f84ed031a3853" Mar 18 20:22:21 crc kubenswrapper[4950]: E0318 20:22:21.786745 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"987d393dafd6ddcade06d315ec14edaa8519ad1e94b51d56111f84ed031a3853\": container with ID starting with 987d393dafd6ddcade06d315ec14edaa8519ad1e94b51d56111f84ed031a3853 not found: ID does not exist" containerID="987d393dafd6ddcade06d315ec14edaa8519ad1e94b51d56111f84ed031a3853" Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.786769 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"987d393dafd6ddcade06d315ec14edaa8519ad1e94b51d56111f84ed031a3853"} err="failed to get container status \"987d393dafd6ddcade06d315ec14edaa8519ad1e94b51d56111f84ed031a3853\": rpc error: code = NotFound desc = could not find container \"987d393dafd6ddcade06d315ec14edaa8519ad1e94b51d56111f84ed031a3853\": container with ID starting with 987d393dafd6ddcade06d315ec14edaa8519ad1e94b51d56111f84ed031a3853 not found: ID does not exist" Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.786786 4950 scope.go:117] "RemoveContainer" containerID="496c67be195eafd70e76f49f971c898c6cec7ce5c39c3df3adfee15685d61ac9" Mar 18 20:22:21 crc kubenswrapper[4950]: E0318 20:22:21.787147 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"496c67be195eafd70e76f49f971c898c6cec7ce5c39c3df3adfee15685d61ac9\": container with ID starting with 496c67be195eafd70e76f49f971c898c6cec7ce5c39c3df3adfee15685d61ac9 not found: ID does not exist" containerID="496c67be195eafd70e76f49f971c898c6cec7ce5c39c3df3adfee15685d61ac9" Mar 18 20:22:21 crc kubenswrapper[4950]: I0318 20:22:21.787173 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"496c67be195eafd70e76f49f971c898c6cec7ce5c39c3df3adfee15685d61ac9"} err="failed to get container status \"496c67be195eafd70e76f49f971c898c6cec7ce5c39c3df3adfee15685d61ac9\": rpc error: code = NotFound desc = could not find container \"496c67be195eafd70e76f49f971c898c6cec7ce5c39c3df3adfee15685d61ac9\": container with ID starting with 496c67be195eafd70e76f49f971c898c6cec7ce5c39c3df3adfee15685d61ac9 not found: ID does not exist" Mar 18 20:22:22 crc kubenswrapper[4950]: I0318 20:22:22.132312 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dmmkn" Mar 18 20:22:22 crc kubenswrapper[4950]: I0318 20:22:22.133317 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dmmkn" Mar 18 20:22:22 crc kubenswrapper[4950]: I0318 20:22:22.182840 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dmmkn" Mar 18 20:22:23 crc kubenswrapper[4950]: I0318 20:22:23.494925 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10835f63-f9b3-4dba-80a4-1af5ca39fb99" path="/var/lib/kubelet/pods/10835f63-f9b3-4dba-80a4-1af5ca39fb99/volumes" Mar 18 20:22:30 crc kubenswrapper[4950]: I0318 20:22:30.729523 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-b85c4d696-w5ws4" Mar 18 20:22:32 crc kubenswrapper[4950]: I0318 20:22:32.178236 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dmmkn" Mar 18 20:22:32 crc kubenswrapper[4950]: I0318 20:22:32.256584 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dmmkn"] Mar 18 20:22:32 crc kubenswrapper[4950]: I0318 20:22:32.786033 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dmmkn" podUID="82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d" containerName="registry-server" containerID="cri-o://22f56e6d93a3bb19b2e3a2652c04ed5b5cf974b5a93e54d08019b5c8cdcfb7b6" gracePeriod=2 Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.140379 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dmmkn" Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.270655 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d-catalog-content\") pod \"82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d\" (UID: \"82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d\") " Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.270741 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d-utilities\") pod \"82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d\" (UID: \"82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d\") " Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.270818 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pz2ht\" (UniqueName: \"kubernetes.io/projected/82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d-kube-api-access-pz2ht\") pod \"82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d\" (UID: \"82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d\") " Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.271686 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d-utilities" (OuterVolumeSpecName: "utilities") pod "82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d" (UID: "82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.278649 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d-kube-api-access-pz2ht" (OuterVolumeSpecName: "kube-api-access-pz2ht") pod "82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d" (UID: "82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d"). InnerVolumeSpecName "kube-api-access-pz2ht". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.321783 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d" (UID: "82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.372177 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.372212 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.372226 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pz2ht\" (UniqueName: \"kubernetes.io/projected/82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d-kube-api-access-pz2ht\") on node \"crc\" DevicePath \"\"" Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.799120 4950 generic.go:334] "Generic (PLEG): container finished" podID="82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d" containerID="22f56e6d93a3bb19b2e3a2652c04ed5b5cf974b5a93e54d08019b5c8cdcfb7b6" exitCode=0 Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.799182 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dmmkn" event={"ID":"82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d","Type":"ContainerDied","Data":"22f56e6d93a3bb19b2e3a2652c04ed5b5cf974b5a93e54d08019b5c8cdcfb7b6"} Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.799204 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dmmkn" Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.799238 4950 scope.go:117] "RemoveContainer" containerID="22f56e6d93a3bb19b2e3a2652c04ed5b5cf974b5a93e54d08019b5c8cdcfb7b6" Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.799220 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dmmkn" event={"ID":"82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d","Type":"ContainerDied","Data":"c2dbe3e8d062c7f7f436a8c3b189edb74fa06c217913d0e421084de685dc5c0b"} Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.830277 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dmmkn"] Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.834307 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dmmkn"] Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.835110 4950 scope.go:117] "RemoveContainer" containerID="7da5cd9a0b38d3ab44d5427cc03dd4ff7561b4f14528eded7d717e1f7f0ccc68" Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.858919 4950 scope.go:117] "RemoveContainer" containerID="ddf3477edafdf4ac8c40fa4d0d930dfd4244c11d24998b162fbc5caa7801b95f" Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.888003 4950 scope.go:117] "RemoveContainer" containerID="22f56e6d93a3bb19b2e3a2652c04ed5b5cf974b5a93e54d08019b5c8cdcfb7b6" Mar 18 20:22:33 crc kubenswrapper[4950]: E0318 20:22:33.888890 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22f56e6d93a3bb19b2e3a2652c04ed5b5cf974b5a93e54d08019b5c8cdcfb7b6\": container with ID starting with 22f56e6d93a3bb19b2e3a2652c04ed5b5cf974b5a93e54d08019b5c8cdcfb7b6 not found: ID does not exist" containerID="22f56e6d93a3bb19b2e3a2652c04ed5b5cf974b5a93e54d08019b5c8cdcfb7b6" Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.888978 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22f56e6d93a3bb19b2e3a2652c04ed5b5cf974b5a93e54d08019b5c8cdcfb7b6"} err="failed to get container status \"22f56e6d93a3bb19b2e3a2652c04ed5b5cf974b5a93e54d08019b5c8cdcfb7b6\": rpc error: code = NotFound desc = could not find container \"22f56e6d93a3bb19b2e3a2652c04ed5b5cf974b5a93e54d08019b5c8cdcfb7b6\": container with ID starting with 22f56e6d93a3bb19b2e3a2652c04ed5b5cf974b5a93e54d08019b5c8cdcfb7b6 not found: ID does not exist" Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.889033 4950 scope.go:117] "RemoveContainer" containerID="7da5cd9a0b38d3ab44d5427cc03dd4ff7561b4f14528eded7d717e1f7f0ccc68" Mar 18 20:22:33 crc kubenswrapper[4950]: E0318 20:22:33.889729 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7da5cd9a0b38d3ab44d5427cc03dd4ff7561b4f14528eded7d717e1f7f0ccc68\": container with ID starting with 7da5cd9a0b38d3ab44d5427cc03dd4ff7561b4f14528eded7d717e1f7f0ccc68 not found: ID does not exist" containerID="7da5cd9a0b38d3ab44d5427cc03dd4ff7561b4f14528eded7d717e1f7f0ccc68" Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.889773 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7da5cd9a0b38d3ab44d5427cc03dd4ff7561b4f14528eded7d717e1f7f0ccc68"} err="failed to get container status \"7da5cd9a0b38d3ab44d5427cc03dd4ff7561b4f14528eded7d717e1f7f0ccc68\": rpc error: code = NotFound desc = could not find container \"7da5cd9a0b38d3ab44d5427cc03dd4ff7561b4f14528eded7d717e1f7f0ccc68\": container with ID starting with 7da5cd9a0b38d3ab44d5427cc03dd4ff7561b4f14528eded7d717e1f7f0ccc68 not found: ID does not exist" Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.889801 4950 scope.go:117] "RemoveContainer" containerID="ddf3477edafdf4ac8c40fa4d0d930dfd4244c11d24998b162fbc5caa7801b95f" Mar 18 20:22:33 crc kubenswrapper[4950]: E0318 20:22:33.890267 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddf3477edafdf4ac8c40fa4d0d930dfd4244c11d24998b162fbc5caa7801b95f\": container with ID starting with ddf3477edafdf4ac8c40fa4d0d930dfd4244c11d24998b162fbc5caa7801b95f not found: ID does not exist" containerID="ddf3477edafdf4ac8c40fa4d0d930dfd4244c11d24998b162fbc5caa7801b95f" Mar 18 20:22:33 crc kubenswrapper[4950]: I0318 20:22:33.890321 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddf3477edafdf4ac8c40fa4d0d930dfd4244c11d24998b162fbc5caa7801b95f"} err="failed to get container status \"ddf3477edafdf4ac8c40fa4d0d930dfd4244c11d24998b162fbc5caa7801b95f\": rpc error: code = NotFound desc = could not find container \"ddf3477edafdf4ac8c40fa4d0d930dfd4244c11d24998b162fbc5caa7801b95f\": container with ID starting with ddf3477edafdf4ac8c40fa4d0d930dfd4244c11d24998b162fbc5caa7801b95f not found: ID does not exist" Mar 18 20:22:35 crc kubenswrapper[4950]: I0318 20:22:35.488881 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d" path="/var/lib/kubelet/pods/82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d/volumes" Mar 18 20:22:38 crc kubenswrapper[4950]: I0318 20:22:38.252868 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mhmk7"] Mar 18 20:22:38 crc kubenswrapper[4950]: E0318 20:22:38.253080 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10835f63-f9b3-4dba-80a4-1af5ca39fb99" containerName="registry-server" Mar 18 20:22:38 crc kubenswrapper[4950]: I0318 20:22:38.253091 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="10835f63-f9b3-4dba-80a4-1af5ca39fb99" containerName="registry-server" Mar 18 20:22:38 crc kubenswrapper[4950]: E0318 20:22:38.253107 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d" containerName="registry-server" Mar 18 20:22:38 crc kubenswrapper[4950]: I0318 20:22:38.253113 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d" containerName="registry-server" Mar 18 20:22:38 crc kubenswrapper[4950]: E0318 20:22:38.253125 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d" containerName="extract-utilities" Mar 18 20:22:38 crc kubenswrapper[4950]: I0318 20:22:38.253132 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d" containerName="extract-utilities" Mar 18 20:22:38 crc kubenswrapper[4950]: E0318 20:22:38.253140 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10835f63-f9b3-4dba-80a4-1af5ca39fb99" containerName="extract-content" Mar 18 20:22:38 crc kubenswrapper[4950]: I0318 20:22:38.253145 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="10835f63-f9b3-4dba-80a4-1af5ca39fb99" containerName="extract-content" Mar 18 20:22:38 crc kubenswrapper[4950]: E0318 20:22:38.253153 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10835f63-f9b3-4dba-80a4-1af5ca39fb99" containerName="extract-utilities" Mar 18 20:22:38 crc kubenswrapper[4950]: I0318 20:22:38.253159 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="10835f63-f9b3-4dba-80a4-1af5ca39fb99" containerName="extract-utilities" Mar 18 20:22:38 crc kubenswrapper[4950]: E0318 20:22:38.253169 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d" containerName="extract-content" Mar 18 20:22:38 crc kubenswrapper[4950]: I0318 20:22:38.253174 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d" containerName="extract-content" Mar 18 20:22:38 crc kubenswrapper[4950]: I0318 20:22:38.253270 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="10835f63-f9b3-4dba-80a4-1af5ca39fb99" containerName="registry-server" Mar 18 20:22:38 crc kubenswrapper[4950]: I0318 20:22:38.253286 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="82eb1bf9-90e0-4c0a-b9ff-8c6300ba920d" containerName="registry-server" Mar 18 20:22:38 crc kubenswrapper[4950]: I0318 20:22:38.254101 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mhmk7" Mar 18 20:22:38 crc kubenswrapper[4950]: I0318 20:22:38.267721 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mhmk7"] Mar 18 20:22:38 crc kubenswrapper[4950]: I0318 20:22:38.334719 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zshsk\" (UniqueName: \"kubernetes.io/projected/3792146f-d91d-436e-b8af-fb90133a58b4-kube-api-access-zshsk\") pod \"community-operators-mhmk7\" (UID: \"3792146f-d91d-436e-b8af-fb90133a58b4\") " pod="openshift-marketplace/community-operators-mhmk7" Mar 18 20:22:38 crc kubenswrapper[4950]: I0318 20:22:38.334802 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3792146f-d91d-436e-b8af-fb90133a58b4-catalog-content\") pod \"community-operators-mhmk7\" (UID: \"3792146f-d91d-436e-b8af-fb90133a58b4\") " pod="openshift-marketplace/community-operators-mhmk7" Mar 18 20:22:38 crc kubenswrapper[4950]: I0318 20:22:38.334983 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3792146f-d91d-436e-b8af-fb90133a58b4-utilities\") pod \"community-operators-mhmk7\" (UID: \"3792146f-d91d-436e-b8af-fb90133a58b4\") " pod="openshift-marketplace/community-operators-mhmk7" Mar 18 20:22:38 crc kubenswrapper[4950]: I0318 20:22:38.436520 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zshsk\" (UniqueName: \"kubernetes.io/projected/3792146f-d91d-436e-b8af-fb90133a58b4-kube-api-access-zshsk\") pod \"community-operators-mhmk7\" (UID: \"3792146f-d91d-436e-b8af-fb90133a58b4\") " pod="openshift-marketplace/community-operators-mhmk7" Mar 18 20:22:38 crc kubenswrapper[4950]: I0318 20:22:38.436590 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3792146f-d91d-436e-b8af-fb90133a58b4-catalog-content\") pod \"community-operators-mhmk7\" (UID: \"3792146f-d91d-436e-b8af-fb90133a58b4\") " pod="openshift-marketplace/community-operators-mhmk7" Mar 18 20:22:38 crc kubenswrapper[4950]: I0318 20:22:38.436618 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3792146f-d91d-436e-b8af-fb90133a58b4-utilities\") pod \"community-operators-mhmk7\" (UID: \"3792146f-d91d-436e-b8af-fb90133a58b4\") " pod="openshift-marketplace/community-operators-mhmk7" Mar 18 20:22:38 crc kubenswrapper[4950]: I0318 20:22:38.437312 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3792146f-d91d-436e-b8af-fb90133a58b4-catalog-content\") pod \"community-operators-mhmk7\" (UID: \"3792146f-d91d-436e-b8af-fb90133a58b4\") " pod="openshift-marketplace/community-operators-mhmk7" Mar 18 20:22:38 crc kubenswrapper[4950]: I0318 20:22:38.437472 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3792146f-d91d-436e-b8af-fb90133a58b4-utilities\") pod \"community-operators-mhmk7\" (UID: \"3792146f-d91d-436e-b8af-fb90133a58b4\") " pod="openshift-marketplace/community-operators-mhmk7" Mar 18 20:22:38 crc kubenswrapper[4950]: I0318 20:22:38.469733 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zshsk\" (UniqueName: \"kubernetes.io/projected/3792146f-d91d-436e-b8af-fb90133a58b4-kube-api-access-zshsk\") pod \"community-operators-mhmk7\" (UID: \"3792146f-d91d-436e-b8af-fb90133a58b4\") " pod="openshift-marketplace/community-operators-mhmk7" Mar 18 20:22:38 crc kubenswrapper[4950]: I0318 20:22:38.567446 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mhmk7" Mar 18 20:22:39 crc kubenswrapper[4950]: I0318 20:22:39.061878 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mhmk7"] Mar 18 20:22:39 crc kubenswrapper[4950]: I0318 20:22:39.836542 4950 generic.go:334] "Generic (PLEG): container finished" podID="3792146f-d91d-436e-b8af-fb90133a58b4" containerID="b52b126695b669833de906e411ff1481eaac9565b5dd67a5176f2ee0160458e0" exitCode=0 Mar 18 20:22:39 crc kubenswrapper[4950]: I0318 20:22:39.836589 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhmk7" event={"ID":"3792146f-d91d-436e-b8af-fb90133a58b4","Type":"ContainerDied","Data":"b52b126695b669833de906e411ff1481eaac9565b5dd67a5176f2ee0160458e0"} Mar 18 20:22:39 crc kubenswrapper[4950]: I0318 20:22:39.837021 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhmk7" event={"ID":"3792146f-d91d-436e-b8af-fb90133a58b4","Type":"ContainerStarted","Data":"2504830a9b2ba50185ea53885618ee884e670af1d63a7a650df1d383dea3be4f"} Mar 18 20:22:40 crc kubenswrapper[4950]: I0318 20:22:40.843926 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhmk7" event={"ID":"3792146f-d91d-436e-b8af-fb90133a58b4","Type":"ContainerStarted","Data":"789e43e181d8ff9497f52c7eeae267bd23ff9a3faf0f78beb05885b54577c322"} Mar 18 20:22:41 crc kubenswrapper[4950]: I0318 20:22:41.854455 4950 generic.go:334] "Generic (PLEG): container finished" podID="3792146f-d91d-436e-b8af-fb90133a58b4" containerID="789e43e181d8ff9497f52c7eeae267bd23ff9a3faf0f78beb05885b54577c322" exitCode=0 Mar 18 20:22:41 crc kubenswrapper[4950]: I0318 20:22:41.854586 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhmk7" event={"ID":"3792146f-d91d-436e-b8af-fb90133a58b4","Type":"ContainerDied","Data":"789e43e181d8ff9497f52c7eeae267bd23ff9a3faf0f78beb05885b54577c322"} Mar 18 20:22:42 crc kubenswrapper[4950]: I0318 20:22:42.860953 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhmk7" event={"ID":"3792146f-d91d-436e-b8af-fb90133a58b4","Type":"ContainerStarted","Data":"08187b42c020f14f29afd1d0d41d8edafb2da9d9b65156d24698b12b454e4e9f"} Mar 18 20:22:42 crc kubenswrapper[4950]: I0318 20:22:42.879179 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mhmk7" podStartSLOduration=2.416439658 podStartE2EDuration="4.879160911s" podCreationTimestamp="2026-03-18 20:22:38 +0000 UTC" firstStartedPulling="2026-03-18 20:22:39.838503502 +0000 UTC m=+973.079345370" lastFinishedPulling="2026-03-18 20:22:42.301224735 +0000 UTC m=+975.542066623" observedRunningTime="2026-03-18 20:22:42.877528718 +0000 UTC m=+976.118370586" watchObservedRunningTime="2026-03-18 20:22:42.879160911 +0000 UTC m=+976.120002779" Mar 18 20:22:44 crc kubenswrapper[4950]: I0318 20:22:44.593766 4950 scope.go:117] "RemoveContainer" containerID="c501268b7c004983273092ea349fea475a7416cf2525156462cb37c5bdc76999" Mar 18 20:22:48 crc kubenswrapper[4950]: I0318 20:22:48.567561 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mhmk7" Mar 18 20:22:48 crc kubenswrapper[4950]: I0318 20:22:48.568124 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mhmk7" Mar 18 20:22:48 crc kubenswrapper[4950]: I0318 20:22:48.621347 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mhmk7" Mar 18 20:22:48 crc kubenswrapper[4950]: I0318 20:22:48.926863 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mhmk7" Mar 18 20:22:48 crc kubenswrapper[4950]: I0318 20:22:48.965772 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mhmk7"] Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.744041 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-l2xgb"] Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.744889 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-l2xgb" Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.748112 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-zplsd" Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.755074 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-xlx8b"] Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.756775 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-xlx8b" Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.763743 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-gsz42" Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.768549 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-l2xgb"] Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.779967 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-h9vll"] Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.780720 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-h9vll" Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.791528 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-xlx8b"] Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.791756 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-vxsgs" Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.800583 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-rgb5v"] Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.801238 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-rgb5v" Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.808935 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-fhprd" Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.825559 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-rgb5v"] Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.857518 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-h9vll"] Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.860038 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-6jt4g"] Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.860798 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-6jt4g" Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.863076 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-ktxmj" Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.901536 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-6jt4g"] Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.902912 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdqs8\" (UniqueName: \"kubernetes.io/projected/19c17169-bafd-4a5d-b255-0c50248f3e1c-kube-api-access-wdqs8\") pod \"cinder-operator-controller-manager-8d58dc466-xlx8b\" (UID: \"19c17169-bafd-4a5d-b255-0c50248f3e1c\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-xlx8b" Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.902973 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzkw2\" (UniqueName: \"kubernetes.io/projected/815a0db2-1d7b-4577-b1bf-cf3114f2a2a3-kube-api-access-zzkw2\") pod \"barbican-operator-controller-manager-59bc569d95-l2xgb\" (UID: \"815a0db2-1d7b-4577-b1bf-cf3114f2a2a3\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-l2xgb" Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.903008 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ctd7\" (UniqueName: \"kubernetes.io/projected/8081f4de-9137-49d2-a66d-bbf265033c5c-kube-api-access-5ctd7\") pod \"designate-operator-controller-manager-588d4d986b-h9vll\" (UID: \"8081f4de-9137-49d2-a66d-bbf265033c5c\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-h9vll" Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.903029 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khtkr\" (UniqueName: \"kubernetes.io/projected/bc63ecc0-0c62-4607-8ce4-8d143f0c2ce6-kube-api-access-khtkr\") pod \"glance-operator-controller-manager-79df6bcc97-rgb5v\" (UID: \"bc63ecc0-0c62-4607-8ce4-8d143f0c2ce6\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-rgb5v" Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.915545 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mhmk7" podUID="3792146f-d91d-436e-b8af-fb90133a58b4" containerName="registry-server" containerID="cri-o://08187b42c020f14f29afd1d0d41d8edafb2da9d9b65156d24698b12b454e4e9f" gracePeriod=2 Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.964980 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-kckr9"] Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.965783 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-7b9c774f96-6cg5j"] Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.966296 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-6cg5j" Mar 18 20:22:50 crc kubenswrapper[4950]: I0318 20:22:50.966681 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-kckr9" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:50.978968 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-5v45z"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.044991 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7b9c774f96-6cg5j"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.045020 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-kckr9"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.007148 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdqs8\" (UniqueName: \"kubernetes.io/projected/19c17169-bafd-4a5d-b255-0c50248f3e1c-kube-api-access-wdqs8\") pod \"cinder-operator-controller-manager-8d58dc466-xlx8b\" (UID: \"19c17169-bafd-4a5d-b255-0c50248f3e1c\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-xlx8b" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.045117 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzkw2\" (UniqueName: \"kubernetes.io/projected/815a0db2-1d7b-4577-b1bf-cf3114f2a2a3-kube-api-access-zzkw2\") pod \"barbican-operator-controller-manager-59bc569d95-l2xgb\" (UID: \"815a0db2-1d7b-4577-b1bf-cf3114f2a2a3\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-l2xgb" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.045147 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxkq6\" (UniqueName: \"kubernetes.io/projected/9046fce8-e07a-4cc6-9577-4ab38ecac0f3-kube-api-access-sxkq6\") pod \"heat-operator-controller-manager-67dd5f86f5-6jt4g\" (UID: \"9046fce8-e07a-4cc6-9577-4ab38ecac0f3\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-6jt4g" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.045203 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ctd7\" (UniqueName: \"kubernetes.io/projected/8081f4de-9137-49d2-a66d-bbf265033c5c-kube-api-access-5ctd7\") pod \"designate-operator-controller-manager-588d4d986b-h9vll\" (UID: \"8081f4de-9137-49d2-a66d-bbf265033c5c\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-h9vll" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.045230 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khtkr\" (UniqueName: \"kubernetes.io/projected/bc63ecc0-0c62-4607-8ce4-8d143f0c2ce6-kube-api-access-khtkr\") pod \"glance-operator-controller-manager-79df6bcc97-rgb5v\" (UID: \"bc63ecc0-0c62-4607-8ce4-8d143f0c2ce6\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-rgb5v" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:50.981535 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:50.981560 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-h4f4m" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:50.991782 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-bsfbl" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.045921 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-5v45z" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.060490 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-5v45z"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.068993 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-n4dx6" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.070034 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdqs8\" (UniqueName: \"kubernetes.io/projected/19c17169-bafd-4a5d-b255-0c50248f3e1c-kube-api-access-wdqs8\") pod \"cinder-operator-controller-manager-8d58dc466-xlx8b\" (UID: \"19c17169-bafd-4a5d-b255-0c50248f3e1c\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-xlx8b" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.073988 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-xlx8b" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.078230 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-dls6n"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.083282 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-dls6n" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.088666 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-ddcxz" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.089681 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzkw2\" (UniqueName: \"kubernetes.io/projected/815a0db2-1d7b-4577-b1bf-cf3114f2a2a3-kube-api-access-zzkw2\") pod \"barbican-operator-controller-manager-59bc569d95-l2xgb\" (UID: \"815a0db2-1d7b-4577-b1bf-cf3114f2a2a3\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-l2xgb" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.093980 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ctd7\" (UniqueName: \"kubernetes.io/projected/8081f4de-9137-49d2-a66d-bbf265033c5c-kube-api-access-5ctd7\") pod \"designate-operator-controller-manager-588d4d986b-h9vll\" (UID: \"8081f4de-9137-49d2-a66d-bbf265033c5c\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-h9vll" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.094869 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khtkr\" (UniqueName: \"kubernetes.io/projected/bc63ecc0-0c62-4607-8ce4-8d143f0c2ce6-kube-api-access-khtkr\") pod \"glance-operator-controller-manager-79df6bcc97-rgb5v\" (UID: \"bc63ecc0-0c62-4607-8ce4-8d143f0c2ce6\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-rgb5v" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.099781 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-h9vll" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.119679 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-dls6n"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.120011 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-rgb5v" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.146832 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xf2t\" (UniqueName: \"kubernetes.io/projected/0277fe72-c2e8-4f07-ae64-bb1232a98518-kube-api-access-5xf2t\") pod \"ironic-operator-controller-manager-6f787dddc9-kckr9\" (UID: \"0277fe72-c2e8-4f07-ae64-bb1232a98518\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-kckr9" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.147398 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/df0f2790-244c-4b4f-b3e9-d82e69bb9044-cert\") pod \"infra-operator-controller-manager-7b9c774f96-6cg5j\" (UID: \"df0f2790-244c-4b4f-b3e9-d82e69bb9044\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-6cg5j" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.147440 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swbbr\" (UniqueName: \"kubernetes.io/projected/df0f2790-244c-4b4f-b3e9-d82e69bb9044-kube-api-access-swbbr\") pod \"infra-operator-controller-manager-7b9c774f96-6cg5j\" (UID: \"df0f2790-244c-4b4f-b3e9-d82e69bb9044\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-6cg5j" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.147527 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxkq6\" (UniqueName: \"kubernetes.io/projected/9046fce8-e07a-4cc6-9577-4ab38ecac0f3-kube-api-access-sxkq6\") pod \"heat-operator-controller-manager-67dd5f86f5-6jt4g\" (UID: \"9046fce8-e07a-4cc6-9577-4ab38ecac0f3\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-6jt4g" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.152552 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-ftljz"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.153317 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-55f864c847-ftljz" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.166652 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-ftljz"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.170564 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-klnj8" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.189991 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-5h94g"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.191364 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-5h94g" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.211321 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxkq6\" (UniqueName: \"kubernetes.io/projected/9046fce8-e07a-4cc6-9577-4ab38ecac0f3-kube-api-access-sxkq6\") pod \"heat-operator-controller-manager-67dd5f86f5-6jt4g\" (UID: \"9046fce8-e07a-4cc6-9577-4ab38ecac0f3\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-6jt4g" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.211760 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-ptfs7" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.242480 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-5h94g"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.249158 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46bfk\" (UniqueName: \"kubernetes.io/projected/516ce0ee-83c3-40a3-bb00-2a88891a5791-kube-api-access-46bfk\") pod \"keystone-operator-controller-manager-768b96df4c-dls6n\" (UID: \"516ce0ee-83c3-40a3-bb00-2a88891a5791\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-dls6n" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.249227 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xf2t\" (UniqueName: \"kubernetes.io/projected/0277fe72-c2e8-4f07-ae64-bb1232a98518-kube-api-access-5xf2t\") pod \"ironic-operator-controller-manager-6f787dddc9-kckr9\" (UID: \"0277fe72-c2e8-4f07-ae64-bb1232a98518\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-kckr9" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.249250 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqr65\" (UniqueName: \"kubernetes.io/projected/edc5286e-aa7d-4ee6-a8f7-3f5cad640484-kube-api-access-lqr65\") pod \"manila-operator-controller-manager-55f864c847-ftljz\" (UID: \"edc5286e-aa7d-4ee6-a8f7-3f5cad640484\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-ftljz" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.249275 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzw6w\" (UniqueName: \"kubernetes.io/projected/926aae26-bb2f-4cac-af12-abbd3585a90a-kube-api-access-vzw6w\") pod \"horizon-operator-controller-manager-8464cc45fb-5v45z\" (UID: \"926aae26-bb2f-4cac-af12-abbd3585a90a\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-5v45z" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.249292 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/df0f2790-244c-4b4f-b3e9-d82e69bb9044-cert\") pod \"infra-operator-controller-manager-7b9c774f96-6cg5j\" (UID: \"df0f2790-244c-4b4f-b3e9-d82e69bb9044\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-6cg5j" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.249313 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swbbr\" (UniqueName: \"kubernetes.io/projected/df0f2790-244c-4b4f-b3e9-d82e69bb9044-kube-api-access-swbbr\") pod \"infra-operator-controller-manager-7b9c774f96-6cg5j\" (UID: \"df0f2790-244c-4b4f-b3e9-d82e69bb9044\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-6cg5j" Mar 18 20:22:51 crc kubenswrapper[4950]: E0318 20:22:51.249675 4950 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 18 20:22:51 crc kubenswrapper[4950]: E0318 20:22:51.249713 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/df0f2790-244c-4b4f-b3e9-d82e69bb9044-cert podName:df0f2790-244c-4b4f-b3e9-d82e69bb9044 nodeName:}" failed. No retries permitted until 2026-03-18 20:22:51.749697015 +0000 UTC m=+984.990538883 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/df0f2790-244c-4b4f-b3e9-d82e69bb9044-cert") pod "infra-operator-controller-manager-7b9c774f96-6cg5j" (UID: "df0f2790-244c-4b4f-b3e9-d82e69bb9044") : secret "infra-operator-webhook-server-cert" not found Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.250818 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-qxz4b"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.252842 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-767865f676-qxz4b" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.255659 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-bsv4c"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.262079 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-bsv4c" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.267250 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-v9hfl" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.268280 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-5tpms" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.270158 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-pvw7d"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.271043 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-pvw7d" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.278067 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-qxz4b"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.281472 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swbbr\" (UniqueName: \"kubernetes.io/projected/df0f2790-244c-4b4f-b3e9-d82e69bb9044-kube-api-access-swbbr\") pod \"infra-operator-controller-manager-7b9c774f96-6cg5j\" (UID: \"df0f2790-244c-4b4f-b3e9-d82e69bb9044\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-6cg5j" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.288476 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-bsv4c"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.290843 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-q8sm7" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.292553 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-pvw7d"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.326272 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xf2t\" (UniqueName: \"kubernetes.io/projected/0277fe72-c2e8-4f07-ae64-bb1232a98518-kube-api-access-5xf2t\") pod \"ironic-operator-controller-manager-6f787dddc9-kckr9\" (UID: \"0277fe72-c2e8-4f07-ae64-bb1232a98518\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-kckr9" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.327452 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-74c4796899rqqwt"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.332363 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-74c4796899rqqwt" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.334499 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-5xzct" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.334645 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.341829 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-zcjsh"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.342635 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5784578c99-zcjsh" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.348546 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-d6t88"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.348766 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-4kd94" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.349321 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-884679f54-d6t88" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.350159 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqr65\" (UniqueName: \"kubernetes.io/projected/edc5286e-aa7d-4ee6-a8f7-3f5cad640484-kube-api-access-lqr65\") pod \"manila-operator-controller-manager-55f864c847-ftljz\" (UID: \"edc5286e-aa7d-4ee6-a8f7-3f5cad640484\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-ftljz" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.350187 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfhtk\" (UniqueName: \"kubernetes.io/projected/ec5a5aa1-65cc-4bb2-8644-fa0c7bb44da4-kube-api-access-pfhtk\") pod \"neutron-operator-controller-manager-767865f676-qxz4b\" (UID: \"ec5a5aa1-65cc-4bb2-8644-fa0c7bb44da4\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-qxz4b" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.350210 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzw6w\" (UniqueName: \"kubernetes.io/projected/926aae26-bb2f-4cac-af12-abbd3585a90a-kube-api-access-vzw6w\") pod \"horizon-operator-controller-manager-8464cc45fb-5v45z\" (UID: \"926aae26-bb2f-4cac-af12-abbd3585a90a\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-5v45z" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.350269 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46bfk\" (UniqueName: \"kubernetes.io/projected/516ce0ee-83c3-40a3-bb00-2a88891a5791-kube-api-access-46bfk\") pod \"keystone-operator-controller-manager-768b96df4c-dls6n\" (UID: \"516ce0ee-83c3-40a3-bb00-2a88891a5791\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-dls6n" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.350291 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf5ml\" (UniqueName: \"kubernetes.io/projected/80caf1bd-ddfd-498d-9b80-ea8cd37a3c72-kube-api-access-xf5ml\") pod \"mariadb-operator-controller-manager-67ccfc9778-5h94g\" (UID: \"80caf1bd-ddfd-498d-9b80-ea8cd37a3c72\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-5h94g" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.361562 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-65xxg" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.362992 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-zcjsh"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.363101 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-l2xgb" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.378473 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-74c4796899rqqwt"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.385432 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-kckr9" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.395220 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-d6t88"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.406470 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqr65\" (UniqueName: \"kubernetes.io/projected/edc5286e-aa7d-4ee6-a8f7-3f5cad640484-kube-api-access-lqr65\") pod \"manila-operator-controller-manager-55f864c847-ftljz\" (UID: \"edc5286e-aa7d-4ee6-a8f7-3f5cad640484\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-ftljz" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.406973 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzw6w\" (UniqueName: \"kubernetes.io/projected/926aae26-bb2f-4cac-af12-abbd3585a90a-kube-api-access-vzw6w\") pod \"horizon-operator-controller-manager-8464cc45fb-5v45z\" (UID: \"926aae26-bb2f-4cac-af12-abbd3585a90a\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-5v45z" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.415963 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46bfk\" (UniqueName: \"kubernetes.io/projected/516ce0ee-83c3-40a3-bb00-2a88891a5791-kube-api-access-46bfk\") pod \"keystone-operator-controller-manager-768b96df4c-dls6n\" (UID: \"516ce0ee-83c3-40a3-bb00-2a88891a5791\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-dls6n" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.446016 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-8k97x"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.446829 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-8k97x" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.449473 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-bfzxt" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.450975 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kspc8\" (UniqueName: \"kubernetes.io/projected/29d3abf6-9968-4603-9c38-d3c1809b7275-kube-api-access-kspc8\") pod \"octavia-operator-controller-manager-5b9f45d989-pvw7d\" (UID: \"29d3abf6-9968-4603-9c38-d3c1809b7275\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-pvw7d" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.451005 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rqnw\" (UniqueName: \"kubernetes.io/projected/6b04874c-d0d9-44e7-8047-b0cc2ba6bdfc-kube-api-access-7rqnw\") pod \"placement-operator-controller-manager-5784578c99-zcjsh\" (UID: \"6b04874c-d0d9-44e7-8047-b0cc2ba6bdfc\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-zcjsh" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.451047 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfhtk\" (UniqueName: \"kubernetes.io/projected/ec5a5aa1-65cc-4bb2-8644-fa0c7bb44da4-kube-api-access-pfhtk\") pod \"neutron-operator-controller-manager-767865f676-qxz4b\" (UID: \"ec5a5aa1-65cc-4bb2-8644-fa0c7bb44da4\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-qxz4b" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.451106 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlbq7\" (UniqueName: \"kubernetes.io/projected/f216b689-dfc9-4042-92f3-71014b3a2698-kube-api-access-qlbq7\") pod \"ovn-operator-controller-manager-884679f54-d6t88\" (UID: \"f216b689-dfc9-4042-92f3-71014b3a2698\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-d6t88" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.451126 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gmfm\" (UniqueName: \"kubernetes.io/projected/608c505a-eb24-4cd9-aa6d-e07d4d8edc0c-kube-api-access-5gmfm\") pod \"nova-operator-controller-manager-5d488d59fb-bsv4c\" (UID: \"608c505a-eb24-4cd9-aa6d-e07d4d8edc0c\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-bsv4c" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.451145 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5a502f8b-719d-4f52-812d-c559a8772461-cert\") pod \"openstack-baremetal-operator-controller-manager-74c4796899rqqwt\" (UID: \"5a502f8b-719d-4f52-812d-c559a8772461\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-74c4796899rqqwt" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.451160 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbdmd\" (UniqueName: \"kubernetes.io/projected/5a502f8b-719d-4f52-812d-c559a8772461-kube-api-access-hbdmd\") pod \"openstack-baremetal-operator-controller-manager-74c4796899rqqwt\" (UID: \"5a502f8b-719d-4f52-812d-c559a8772461\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-74c4796899rqqwt" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.451189 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf5ml\" (UniqueName: \"kubernetes.io/projected/80caf1bd-ddfd-498d-9b80-ea8cd37a3c72-kube-api-access-xf5ml\") pod \"mariadb-operator-controller-manager-67ccfc9778-5h94g\" (UID: \"80caf1bd-ddfd-498d-9b80-ea8cd37a3c72\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-5h94g" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.478067 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-5v45z" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.487479 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-6jt4g" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.513976 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-dls6n" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.529393 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfhtk\" (UniqueName: \"kubernetes.io/projected/ec5a5aa1-65cc-4bb2-8644-fa0c7bb44da4-kube-api-access-pfhtk\") pod \"neutron-operator-controller-manager-767865f676-qxz4b\" (UID: \"ec5a5aa1-65cc-4bb2-8644-fa0c7bb44da4\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-qxz4b" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.535695 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf5ml\" (UniqueName: \"kubernetes.io/projected/80caf1bd-ddfd-498d-9b80-ea8cd37a3c72-kube-api-access-xf5ml\") pod \"mariadb-operator-controller-manager-67ccfc9778-5h94g\" (UID: \"80caf1bd-ddfd-498d-9b80-ea8cd37a3c72\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-5h94g" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.553427 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f7rz\" (UniqueName: \"kubernetes.io/projected/b8e294f2-ac3e-4dbc-b2eb-cd4d2b947789-kube-api-access-7f7rz\") pod \"telemetry-operator-controller-manager-d6b694c5-8k97x\" (UID: \"b8e294f2-ac3e-4dbc-b2eb-cd4d2b947789\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-8k97x" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.557623 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlbq7\" (UniqueName: \"kubernetes.io/projected/f216b689-dfc9-4042-92f3-71014b3a2698-kube-api-access-qlbq7\") pod \"ovn-operator-controller-manager-884679f54-d6t88\" (UID: \"f216b689-dfc9-4042-92f3-71014b3a2698\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-d6t88" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.557701 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gmfm\" (UniqueName: \"kubernetes.io/projected/608c505a-eb24-4cd9-aa6d-e07d4d8edc0c-kube-api-access-5gmfm\") pod \"nova-operator-controller-manager-5d488d59fb-bsv4c\" (UID: \"608c505a-eb24-4cd9-aa6d-e07d4d8edc0c\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-bsv4c" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.557733 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5a502f8b-719d-4f52-812d-c559a8772461-cert\") pod \"openstack-baremetal-operator-controller-manager-74c4796899rqqwt\" (UID: \"5a502f8b-719d-4f52-812d-c559a8772461\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-74c4796899rqqwt" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.557758 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbdmd\" (UniqueName: \"kubernetes.io/projected/5a502f8b-719d-4f52-812d-c559a8772461-kube-api-access-hbdmd\") pod \"openstack-baremetal-operator-controller-manager-74c4796899rqqwt\" (UID: \"5a502f8b-719d-4f52-812d-c559a8772461\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-74c4796899rqqwt" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.557876 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kspc8\" (UniqueName: \"kubernetes.io/projected/29d3abf6-9968-4603-9c38-d3c1809b7275-kube-api-access-kspc8\") pod \"octavia-operator-controller-manager-5b9f45d989-pvw7d\" (UID: \"29d3abf6-9968-4603-9c38-d3c1809b7275\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-pvw7d" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.557914 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rqnw\" (UniqueName: \"kubernetes.io/projected/6b04874c-d0d9-44e7-8047-b0cc2ba6bdfc-kube-api-access-7rqnw\") pod \"placement-operator-controller-manager-5784578c99-zcjsh\" (UID: \"6b04874c-d0d9-44e7-8047-b0cc2ba6bdfc\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-zcjsh" Mar 18 20:22:51 crc kubenswrapper[4950]: E0318 20:22:51.558702 4950 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 18 20:22:51 crc kubenswrapper[4950]: E0318 20:22:51.558746 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5a502f8b-719d-4f52-812d-c559a8772461-cert podName:5a502f8b-719d-4f52-812d-c559a8772461 nodeName:}" failed. No retries permitted until 2026-03-18 20:22:52.058731305 +0000 UTC m=+985.299573173 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5a502f8b-719d-4f52-812d-c559a8772461-cert") pod "openstack-baremetal-operator-controller-manager-74c4796899rqqwt" (UID: "5a502f8b-719d-4f52-812d-c559a8772461") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.568098 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-55f864c847-ftljz" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.593355 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-5h94g" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.600360 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-tkbrk"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.601154 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-wnrmf"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.601819 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-c674c5965-tkbrk" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.602319 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-wnrmf" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.606190 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-8k97x"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.614359 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-wcjtn" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.614555 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-8djp9" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.634188 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-767865f676-qxz4b" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.641484 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-tkbrk"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.646450 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kspc8\" (UniqueName: \"kubernetes.io/projected/29d3abf6-9968-4603-9c38-d3c1809b7275-kube-api-access-kspc8\") pod \"octavia-operator-controller-manager-5b9f45d989-pvw7d\" (UID: \"29d3abf6-9968-4603-9c38-d3c1809b7275\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-pvw7d" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.651261 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbdmd\" (UniqueName: \"kubernetes.io/projected/5a502f8b-719d-4f52-812d-c559a8772461-kube-api-access-hbdmd\") pod \"openstack-baremetal-operator-controller-manager-74c4796899rqqwt\" (UID: \"5a502f8b-719d-4f52-812d-c559a8772461\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-74c4796899rqqwt" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.658798 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f7rz\" (UniqueName: \"kubernetes.io/projected/b8e294f2-ac3e-4dbc-b2eb-cd4d2b947789-kube-api-access-7f7rz\") pod \"telemetry-operator-controller-manager-d6b694c5-8k97x\" (UID: \"b8e294f2-ac3e-4dbc-b2eb-cd4d2b947789\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-8k97x" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.664142 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gmfm\" (UniqueName: \"kubernetes.io/projected/608c505a-eb24-4cd9-aa6d-e07d4d8edc0c-kube-api-access-5gmfm\") pod \"nova-operator-controller-manager-5d488d59fb-bsv4c\" (UID: \"608c505a-eb24-4cd9-aa6d-e07d4d8edc0c\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-bsv4c" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.664530 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rqnw\" (UniqueName: \"kubernetes.io/projected/6b04874c-d0d9-44e7-8047-b0cc2ba6bdfc-kube-api-access-7rqnw\") pod \"placement-operator-controller-manager-5784578c99-zcjsh\" (UID: \"6b04874c-d0d9-44e7-8047-b0cc2ba6bdfc\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-zcjsh" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.671305 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlbq7\" (UniqueName: \"kubernetes.io/projected/f216b689-dfc9-4042-92f3-71014b3a2698-kube-api-access-qlbq7\") pod \"ovn-operator-controller-manager-884679f54-d6t88\" (UID: \"f216b689-dfc9-4042-92f3-71014b3a2698\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-d6t88" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.679507 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-wnrmf"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.684122 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-884679f54-d6t88" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.698466 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-v6qqx"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.699535 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-v6qqx" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.700824 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-pvw7d" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.703033 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-5fx6t" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.705296 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f7rz\" (UniqueName: \"kubernetes.io/projected/b8e294f2-ac3e-4dbc-b2eb-cd4d2b947789-kube-api-access-7f7rz\") pod \"telemetry-operator-controller-manager-d6b694c5-8k97x\" (UID: \"b8e294f2-ac3e-4dbc-b2eb-cd4d2b947789\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-8k97x" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.717461 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-8k97x" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.743245 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-v6qqx"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.761358 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8mcp\" (UniqueName: \"kubernetes.io/projected/e4b560c4-a26f-4acf-bdc3-bb9cabbbd982-kube-api-access-k8mcp\") pod \"test-operator-controller-manager-5c5cb9c4d7-wnrmf\" (UID: \"e4b560c4-a26f-4acf-bdc3-bb9cabbbd982\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-wnrmf" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.761562 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/df0f2790-244c-4b4f-b3e9-d82e69bb9044-cert\") pod \"infra-operator-controller-manager-7b9c774f96-6cg5j\" (UID: \"df0f2790-244c-4b4f-b3e9-d82e69bb9044\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-6cg5j" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.761638 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgljg\" (UniqueName: \"kubernetes.io/projected/598a0d70-d050-4d4c-8ef5-e274784e5c92-kube-api-access-xgljg\") pod \"swift-operator-controller-manager-c674c5965-tkbrk\" (UID: \"598a0d70-d050-4d4c-8ef5-e274784e5c92\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-tkbrk" Mar 18 20:22:51 crc kubenswrapper[4950]: E0318 20:22:51.761784 4950 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 18 20:22:51 crc kubenswrapper[4950]: E0318 20:22:51.761825 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/df0f2790-244c-4b4f-b3e9-d82e69bb9044-cert podName:df0f2790-244c-4b4f-b3e9-d82e69bb9044 nodeName:}" failed. No retries permitted until 2026-03-18 20:22:52.761810182 +0000 UTC m=+986.002652050 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/df0f2790-244c-4b4f-b3e9-d82e69bb9044-cert") pod "infra-operator-controller-manager-7b9c774f96-6cg5j" (UID: "df0f2790-244c-4b4f-b3e9-d82e69bb9044") : secret "infra-operator-webhook-server-cert" not found Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.811195 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5784578c99-zcjsh" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.868463 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgljg\" (UniqueName: \"kubernetes.io/projected/598a0d70-d050-4d4c-8ef5-e274784e5c92-kube-api-access-xgljg\") pod \"swift-operator-controller-manager-c674c5965-tkbrk\" (UID: \"598a0d70-d050-4d4c-8ef5-e274784e5c92\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-tkbrk" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.868519 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8mcp\" (UniqueName: \"kubernetes.io/projected/e4b560c4-a26f-4acf-bdc3-bb9cabbbd982-kube-api-access-k8mcp\") pod \"test-operator-controller-manager-5c5cb9c4d7-wnrmf\" (UID: \"e4b560c4-a26f-4acf-bdc3-bb9cabbbd982\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-wnrmf" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.868577 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rcb4\" (UniqueName: \"kubernetes.io/projected/e3196503-22ca-4f3a-ac09-f37171f08d37-kube-api-access-2rcb4\") pod \"watcher-operator-controller-manager-6c4d75f7f9-v6qqx\" (UID: \"e3196503-22ca-4f3a-ac09-f37171f08d37\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-v6qqx" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.908846 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgljg\" (UniqueName: \"kubernetes.io/projected/598a0d70-d050-4d4c-8ef5-e274784e5c92-kube-api-access-xgljg\") pod \"swift-operator-controller-manager-c674c5965-tkbrk\" (UID: \"598a0d70-d050-4d4c-8ef5-e274784e5c92\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-tkbrk" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.909665 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8mcp\" (UniqueName: \"kubernetes.io/projected/e4b560c4-a26f-4acf-bdc3-bb9cabbbd982-kube-api-access-k8mcp\") pod \"test-operator-controller-manager-5c5cb9c4d7-wnrmf\" (UID: \"e4b560c4-a26f-4acf-bdc3-bb9cabbbd982\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-wnrmf" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.913545 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr"] Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.914401 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.922465 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-8glkq" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.925335 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.925395 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.963719 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-bsv4c" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.971353 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rcb4\" (UniqueName: \"kubernetes.io/projected/e3196503-22ca-4f3a-ac09-f37171f08d37-kube-api-access-2rcb4\") pod \"watcher-operator-controller-manager-6c4d75f7f9-v6qqx\" (UID: \"e3196503-22ca-4f3a-ac09-f37171f08d37\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-v6qqx" Mar 18 20:22:51 crc kubenswrapper[4950]: I0318 20:22:51.985331 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr"] Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.012040 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rcb4\" (UniqueName: \"kubernetes.io/projected/e3196503-22ca-4f3a-ac09-f37171f08d37-kube-api-access-2rcb4\") pod \"watcher-operator-controller-manager-6c4d75f7f9-v6qqx\" (UID: \"e3196503-22ca-4f3a-ac09-f37171f08d37\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-v6qqx" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.046829 4950 generic.go:334] "Generic (PLEG): container finished" podID="3792146f-d91d-436e-b8af-fb90133a58b4" containerID="08187b42c020f14f29afd1d0d41d8edafb2da9d9b65156d24698b12b454e4e9f" exitCode=0 Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.046874 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhmk7" event={"ID":"3792146f-d91d-436e-b8af-fb90133a58b4","Type":"ContainerDied","Data":"08187b42c020f14f29afd1d0d41d8edafb2da9d9b65156d24698b12b454e4e9f"} Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.075717 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-metrics-certs\") pod \"openstack-operator-controller-manager-86bd8996f6-fxpwr\" (UID: \"151d2334-a121-46f8-947b-3b17a21abcaf\") " pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.075764 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7v8p7\" (UniqueName: \"kubernetes.io/projected/151d2334-a121-46f8-947b-3b17a21abcaf-kube-api-access-7v8p7\") pod \"openstack-operator-controller-manager-86bd8996f6-fxpwr\" (UID: \"151d2334-a121-46f8-947b-3b17a21abcaf\") " pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.075816 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-webhook-certs\") pod \"openstack-operator-controller-manager-86bd8996f6-fxpwr\" (UID: \"151d2334-a121-46f8-947b-3b17a21abcaf\") " pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.075850 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5a502f8b-719d-4f52-812d-c559a8772461-cert\") pod \"openstack-baremetal-operator-controller-manager-74c4796899rqqwt\" (UID: \"5a502f8b-719d-4f52-812d-c559a8772461\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-74c4796899rqqwt" Mar 18 20:22:52 crc kubenswrapper[4950]: E0318 20:22:52.075966 4950 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 18 20:22:52 crc kubenswrapper[4950]: E0318 20:22:52.076023 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5a502f8b-719d-4f52-812d-c559a8772461-cert podName:5a502f8b-719d-4f52-812d-c559a8772461 nodeName:}" failed. No retries permitted until 2026-03-18 20:22:53.076007308 +0000 UTC m=+986.316849176 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5a502f8b-719d-4f52-812d-c559a8772461-cert") pod "openstack-baremetal-operator-controller-manager-74c4796899rqqwt" (UID: "5a502f8b-719d-4f52-812d-c559a8772461") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.106656 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-wnrmf" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.117733 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7vkgn"] Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.118623 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7vkgn" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.134867 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-qf88m" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.135546 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-c674c5965-tkbrk" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.148474 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7vkgn"] Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.160473 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-v6qqx" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.191115 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-metrics-certs\") pod \"openstack-operator-controller-manager-86bd8996f6-fxpwr\" (UID: \"151d2334-a121-46f8-947b-3b17a21abcaf\") " pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.191164 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7v8p7\" (UniqueName: \"kubernetes.io/projected/151d2334-a121-46f8-947b-3b17a21abcaf-kube-api-access-7v8p7\") pod \"openstack-operator-controller-manager-86bd8996f6-fxpwr\" (UID: \"151d2334-a121-46f8-947b-3b17a21abcaf\") " pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.191210 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-webhook-certs\") pod \"openstack-operator-controller-manager-86bd8996f6-fxpwr\" (UID: \"151d2334-a121-46f8-947b-3b17a21abcaf\") " pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:22:52 crc kubenswrapper[4950]: E0318 20:22:52.191345 4950 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 18 20:22:52 crc kubenswrapper[4950]: E0318 20:22:52.191396 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-webhook-certs podName:151d2334-a121-46f8-947b-3b17a21abcaf nodeName:}" failed. No retries permitted until 2026-03-18 20:22:52.69138112 +0000 UTC m=+985.932222978 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-webhook-certs") pod "openstack-operator-controller-manager-86bd8996f6-fxpwr" (UID: "151d2334-a121-46f8-947b-3b17a21abcaf") : secret "webhook-server-cert" not found Mar 18 20:22:52 crc kubenswrapper[4950]: E0318 20:22:52.191663 4950 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 18 20:22:52 crc kubenswrapper[4950]: E0318 20:22:52.191690 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-metrics-certs podName:151d2334-a121-46f8-947b-3b17a21abcaf nodeName:}" failed. No retries permitted until 2026-03-18 20:22:52.691682768 +0000 UTC m=+985.932524626 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-metrics-certs") pod "openstack-operator-controller-manager-86bd8996f6-fxpwr" (UID: "151d2334-a121-46f8-947b-3b17a21abcaf") : secret "metrics-server-cert" not found Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.238441 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7v8p7\" (UniqueName: \"kubernetes.io/projected/151d2334-a121-46f8-947b-3b17a21abcaf-kube-api-access-7v8p7\") pod \"openstack-operator-controller-manager-86bd8996f6-fxpwr\" (UID: \"151d2334-a121-46f8-947b-3b17a21abcaf\") " pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.302303 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5f2sv\" (UniqueName: \"kubernetes.io/projected/283bb5c8-8799-4b57-a3c0-3a8b28f3c2d6-kube-api-access-5f2sv\") pod \"rabbitmq-cluster-operator-manager-668c99d594-7vkgn\" (UID: \"283bb5c8-8799-4b57-a3c0-3a8b28f3c2d6\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7vkgn" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.359645 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mhmk7" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.403469 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5f2sv\" (UniqueName: \"kubernetes.io/projected/283bb5c8-8799-4b57-a3c0-3a8b28f3c2d6-kube-api-access-5f2sv\") pod \"rabbitmq-cluster-operator-manager-668c99d594-7vkgn\" (UID: \"283bb5c8-8799-4b57-a3c0-3a8b28f3c2d6\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7vkgn" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.437887 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-rgb5v"] Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.456916 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5f2sv\" (UniqueName: \"kubernetes.io/projected/283bb5c8-8799-4b57-a3c0-3a8b28f3c2d6-kube-api-access-5f2sv\") pod \"rabbitmq-cluster-operator-manager-668c99d594-7vkgn\" (UID: \"283bb5c8-8799-4b57-a3c0-3a8b28f3c2d6\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7vkgn" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.504084 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3792146f-d91d-436e-b8af-fb90133a58b4-utilities\") pod \"3792146f-d91d-436e-b8af-fb90133a58b4\" (UID: \"3792146f-d91d-436e-b8af-fb90133a58b4\") " Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.504567 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zshsk\" (UniqueName: \"kubernetes.io/projected/3792146f-d91d-436e-b8af-fb90133a58b4-kube-api-access-zshsk\") pod \"3792146f-d91d-436e-b8af-fb90133a58b4\" (UID: \"3792146f-d91d-436e-b8af-fb90133a58b4\") " Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.505438 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3792146f-d91d-436e-b8af-fb90133a58b4-catalog-content\") pod \"3792146f-d91d-436e-b8af-fb90133a58b4\" (UID: \"3792146f-d91d-436e-b8af-fb90133a58b4\") " Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.506367 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3792146f-d91d-436e-b8af-fb90133a58b4-utilities" (OuterVolumeSpecName: "utilities") pod "3792146f-d91d-436e-b8af-fb90133a58b4" (UID: "3792146f-d91d-436e-b8af-fb90133a58b4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.507726 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3792146f-d91d-436e-b8af-fb90133a58b4-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.512511 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3792146f-d91d-436e-b8af-fb90133a58b4-kube-api-access-zshsk" (OuterVolumeSpecName: "kube-api-access-zshsk") pod "3792146f-d91d-436e-b8af-fb90133a58b4" (UID: "3792146f-d91d-436e-b8af-fb90133a58b4"). InnerVolumeSpecName "kube-api-access-zshsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.571700 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3792146f-d91d-436e-b8af-fb90133a58b4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3792146f-d91d-436e-b8af-fb90133a58b4" (UID: "3792146f-d91d-436e-b8af-fb90133a58b4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.603485 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7vkgn" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.615863 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3792146f-d91d-436e-b8af-fb90133a58b4-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.615965 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zshsk\" (UniqueName: \"kubernetes.io/projected/3792146f-d91d-436e-b8af-fb90133a58b4-kube-api-access-zshsk\") on node \"crc\" DevicePath \"\"" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.719339 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-webhook-certs\") pod \"openstack-operator-controller-manager-86bd8996f6-fxpwr\" (UID: \"151d2334-a121-46f8-947b-3b17a21abcaf\") " pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.719453 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-metrics-certs\") pod \"openstack-operator-controller-manager-86bd8996f6-fxpwr\" (UID: \"151d2334-a121-46f8-947b-3b17a21abcaf\") " pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:22:52 crc kubenswrapper[4950]: E0318 20:22:52.719576 4950 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 18 20:22:52 crc kubenswrapper[4950]: E0318 20:22:52.719638 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-metrics-certs podName:151d2334-a121-46f8-947b-3b17a21abcaf nodeName:}" failed. No retries permitted until 2026-03-18 20:22:53.719605459 +0000 UTC m=+986.960447327 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-metrics-certs") pod "openstack-operator-controller-manager-86bd8996f6-fxpwr" (UID: "151d2334-a121-46f8-947b-3b17a21abcaf") : secret "metrics-server-cert" not found Mar 18 20:22:52 crc kubenswrapper[4950]: E0318 20:22:52.719965 4950 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 18 20:22:52 crc kubenswrapper[4950]: E0318 20:22:52.719989 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-webhook-certs podName:151d2334-a121-46f8-947b-3b17a21abcaf nodeName:}" failed. No retries permitted until 2026-03-18 20:22:53.719981739 +0000 UTC m=+986.960823607 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-webhook-certs") pod "openstack-operator-controller-manager-86bd8996f6-fxpwr" (UID: "151d2334-a121-46f8-947b-3b17a21abcaf") : secret "webhook-server-cert" not found Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.821178 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/df0f2790-244c-4b4f-b3e9-d82e69bb9044-cert\") pod \"infra-operator-controller-manager-7b9c774f96-6cg5j\" (UID: \"df0f2790-244c-4b4f-b3e9-d82e69bb9044\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-6cg5j" Mar 18 20:22:52 crc kubenswrapper[4950]: E0318 20:22:52.821337 4950 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 18 20:22:52 crc kubenswrapper[4950]: E0318 20:22:52.821400 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/df0f2790-244c-4b4f-b3e9-d82e69bb9044-cert podName:df0f2790-244c-4b4f-b3e9-d82e69bb9044 nodeName:}" failed. No retries permitted until 2026-03-18 20:22:54.821384364 +0000 UTC m=+988.062226232 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/df0f2790-244c-4b4f-b3e9-d82e69bb9044-cert") pod "infra-operator-controller-manager-7b9c774f96-6cg5j" (UID: "df0f2790-244c-4b4f-b3e9-d82e69bb9044") : secret "infra-operator-webhook-server-cert" not found Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.924990 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-h9vll"] Mar 18 20:22:52 crc kubenswrapper[4950]: I0318 20:22:52.942072 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-xlx8b"] Mar 18 20:22:52 crc kubenswrapper[4950]: W0318 20:22:52.945859 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8081f4de_9137_49d2_a66d_bbf265033c5c.slice/crio-b8ab30e88da23ac5ed588da81045254cbafe9d38d1c202b26df408cd5f17ee34 WatchSource:0}: Error finding container b8ab30e88da23ac5ed588da81045254cbafe9d38d1c202b26df408cd5f17ee34: Status 404 returned error can't find the container with id b8ab30e88da23ac5ed588da81045254cbafe9d38d1c202b26df408cd5f17ee34 Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.033947 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-dls6n"] Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.042611 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-6jt4g"] Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.049959 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-5v45z"] Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.062306 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-ftljz"] Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.076780 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-xlx8b" event={"ID":"19c17169-bafd-4a5d-b255-0c50248f3e1c","Type":"ContainerStarted","Data":"8f90797975d40ff0d57d1823a06e66045aa7cb2c37deee9cf667eaa73c375111"} Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.079348 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-5v45z" event={"ID":"926aae26-bb2f-4cac-af12-abbd3585a90a","Type":"ContainerStarted","Data":"3548289a18102e5347fc4ae39a8138506fd5109c07c48041012ca2575e54eba4"} Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.085455 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhmk7" event={"ID":"3792146f-d91d-436e-b8af-fb90133a58b4","Type":"ContainerDied","Data":"2504830a9b2ba50185ea53885618ee884e670af1d63a7a650df1d383dea3be4f"} Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.085535 4950 scope.go:117] "RemoveContainer" containerID="08187b42c020f14f29afd1d0d41d8edafb2da9d9b65156d24698b12b454e4e9f" Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.085728 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mhmk7" Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.088461 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-6jt4g" event={"ID":"9046fce8-e07a-4cc6-9577-4ab38ecac0f3","Type":"ContainerStarted","Data":"48e87c003f72ab2950746a6664f954cf52860ff0db23fcfc1daeda4598959fbc"} Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.090834 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-55f864c847-ftljz" event={"ID":"edc5286e-aa7d-4ee6-a8f7-3f5cad640484","Type":"ContainerStarted","Data":"401fb6afd35dee007f460dbc2f949d39c7a21c0503b18e8cedd832c5edf3aa93"} Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.091767 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-h9vll" event={"ID":"8081f4de-9137-49d2-a66d-bbf265033c5c","Type":"ContainerStarted","Data":"b8ab30e88da23ac5ed588da81045254cbafe9d38d1c202b26df408cd5f17ee34"} Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.093572 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-rgb5v" event={"ID":"bc63ecc0-0c62-4607-8ce4-8d143f0c2ce6","Type":"ContainerStarted","Data":"62e7bb0b977aec92de7616775b62b3478dae4fdf437c5366ae050a0fcbcf270c"} Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.097830 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-dls6n" event={"ID":"516ce0ee-83c3-40a3-bb00-2a88891a5791","Type":"ContainerStarted","Data":"d544ffd8e3b419b5d48dc17a2fb08f072026918ef3e8dd6f1dbdb43bc15c2baf"} Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.112911 4950 scope.go:117] "RemoveContainer" containerID="789e43e181d8ff9497f52c7eeae267bd23ff9a3faf0f78beb05885b54577c322" Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.128562 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5a502f8b-719d-4f52-812d-c559a8772461-cert\") pod \"openstack-baremetal-operator-controller-manager-74c4796899rqqwt\" (UID: \"5a502f8b-719d-4f52-812d-c559a8772461\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-74c4796899rqqwt" Mar 18 20:22:53 crc kubenswrapper[4950]: E0318 20:22:53.128999 4950 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 18 20:22:53 crc kubenswrapper[4950]: E0318 20:22:53.129074 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5a502f8b-719d-4f52-812d-c559a8772461-cert podName:5a502f8b-719d-4f52-812d-c559a8772461 nodeName:}" failed. No retries permitted until 2026-03-18 20:22:55.129056008 +0000 UTC m=+988.369897876 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5a502f8b-719d-4f52-812d-c559a8772461-cert") pod "openstack-baremetal-operator-controller-manager-74c4796899rqqwt" (UID: "5a502f8b-719d-4f52-812d-c559a8772461") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.147642 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mhmk7"] Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.158089 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mhmk7"] Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.160393 4950 scope.go:117] "RemoveContainer" containerID="b52b126695b669833de906e411ff1481eaac9565b5dd67a5176f2ee0160458e0" Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.202661 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-d6t88"] Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.217841 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-kckr9"] Mar 18 20:22:53 crc kubenswrapper[4950]: W0318 20:22:53.221033 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf216b689_dfc9_4042_92f3_71014b3a2698.slice/crio-540b69de755027b9f05f4db4112dc2c1580ad5eecead4e39275a40e0b7c75e2e WatchSource:0}: Error finding container 540b69de755027b9f05f4db4112dc2c1580ad5eecead4e39275a40e0b7c75e2e: Status 404 returned error can't find the container with id 540b69de755027b9f05f4db4112dc2c1580ad5eecead4e39275a40e0b7c75e2e Mar 18 20:22:53 crc kubenswrapper[4950]: W0318 20:22:53.229095 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0277fe72_c2e8_4f07_ae64_bb1232a98518.slice/crio-75a0993688adcab49d2bc9f2cd769e5a40fead9e9e6eaf372e3e6e4f118011f2 WatchSource:0}: Error finding container 75a0993688adcab49d2bc9f2cd769e5a40fead9e9e6eaf372e3e6e4f118011f2: Status 404 returned error can't find the container with id 75a0993688adcab49d2bc9f2cd769e5a40fead9e9e6eaf372e3e6e4f118011f2 Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.239002 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-l2xgb"] Mar 18 20:22:53 crc kubenswrapper[4950]: W0318 20:22:53.251983 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod815a0db2_1d7b_4577_b1bf_cf3114f2a2a3.slice/crio-e89e2b0d2b87d4463f59a238d67b1243e8ee7a4b06d7d43bb840c24695992424 WatchSource:0}: Error finding container e89e2b0d2b87d4463f59a238d67b1243e8ee7a4b06d7d43bb840c24695992424: Status 404 returned error can't find the container with id e89e2b0d2b87d4463f59a238d67b1243e8ee7a4b06d7d43bb840c24695992424 Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.400891 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-wnrmf"] Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.412958 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-qxz4b"] Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.429047 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-v6qqx"] Mar 18 20:22:53 crc kubenswrapper[4950]: W0318 20:22:53.442907 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3196503_22ca_4f3a_ac09_f37171f08d37.slice/crio-0283a230762a44466ce0373bb448882f7662e546907cc7e92ab1fccb9cfc4174 WatchSource:0}: Error finding container 0283a230762a44466ce0373bb448882f7662e546907cc7e92ab1fccb9cfc4174: Status 404 returned error can't find the container with id 0283a230762a44466ce0373bb448882f7662e546907cc7e92ab1fccb9cfc4174 Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.449235 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-5h94g"] Mar 18 20:22:53 crc kubenswrapper[4950]: E0318 20:22:53.471634 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:6e7552996253fc66667eaa3eb0e11b4e97145efa2ae577155ceabf8e9913ddc1,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xf5ml,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-67ccfc9778-5h94g_openstack-operators(80caf1bd-ddfd-498d-9b80-ea8cd37a3c72): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 18 20:22:53 crc kubenswrapper[4950]: E0318 20:22:53.474626 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-5h94g" podUID="80caf1bd-ddfd-498d-9b80-ea8cd37a3c72" Mar 18 20:22:53 crc kubenswrapper[4950]: E0318 20:22:53.497349 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:425fd66675becbe0ca2b2fe1a5a6694ac6e0b1cdce9a77a7a37f99785eadc74a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kspc8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-5b9f45d989-pvw7d_openstack-operators(29d3abf6-9968-4603-9c38-d3c1809b7275): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 18 20:22:53 crc kubenswrapper[4950]: E0318 20:22:53.498855 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-pvw7d" podUID="29d3abf6-9968-4603-9c38-d3c1809b7275" Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.500398 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3792146f-d91d-436e-b8af-fb90133a58b4" path="/var/lib/kubelet/pods/3792146f-d91d-436e-b8af-fb90133a58b4/volumes" Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.500958 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-tkbrk"] Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.500977 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-pvw7d"] Mar 18 20:22:53 crc kubenswrapper[4950]: E0318 20:22:53.506064 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:866844c5b88e1e0518ceb7490cac9d093da3fb8b2f27ba7bd9bd89f946b9ee6e,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xgljg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-c674c5965-tkbrk_openstack-operators(598a0d70-d050-4d4c-8ef5-e274784e5c92): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 18 20:22:53 crc kubenswrapper[4950]: E0318 20:22:53.507387 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-c674c5965-tkbrk" podUID="598a0d70-d050-4d4c-8ef5-e274784e5c92" Mar 18 20:22:53 crc kubenswrapper[4950]: W0318 20:22:53.515256 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod608c505a_eb24_4cd9_aa6d_e07d4d8edc0c.slice/crio-2aca2e4bc7ffa89a250fa3071be68758ca4c758d64c452d509d0568814bde741 WatchSource:0}: Error finding container 2aca2e4bc7ffa89a250fa3071be68758ca4c758d64c452d509d0568814bde741: Status 404 returned error can't find the container with id 2aca2e4bc7ffa89a250fa3071be68758ca4c758d64c452d509d0568814bde741 Mar 18 20:22:53 crc kubenswrapper[4950]: E0318 20:22:53.521918 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5gmfm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-5d488d59fb-bsv4c_openstack-operators(608c505a-eb24-4cd9-aa6d-e07d4d8edc0c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.521988 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-bsv4c"] Mar 18 20:22:53 crc kubenswrapper[4950]: E0318 20:22:53.525540 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-bsv4c" podUID="608c505a-eb24-4cd9-aa6d-e07d4d8edc0c" Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.530322 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-8k97x"] Mar 18 20:22:53 crc kubenswrapper[4950]: E0318 20:22:53.531085 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7rqnw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5784578c99-zcjsh_openstack-operators(6b04874c-d0d9-44e7-8047-b0cc2ba6bdfc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 18 20:22:53 crc kubenswrapper[4950]: E0318 20:22:53.532600 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-5784578c99-zcjsh" podUID="6b04874c-d0d9-44e7-8047-b0cc2ba6bdfc" Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.535653 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-zcjsh"] Mar 18 20:22:53 crc kubenswrapper[4950]: W0318 20:22:53.536325 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8e294f2_ac3e_4dbc_b2eb_cd4d2b947789.slice/crio-cba298bee733cc0b7d06c3591134615ea670d0a6ff54d68cdf3631eb7e3fb43c WatchSource:0}: Error finding container cba298bee733cc0b7d06c3591134615ea670d0a6ff54d68cdf3631eb7e3fb43c: Status 404 returned error can't find the container with id cba298bee733cc0b7d06c3591134615ea670d0a6ff54d68cdf3631eb7e3fb43c Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.539532 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7vkgn"] Mar 18 20:22:53 crc kubenswrapper[4950]: E0318 20:22:53.541842 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5f2sv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-7vkgn_openstack-operators(283bb5c8-8799-4b57-a3c0-3a8b28f3c2d6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 18 20:22:53 crc kubenswrapper[4950]: E0318 20:22:53.543700 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7vkgn" podUID="283bb5c8-8799-4b57-a3c0-3a8b28f3c2d6" Mar 18 20:22:53 crc kubenswrapper[4950]: E0318 20:22:53.547805 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7f7rz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-d6b694c5-8k97x_openstack-operators(b8e294f2-ac3e-4dbc-b2eb-cd4d2b947789): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 18 20:22:53 crc kubenswrapper[4950]: E0318 20:22:53.549860 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-8k97x" podUID="b8e294f2-ac3e-4dbc-b2eb-cd4d2b947789" Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.747042 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-metrics-certs\") pod \"openstack-operator-controller-manager-86bd8996f6-fxpwr\" (UID: \"151d2334-a121-46f8-947b-3b17a21abcaf\") " pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:22:53 crc kubenswrapper[4950]: E0318 20:22:53.747223 4950 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 18 20:22:53 crc kubenswrapper[4950]: I0318 20:22:53.747241 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-webhook-certs\") pod \"openstack-operator-controller-manager-86bd8996f6-fxpwr\" (UID: \"151d2334-a121-46f8-947b-3b17a21abcaf\") " pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:22:53 crc kubenswrapper[4950]: E0318 20:22:53.747291 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-metrics-certs podName:151d2334-a121-46f8-947b-3b17a21abcaf nodeName:}" failed. No retries permitted until 2026-03-18 20:22:55.747273743 +0000 UTC m=+988.988115611 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-metrics-certs") pod "openstack-operator-controller-manager-86bd8996f6-fxpwr" (UID: "151d2334-a121-46f8-947b-3b17a21abcaf") : secret "metrics-server-cert" not found Mar 18 20:22:53 crc kubenswrapper[4950]: E0318 20:22:53.747375 4950 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 18 20:22:53 crc kubenswrapper[4950]: E0318 20:22:53.747451 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-webhook-certs podName:151d2334-a121-46f8-947b-3b17a21abcaf nodeName:}" failed. No retries permitted until 2026-03-18 20:22:55.747432088 +0000 UTC m=+988.988274026 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-webhook-certs") pod "openstack-operator-controller-manager-86bd8996f6-fxpwr" (UID: "151d2334-a121-46f8-947b-3b17a21abcaf") : secret "webhook-server-cert" not found Mar 18 20:22:54 crc kubenswrapper[4950]: I0318 20:22:54.142494 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-wnrmf" event={"ID":"e4b560c4-a26f-4acf-bdc3-bb9cabbbd982","Type":"ContainerStarted","Data":"f0659707fb5cde8f56c957cf83313c7f2868057539baf15f89ff764ea9f55fe1"} Mar 18 20:22:54 crc kubenswrapper[4950]: I0318 20:22:54.152032 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-pvw7d" event={"ID":"29d3abf6-9968-4603-9c38-d3c1809b7275","Type":"ContainerStarted","Data":"a1527b8ca1da9c2249266a1b5ba112b0babfb2692f95deb85d1a913538d08f31"} Mar 18 20:22:54 crc kubenswrapper[4950]: I0318 20:22:54.153024 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-l2xgb" event={"ID":"815a0db2-1d7b-4577-b1bf-cf3114f2a2a3","Type":"ContainerStarted","Data":"e89e2b0d2b87d4463f59a238d67b1243e8ee7a4b06d7d43bb840c24695992424"} Mar 18 20:22:54 crc kubenswrapper[4950]: E0318 20:22:54.153891 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:425fd66675becbe0ca2b2fe1a5a6694ac6e0b1cdce9a77a7a37f99785eadc74a\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-pvw7d" podUID="29d3abf6-9968-4603-9c38-d3c1809b7275" Mar 18 20:22:54 crc kubenswrapper[4950]: I0318 20:22:54.156180 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-kckr9" event={"ID":"0277fe72-c2e8-4f07-ae64-bb1232a98518","Type":"ContainerStarted","Data":"75a0993688adcab49d2bc9f2cd769e5a40fead9e9e6eaf372e3e6e4f118011f2"} Mar 18 20:22:54 crc kubenswrapper[4950]: I0318 20:22:54.157756 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-c674c5965-tkbrk" event={"ID":"598a0d70-d050-4d4c-8ef5-e274784e5c92","Type":"ContainerStarted","Data":"132ea8b7b12e578bd8caa62c9475dea890ff565c4387816bf0ccbeb70ba5a5d7"} Mar 18 20:22:54 crc kubenswrapper[4950]: E0318 20:22:54.159774 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:866844c5b88e1e0518ceb7490cac9d093da3fb8b2f27ba7bd9bd89f946b9ee6e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-c674c5965-tkbrk" podUID="598a0d70-d050-4d4c-8ef5-e274784e5c92" Mar 18 20:22:54 crc kubenswrapper[4950]: I0318 20:22:54.172775 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-8k97x" event={"ID":"b8e294f2-ac3e-4dbc-b2eb-cd4d2b947789","Type":"ContainerStarted","Data":"cba298bee733cc0b7d06c3591134615ea670d0a6ff54d68cdf3631eb7e3fb43c"} Mar 18 20:22:54 crc kubenswrapper[4950]: I0318 20:22:54.174682 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-5h94g" event={"ID":"80caf1bd-ddfd-498d-9b80-ea8cd37a3c72","Type":"ContainerStarted","Data":"1dedc8385f42b0474f50b781040bd1c3f3af156533d3a3f5d42eaaa1503ebb29"} Mar 18 20:22:54 crc kubenswrapper[4950]: E0318 20:22:54.174853 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-8k97x" podUID="b8e294f2-ac3e-4dbc-b2eb-cd4d2b947789" Mar 18 20:22:54 crc kubenswrapper[4950]: I0318 20:22:54.177212 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-767865f676-qxz4b" event={"ID":"ec5a5aa1-65cc-4bb2-8644-fa0c7bb44da4","Type":"ContainerStarted","Data":"4776111e4addbecd370f111baccc8b9b8b6ae476a0048528504c0a6c55dbf1e1"} Mar 18 20:22:54 crc kubenswrapper[4950]: E0318 20:22:54.179907 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:6e7552996253fc66667eaa3eb0e11b4e97145efa2ae577155ceabf8e9913ddc1\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-5h94g" podUID="80caf1bd-ddfd-498d-9b80-ea8cd37a3c72" Mar 18 20:22:54 crc kubenswrapper[4950]: I0318 20:22:54.179903 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5784578c99-zcjsh" event={"ID":"6b04874c-d0d9-44e7-8047-b0cc2ba6bdfc","Type":"ContainerStarted","Data":"e1c81e7ff8f582776a0ed69a86a13fc7d737e675edbf5f6c0c5f92fdb208ad2d"} Mar 18 20:22:54 crc kubenswrapper[4950]: I0318 20:22:54.183098 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-v6qqx" event={"ID":"e3196503-22ca-4f3a-ac09-f37171f08d37","Type":"ContainerStarted","Data":"0283a230762a44466ce0373bb448882f7662e546907cc7e92ab1fccb9cfc4174"} Mar 18 20:22:54 crc kubenswrapper[4950]: E0318 20:22:54.183833 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5784578c99-zcjsh" podUID="6b04874c-d0d9-44e7-8047-b0cc2ba6bdfc" Mar 18 20:22:54 crc kubenswrapper[4950]: I0318 20:22:54.185321 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7vkgn" event={"ID":"283bb5c8-8799-4b57-a3c0-3a8b28f3c2d6","Type":"ContainerStarted","Data":"485eee6067de00481ec7a3ee63ed58c2e4b1d1eb6d37a1c668d0119fc4b2f45b"} Mar 18 20:22:54 crc kubenswrapper[4950]: E0318 20:22:54.191167 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7vkgn" podUID="283bb5c8-8799-4b57-a3c0-3a8b28f3c2d6" Mar 18 20:22:54 crc kubenswrapper[4950]: I0318 20:22:54.193851 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-bsv4c" event={"ID":"608c505a-eb24-4cd9-aa6d-e07d4d8edc0c","Type":"ContainerStarted","Data":"2aca2e4bc7ffa89a250fa3071be68758ca4c758d64c452d509d0568814bde741"} Mar 18 20:22:54 crc kubenswrapper[4950]: E0318 20:22:54.196159 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a\\\"\"" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-bsv4c" podUID="608c505a-eb24-4cd9-aa6d-e07d4d8edc0c" Mar 18 20:22:54 crc kubenswrapper[4950]: I0318 20:22:54.201881 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-884679f54-d6t88" event={"ID":"f216b689-dfc9-4042-92f3-71014b3a2698","Type":"ContainerStarted","Data":"540b69de755027b9f05f4db4112dc2c1580ad5eecead4e39275a40e0b7c75e2e"} Mar 18 20:22:54 crc kubenswrapper[4950]: I0318 20:22:54.862160 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/df0f2790-244c-4b4f-b3e9-d82e69bb9044-cert\") pod \"infra-operator-controller-manager-7b9c774f96-6cg5j\" (UID: \"df0f2790-244c-4b4f-b3e9-d82e69bb9044\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-6cg5j" Mar 18 20:22:54 crc kubenswrapper[4950]: E0318 20:22:54.862357 4950 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 18 20:22:54 crc kubenswrapper[4950]: E0318 20:22:54.862461 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/df0f2790-244c-4b4f-b3e9-d82e69bb9044-cert podName:df0f2790-244c-4b4f-b3e9-d82e69bb9044 nodeName:}" failed. No retries permitted until 2026-03-18 20:22:58.862441257 +0000 UTC m=+992.103283125 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/df0f2790-244c-4b4f-b3e9-d82e69bb9044-cert") pod "infra-operator-controller-manager-7b9c774f96-6cg5j" (UID: "df0f2790-244c-4b4f-b3e9-d82e69bb9044") : secret "infra-operator-webhook-server-cert" not found Mar 18 20:22:55 crc kubenswrapper[4950]: E0318 20:22:55.166395 4950 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 18 20:22:55 crc kubenswrapper[4950]: E0318 20:22:55.166469 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5a502f8b-719d-4f52-812d-c559a8772461-cert podName:5a502f8b-719d-4f52-812d-c559a8772461 nodeName:}" failed. No retries permitted until 2026-03-18 20:22:59.166455456 +0000 UTC m=+992.407297324 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5a502f8b-719d-4f52-812d-c559a8772461-cert") pod "openstack-baremetal-operator-controller-manager-74c4796899rqqwt" (UID: "5a502f8b-719d-4f52-812d-c559a8772461") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 18 20:22:55 crc kubenswrapper[4950]: I0318 20:22:55.166300 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5a502f8b-719d-4f52-812d-c559a8772461-cert\") pod \"openstack-baremetal-operator-controller-manager-74c4796899rqqwt\" (UID: \"5a502f8b-719d-4f52-812d-c559a8772461\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-74c4796899rqqwt" Mar 18 20:22:55 crc kubenswrapper[4950]: E0318 20:22:55.218125 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5784578c99-zcjsh" podUID="6b04874c-d0d9-44e7-8047-b0cc2ba6bdfc" Mar 18 20:22:55 crc kubenswrapper[4950]: E0318 20:22:55.218648 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:6e7552996253fc66667eaa3eb0e11b4e97145efa2ae577155ceabf8e9913ddc1\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-5h94g" podUID="80caf1bd-ddfd-498d-9b80-ea8cd37a3c72" Mar 18 20:22:55 crc kubenswrapper[4950]: E0318 20:22:55.218671 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7vkgn" podUID="283bb5c8-8799-4b57-a3c0-3a8b28f3c2d6" Mar 18 20:22:55 crc kubenswrapper[4950]: E0318 20:22:55.219109 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:425fd66675becbe0ca2b2fe1a5a6694ac6e0b1cdce9a77a7a37f99785eadc74a\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-pvw7d" podUID="29d3abf6-9968-4603-9c38-d3c1809b7275" Mar 18 20:22:55 crc kubenswrapper[4950]: E0318 20:22:55.219165 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a\\\"\"" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-bsv4c" podUID="608c505a-eb24-4cd9-aa6d-e07d4d8edc0c" Mar 18 20:22:55 crc kubenswrapper[4950]: E0318 20:22:55.223039 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-8k97x" podUID="b8e294f2-ac3e-4dbc-b2eb-cd4d2b947789" Mar 18 20:22:55 crc kubenswrapper[4950]: E0318 20:22:55.235736 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:866844c5b88e1e0518ceb7490cac9d093da3fb8b2f27ba7bd9bd89f946b9ee6e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-c674c5965-tkbrk" podUID="598a0d70-d050-4d4c-8ef5-e274784e5c92" Mar 18 20:22:55 crc kubenswrapper[4950]: I0318 20:22:55.778764 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-metrics-certs\") pod \"openstack-operator-controller-manager-86bd8996f6-fxpwr\" (UID: \"151d2334-a121-46f8-947b-3b17a21abcaf\") " pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:22:55 crc kubenswrapper[4950]: I0318 20:22:55.778842 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-webhook-certs\") pod \"openstack-operator-controller-manager-86bd8996f6-fxpwr\" (UID: \"151d2334-a121-46f8-947b-3b17a21abcaf\") " pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:22:55 crc kubenswrapper[4950]: E0318 20:22:55.778991 4950 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 18 20:22:55 crc kubenswrapper[4950]: E0318 20:22:55.779038 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-webhook-certs podName:151d2334-a121-46f8-947b-3b17a21abcaf nodeName:}" failed. No retries permitted until 2026-03-18 20:22:59.779024862 +0000 UTC m=+993.019866730 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-webhook-certs") pod "openstack-operator-controller-manager-86bd8996f6-fxpwr" (UID: "151d2334-a121-46f8-947b-3b17a21abcaf") : secret "webhook-server-cert" not found Mar 18 20:22:55 crc kubenswrapper[4950]: E0318 20:22:55.779276 4950 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 18 20:22:55 crc kubenswrapper[4950]: E0318 20:22:55.779342 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-metrics-certs podName:151d2334-a121-46f8-947b-3b17a21abcaf nodeName:}" failed. No retries permitted until 2026-03-18 20:22:59.77932504 +0000 UTC m=+993.020166908 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-metrics-certs") pod "openstack-operator-controller-manager-86bd8996f6-fxpwr" (UID: "151d2334-a121-46f8-947b-3b17a21abcaf") : secret "metrics-server-cert" not found Mar 18 20:22:58 crc kubenswrapper[4950]: I0318 20:22:58.923091 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/df0f2790-244c-4b4f-b3e9-d82e69bb9044-cert\") pod \"infra-operator-controller-manager-7b9c774f96-6cg5j\" (UID: \"df0f2790-244c-4b4f-b3e9-d82e69bb9044\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-6cg5j" Mar 18 20:22:58 crc kubenswrapper[4950]: E0318 20:22:58.923296 4950 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 18 20:22:58 crc kubenswrapper[4950]: E0318 20:22:58.923383 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/df0f2790-244c-4b4f-b3e9-d82e69bb9044-cert podName:df0f2790-244c-4b4f-b3e9-d82e69bb9044 nodeName:}" failed. No retries permitted until 2026-03-18 20:23:06.923360786 +0000 UTC m=+1000.164202654 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/df0f2790-244c-4b4f-b3e9-d82e69bb9044-cert") pod "infra-operator-controller-manager-7b9c774f96-6cg5j" (UID: "df0f2790-244c-4b4f-b3e9-d82e69bb9044") : secret "infra-operator-webhook-server-cert" not found Mar 18 20:22:59 crc kubenswrapper[4950]: I0318 20:22:59.226259 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5a502f8b-719d-4f52-812d-c559a8772461-cert\") pod \"openstack-baremetal-operator-controller-manager-74c4796899rqqwt\" (UID: \"5a502f8b-719d-4f52-812d-c559a8772461\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-74c4796899rqqwt" Mar 18 20:22:59 crc kubenswrapper[4950]: E0318 20:22:59.226382 4950 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 18 20:22:59 crc kubenswrapper[4950]: E0318 20:22:59.226454 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5a502f8b-719d-4f52-812d-c559a8772461-cert podName:5a502f8b-719d-4f52-812d-c559a8772461 nodeName:}" failed. No retries permitted until 2026-03-18 20:23:07.22643446 +0000 UTC m=+1000.467276328 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5a502f8b-719d-4f52-812d-c559a8772461-cert") pod "openstack-baremetal-operator-controller-manager-74c4796899rqqwt" (UID: "5a502f8b-719d-4f52-812d-c559a8772461") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 18 20:22:59 crc kubenswrapper[4950]: I0318 20:22:59.835350 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-metrics-certs\") pod \"openstack-operator-controller-manager-86bd8996f6-fxpwr\" (UID: \"151d2334-a121-46f8-947b-3b17a21abcaf\") " pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:22:59 crc kubenswrapper[4950]: I0318 20:22:59.835856 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-webhook-certs\") pod \"openstack-operator-controller-manager-86bd8996f6-fxpwr\" (UID: \"151d2334-a121-46f8-947b-3b17a21abcaf\") " pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:22:59 crc kubenswrapper[4950]: E0318 20:22:59.835558 4950 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 18 20:22:59 crc kubenswrapper[4950]: E0318 20:22:59.837194 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-metrics-certs podName:151d2334-a121-46f8-947b-3b17a21abcaf nodeName:}" failed. No retries permitted until 2026-03-18 20:23:07.837156347 +0000 UTC m=+1001.077998235 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-metrics-certs") pod "openstack-operator-controller-manager-86bd8996f6-fxpwr" (UID: "151d2334-a121-46f8-947b-3b17a21abcaf") : secret "metrics-server-cert" not found Mar 18 20:22:59 crc kubenswrapper[4950]: E0318 20:22:59.836407 4950 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 18 20:22:59 crc kubenswrapper[4950]: E0318 20:22:59.837332 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-webhook-certs podName:151d2334-a121-46f8-947b-3b17a21abcaf nodeName:}" failed. No retries permitted until 2026-03-18 20:23:07.837299021 +0000 UTC m=+1001.078140959 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-webhook-certs") pod "openstack-operator-controller-manager-86bd8996f6-fxpwr" (UID: "151d2334-a121-46f8-947b-3b17a21abcaf") : secret "webhook-server-cert" not found Mar 18 20:23:06 crc kubenswrapper[4950]: E0318 20:23:06.068044 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:703ad3a2b749bce100f1e2a445312b65dc3b8b45e8c8ba59f311d3f8f3368113" Mar 18 20:23:06 crc kubenswrapper[4950]: E0318 20:23:06.068825 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:703ad3a2b749bce100f1e2a445312b65dc3b8b45e8c8ba59f311d3f8f3368113,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vzw6w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-8464cc45fb-5v45z_openstack-operators(926aae26-bb2f-4cac-af12-abbd3585a90a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 18 20:23:06 crc kubenswrapper[4950]: E0318 20:23:06.070165 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-5v45z" podUID="926aae26-bb2f-4cac-af12-abbd3585a90a" Mar 18 20:23:06 crc kubenswrapper[4950]: E0318 20:23:06.312509 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:703ad3a2b749bce100f1e2a445312b65dc3b8b45e8c8ba59f311d3f8f3368113\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-5v45z" podUID="926aae26-bb2f-4cac-af12-abbd3585a90a" Mar 18 20:23:06 crc kubenswrapper[4950]: E0318 20:23:06.794091 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c6ef5db244d874430a56c3cc9d27662e4bd57cdaa489e1f6059abcacf3aa0900" Mar 18 20:23:06 crc kubenswrapper[4950]: E0318 20:23:06.794256 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c6ef5db244d874430a56c3cc9d27662e4bd57cdaa489e1f6059abcacf3aa0900,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sxkq6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-67dd5f86f5-6jt4g_openstack-operators(9046fce8-e07a-4cc6-9577-4ab38ecac0f3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 18 20:23:06 crc kubenswrapper[4950]: E0318 20:23:06.795458 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-6jt4g" podUID="9046fce8-e07a-4cc6-9577-4ab38ecac0f3" Mar 18 20:23:06 crc kubenswrapper[4950]: I0318 20:23:06.954128 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/df0f2790-244c-4b4f-b3e9-d82e69bb9044-cert\") pod \"infra-operator-controller-manager-7b9c774f96-6cg5j\" (UID: \"df0f2790-244c-4b4f-b3e9-d82e69bb9044\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-6cg5j" Mar 18 20:23:06 crc kubenswrapper[4950]: I0318 20:23:06.960018 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/df0f2790-244c-4b4f-b3e9-d82e69bb9044-cert\") pod \"infra-operator-controller-manager-7b9c774f96-6cg5j\" (UID: \"df0f2790-244c-4b4f-b3e9-d82e69bb9044\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-6cg5j" Mar 18 20:23:06 crc kubenswrapper[4950]: I0318 20:23:06.969667 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-6cg5j" Mar 18 20:23:07 crc kubenswrapper[4950]: I0318 20:23:07.258731 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5a502f8b-719d-4f52-812d-c559a8772461-cert\") pod \"openstack-baremetal-operator-controller-manager-74c4796899rqqwt\" (UID: \"5a502f8b-719d-4f52-812d-c559a8772461\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-74c4796899rqqwt" Mar 18 20:23:07 crc kubenswrapper[4950]: I0318 20:23:07.275853 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5a502f8b-719d-4f52-812d-c559a8772461-cert\") pod \"openstack-baremetal-operator-controller-manager-74c4796899rqqwt\" (UID: \"5a502f8b-719d-4f52-812d-c559a8772461\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-74c4796899rqqwt" Mar 18 20:23:07 crc kubenswrapper[4950]: I0318 20:23:07.316060 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-74c4796899rqqwt" Mar 18 20:23:07 crc kubenswrapper[4950]: E0318 20:23:07.321632 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:c6ef5db244d874430a56c3cc9d27662e4bd57cdaa489e1f6059abcacf3aa0900\\\"\"" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-6jt4g" podUID="9046fce8-e07a-4cc6-9577-4ab38ecac0f3" Mar 18 20:23:07 crc kubenswrapper[4950]: I0318 20:23:07.868060 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-metrics-certs\") pod \"openstack-operator-controller-manager-86bd8996f6-fxpwr\" (UID: \"151d2334-a121-46f8-947b-3b17a21abcaf\") " pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:23:07 crc kubenswrapper[4950]: I0318 20:23:07.868622 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-webhook-certs\") pod \"openstack-operator-controller-manager-86bd8996f6-fxpwr\" (UID: \"151d2334-a121-46f8-947b-3b17a21abcaf\") " pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:23:07 crc kubenswrapper[4950]: E0318 20:23:07.868837 4950 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 18 20:23:07 crc kubenswrapper[4950]: E0318 20:23:07.868932 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-webhook-certs podName:151d2334-a121-46f8-947b-3b17a21abcaf nodeName:}" failed. No retries permitted until 2026-03-18 20:23:23.868909118 +0000 UTC m=+1017.109750986 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-webhook-certs") pod "openstack-operator-controller-manager-86bd8996f6-fxpwr" (UID: "151d2334-a121-46f8-947b-3b17a21abcaf") : secret "webhook-server-cert" not found Mar 18 20:23:07 crc kubenswrapper[4950]: I0318 20:23:07.874661 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-metrics-certs\") pod \"openstack-operator-controller-manager-86bd8996f6-fxpwr\" (UID: \"151d2334-a121-46f8-947b-3b17a21abcaf\") " pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:23:08 crc kubenswrapper[4950]: E0318 20:23:08.020301 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:9dd26bc51e7757d84736528d4988a1f980ad50ccb070aef6fc252e32c5c423a8" Mar 18 20:23:08 crc kubenswrapper[4950]: E0318 20:23:08.020512 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:9dd26bc51e7757d84736528d4988a1f980ad50ccb070aef6fc252e32c5c423a8,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5xf2t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6f787dddc9-kckr9_openstack-operators(0277fe72-c2e8-4f07-ae64-bb1232a98518): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 18 20:23:08 crc kubenswrapper[4950]: E0318 20:23:08.021838 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-kckr9" podUID="0277fe72-c2e8-4f07-ae64-bb1232a98518" Mar 18 20:23:08 crc kubenswrapper[4950]: E0318 20:23:08.326860 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:9dd26bc51e7757d84736528d4988a1f980ad50ccb070aef6fc252e32c5c423a8\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-kckr9" podUID="0277fe72-c2e8-4f07-ae64-bb1232a98518" Mar 18 20:23:08 crc kubenswrapper[4950]: E0318 20:23:08.720494 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:12841b27173f5f1beeb83112e057c8753f4cf411f583fba4f0610fac0f60b7ad" Mar 18 20:23:08 crc kubenswrapper[4950]: E0318 20:23:08.720893 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:12841b27173f5f1beeb83112e057c8753f4cf411f583fba4f0610fac0f60b7ad,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5ctd7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-588d4d986b-h9vll_openstack-operators(8081f4de-9137-49d2-a66d-bbf265033c5c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 18 20:23:08 crc kubenswrapper[4950]: E0318 20:23:08.722916 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-h9vll" podUID="8081f4de-9137-49d2-a66d-bbf265033c5c" Mar 18 20:23:09 crc kubenswrapper[4950]: E0318 20:23:09.338150 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:ec36a9083657587022f8471c9d5a71b87a7895398496e7fc546c73aa1eae4b56" Mar 18 20:23:09 crc kubenswrapper[4950]: E0318 20:23:09.339158 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:ec36a9083657587022f8471c9d5a71b87a7895398496e7fc546c73aa1eae4b56,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-46bfk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-768b96df4c-dls6n_openstack-operators(516ce0ee-83c3-40a3-bb00-2a88891a5791): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 18 20:23:09 crc kubenswrapper[4950]: E0318 20:23:09.340351 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-dls6n" podUID="516ce0ee-83c3-40a3-bb00-2a88891a5791" Mar 18 20:23:09 crc kubenswrapper[4950]: E0318 20:23:09.353964 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:12841b27173f5f1beeb83112e057c8753f4cf411f583fba4f0610fac0f60b7ad\\\"\"" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-h9vll" podUID="8081f4de-9137-49d2-a66d-bbf265033c5c" Mar 18 20:23:09 crc kubenswrapper[4950]: I0318 20:23:09.916953 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7b9c774f96-6cg5j"] Mar 18 20:23:09 crc kubenswrapper[4950]: W0318 20:23:09.964173 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf0f2790_244c_4b4f_b3e9_d82e69bb9044.slice/crio-ad300d0bdad2b2d1606093a839f3c60ecc0b1ac5f94584c0f2dca388fab18326 WatchSource:0}: Error finding container ad300d0bdad2b2d1606093a839f3c60ecc0b1ac5f94584c0f2dca388fab18326: Status 404 returned error can't find the container with id ad300d0bdad2b2d1606093a839f3c60ecc0b1ac5f94584c0f2dca388fab18326 Mar 18 20:23:09 crc kubenswrapper[4950]: I0318 20:23:09.970244 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-74c4796899rqqwt"] Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.357276 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-55f864c847-ftljz" event={"ID":"edc5286e-aa7d-4ee6-a8f7-3f5cad640484","Type":"ContainerStarted","Data":"71b06c9c9b0e685836b1fea4bd8fa63132b51dfd95aae999ac8966e0d71f7f4b"} Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.358302 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-55f864c847-ftljz" Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.385726 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-v6qqx" event={"ID":"e3196503-22ca-4f3a-ac09-f37171f08d37","Type":"ContainerStarted","Data":"c60ff2a3a27caf448d98a808fcaf46de86667c88924e4786518d1dfa2810936f"} Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.386461 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-v6qqx" Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.396284 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-rgb5v" event={"ID":"bc63ecc0-0c62-4607-8ce4-8d143f0c2ce6","Type":"ContainerStarted","Data":"3983a862c2709ddb8e25c796e25be24f15f49eb5fbe65e849ece1968664bbd03"} Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.396507 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-rgb5v" Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.405259 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-55f864c847-ftljz" podStartSLOduration=3.063980185 podStartE2EDuration="19.405239745s" podCreationTimestamp="2026-03-18 20:22:51 +0000 UTC" firstStartedPulling="2026-03-18 20:22:53.056216734 +0000 UTC m=+986.297058602" lastFinishedPulling="2026-03-18 20:23:09.397476294 +0000 UTC m=+1002.638318162" observedRunningTime="2026-03-18 20:23:10.40050504 +0000 UTC m=+1003.641346908" watchObservedRunningTime="2026-03-18 20:23:10.405239745 +0000 UTC m=+1003.646081613" Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.411683 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-wnrmf" event={"ID":"e4b560c4-a26f-4acf-bdc3-bb9cabbbd982","Type":"ContainerStarted","Data":"49cef0d80b3abd0dbbe460cd8777e7a51b095b4b3bc21ff5e1d36e2751a495a6"} Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.411855 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-wnrmf" Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.421403 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-l2xgb" event={"ID":"815a0db2-1d7b-4577-b1bf-cf3114f2a2a3","Type":"ContainerStarted","Data":"f07d168859861b7bbd79b7c78add8cb560f937e9eeef85b869eb45343df57196"} Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.421645 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-l2xgb" Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.422822 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-74c4796899rqqwt" event={"ID":"5a502f8b-719d-4f52-812d-c559a8772461","Type":"ContainerStarted","Data":"7672746ed246c14add0095ab227907ba9b73c820e1830e1f3ea23e60c14c5f37"} Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.428871 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-767865f676-qxz4b" event={"ID":"ec5a5aa1-65cc-4bb2-8644-fa0c7bb44da4","Type":"ContainerStarted","Data":"72197d27ad1c201be98a7a99b4c1964b80c0963ca4c503ffdb869b44f49a1651"} Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.429571 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-767865f676-qxz4b" Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.431299 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-xlx8b" event={"ID":"19c17169-bafd-4a5d-b255-0c50248f3e1c","Type":"ContainerStarted","Data":"0cc484ff49d4c6eca5e4d0b672af5e561a2c41404633295abe5c7281a389fffc"} Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.431645 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-xlx8b" Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.433108 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-884679f54-d6t88" event={"ID":"f216b689-dfc9-4042-92f3-71014b3a2698","Type":"ContainerStarted","Data":"2991f40ff9459880e85c0e74bfd4d7cd8076d7304494675b5a97201ba2f62f83"} Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.433485 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-884679f54-d6t88" Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.446014 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-6cg5j" event={"ID":"df0f2790-244c-4b4f-b3e9-d82e69bb9044","Type":"ContainerStarted","Data":"ad300d0bdad2b2d1606093a839f3c60ecc0b1ac5f94584c0f2dca388fab18326"} Mar 18 20:23:10 crc kubenswrapper[4950]: E0318 20:23:10.447862 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:ec36a9083657587022f8471c9d5a71b87a7895398496e7fc546c73aa1eae4b56\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-dls6n" podUID="516ce0ee-83c3-40a3-bb00-2a88891a5791" Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.465926 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-rgb5v" podStartSLOduration=3.615711896 podStartE2EDuration="20.465906229s" podCreationTimestamp="2026-03-18 20:22:50 +0000 UTC" firstStartedPulling="2026-03-18 20:22:52.503705516 +0000 UTC m=+985.744547384" lastFinishedPulling="2026-03-18 20:23:09.353899849 +0000 UTC m=+1002.594741717" observedRunningTime="2026-03-18 20:23:10.444599319 +0000 UTC m=+1003.685441187" watchObservedRunningTime="2026-03-18 20:23:10.465906229 +0000 UTC m=+1003.706748097" Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.468268 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-v6qqx" podStartSLOduration=3.5260710570000002 podStartE2EDuration="19.468261371s" podCreationTimestamp="2026-03-18 20:22:51 +0000 UTC" firstStartedPulling="2026-03-18 20:22:53.456940474 +0000 UTC m=+986.697782342" lastFinishedPulling="2026-03-18 20:23:09.399130788 +0000 UTC m=+1002.639972656" observedRunningTime="2026-03-18 20:23:10.465314293 +0000 UTC m=+1003.706156171" watchObservedRunningTime="2026-03-18 20:23:10.468261371 +0000 UTC m=+1003.709103239" Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.563788 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-wnrmf" podStartSLOduration=3.63921355 podStartE2EDuration="19.563772781s" podCreationTimestamp="2026-03-18 20:22:51 +0000 UTC" firstStartedPulling="2026-03-18 20:22:53.42939559 +0000 UTC m=+986.670237448" lastFinishedPulling="2026-03-18 20:23:09.353954811 +0000 UTC m=+1002.594796679" observedRunningTime="2026-03-18 20:23:10.500899168 +0000 UTC m=+1003.741741046" watchObservedRunningTime="2026-03-18 20:23:10.563772781 +0000 UTC m=+1003.804614649" Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.564376 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-767865f676-qxz4b" podStartSLOduration=3.603497761 podStartE2EDuration="19.564368476s" podCreationTimestamp="2026-03-18 20:22:51 +0000 UTC" firstStartedPulling="2026-03-18 20:22:53.443003458 +0000 UTC m=+986.683845326" lastFinishedPulling="2026-03-18 20:23:09.403874173 +0000 UTC m=+1002.644716041" observedRunningTime="2026-03-18 20:23:10.560079944 +0000 UTC m=+1003.800921812" watchObservedRunningTime="2026-03-18 20:23:10.564368476 +0000 UTC m=+1003.805210344" Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.595954 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-884679f54-d6t88" podStartSLOduration=3.441789162 podStartE2EDuration="19.595940226s" podCreationTimestamp="2026-03-18 20:22:51 +0000 UTC" firstStartedPulling="2026-03-18 20:22:53.245538849 +0000 UTC m=+986.486380717" lastFinishedPulling="2026-03-18 20:23:09.399689913 +0000 UTC m=+1002.640531781" observedRunningTime="2026-03-18 20:23:10.594813386 +0000 UTC m=+1003.835655254" watchObservedRunningTime="2026-03-18 20:23:10.595940226 +0000 UTC m=+1003.836782094" Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.632309 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-xlx8b" podStartSLOduration=4.234978368 podStartE2EDuration="20.632293001s" podCreationTimestamp="2026-03-18 20:22:50 +0000 UTC" firstStartedPulling="2026-03-18 20:22:52.956572136 +0000 UTC m=+986.197414004" lastFinishedPulling="2026-03-18 20:23:09.353886769 +0000 UTC m=+1002.594728637" observedRunningTime="2026-03-18 20:23:10.631297065 +0000 UTC m=+1003.872138943" watchObservedRunningTime="2026-03-18 20:23:10.632293001 +0000 UTC m=+1003.873134869" Mar 18 20:23:10 crc kubenswrapper[4950]: I0318 20:23:10.690151 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-l2xgb" podStartSLOduration=4.546410182 podStartE2EDuration="20.690136701s" podCreationTimestamp="2026-03-18 20:22:50 +0000 UTC" firstStartedPulling="2026-03-18 20:22:53.260061471 +0000 UTC m=+986.500903329" lastFinishedPulling="2026-03-18 20:23:09.40378799 +0000 UTC m=+1002.644629848" observedRunningTime="2026-03-18 20:23:10.662387012 +0000 UTC m=+1003.903228880" watchObservedRunningTime="2026-03-18 20:23:10.690136701 +0000 UTC m=+1003.930978569" Mar 18 20:23:21 crc kubenswrapper[4950]: I0318 20:23:21.077960 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-xlx8b" Mar 18 20:23:21 crc kubenswrapper[4950]: I0318 20:23:21.122910 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-rgb5v" Mar 18 20:23:21 crc kubenswrapper[4950]: I0318 20:23:21.366275 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-l2xgb" Mar 18 20:23:21 crc kubenswrapper[4950]: I0318 20:23:21.572368 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-55f864c847-ftljz" Mar 18 20:23:21 crc kubenswrapper[4950]: I0318 20:23:21.639576 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-767865f676-qxz4b" Mar 18 20:23:21 crc kubenswrapper[4950]: I0318 20:23:21.688213 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-884679f54-d6t88" Mar 18 20:23:22 crc kubenswrapper[4950]: I0318 20:23:22.113039 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-wnrmf" Mar 18 20:23:22 crc kubenswrapper[4950]: I0318 20:23:22.164010 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-v6qqx" Mar 18 20:23:23 crc kubenswrapper[4950]: I0318 20:23:23.958172 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-webhook-certs\") pod \"openstack-operator-controller-manager-86bd8996f6-fxpwr\" (UID: \"151d2334-a121-46f8-947b-3b17a21abcaf\") " pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:23:23 crc kubenswrapper[4950]: I0318 20:23:23.966248 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/151d2334-a121-46f8-947b-3b17a21abcaf-webhook-certs\") pod \"openstack-operator-controller-manager-86bd8996f6-fxpwr\" (UID: \"151d2334-a121-46f8-947b-3b17a21abcaf\") " pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.072207 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.565043 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-kckr9" event={"ID":"0277fe72-c2e8-4f07-ae64-bb1232a98518","Type":"ContainerStarted","Data":"0b20f782319d2239343aac3e5c0ada807720c2331d79d64a5d39565e135c7467"} Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.566188 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-kckr9" Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.567205 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-8k97x" event={"ID":"b8e294f2-ac3e-4dbc-b2eb-cd4d2b947789","Type":"ContainerStarted","Data":"1245b342a0dae8643244584be44519eb65901e611c58524eaaba40be51b0ece9"} Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.567531 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-8k97x" Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.568430 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-5h94g" event={"ID":"80caf1bd-ddfd-498d-9b80-ea8cd37a3c72","Type":"ContainerStarted","Data":"14acd4761774c2da16dd403f93429ee33405df0dc37162ae1d635746a5567ddb"} Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.568742 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-5h94g" Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.569632 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-dls6n" event={"ID":"516ce0ee-83c3-40a3-bb00-2a88891a5791","Type":"ContainerStarted","Data":"d42430cd746bcd5744ae2fe0574d31cfb7024fdf5c151a751f283213ef0cb402"} Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.569936 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-dls6n" Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.570788 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5784578c99-zcjsh" event={"ID":"6b04874c-d0d9-44e7-8047-b0cc2ba6bdfc","Type":"ContainerStarted","Data":"e9027b25274f62f98f30ff40b25cdde377f269954e5661a786ccde8c91b7b143"} Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.571092 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5784578c99-zcjsh" Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.576883 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-c674c5965-tkbrk" event={"ID":"598a0d70-d050-4d4c-8ef5-e274784e5c92","Type":"ContainerStarted","Data":"f8cee283d20894bd882b31796250f4d8512a86c478976cdcbbcd4a941ad8cc2d"} Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.577248 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-c674c5965-tkbrk" Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.578232 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7vkgn" event={"ID":"283bb5c8-8799-4b57-a3c0-3a8b28f3c2d6","Type":"ContainerStarted","Data":"3f20c0ad2d450c38afec89e13bf66b453e30508aee08729e9e0c5b1e6a6034b6"} Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.579698 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-5v45z" event={"ID":"926aae26-bb2f-4cac-af12-abbd3585a90a","Type":"ContainerStarted","Data":"dac290228d4ec5389268c53e124807416a0178f081f9b9b59f211f2ff7692975"} Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.580019 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-5v45z" Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.581086 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-6cg5j" event={"ID":"df0f2790-244c-4b4f-b3e9-d82e69bb9044","Type":"ContainerStarted","Data":"c7252b09a7dc964c2f6fa38f7c1ffdcace0230d83ebfc8622d28d4efbe9a6172"} Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.581434 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-6cg5j" Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.586562 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-h9vll" event={"ID":"8081f4de-9137-49d2-a66d-bbf265033c5c","Type":"ContainerStarted","Data":"240b2f15bd22dc1387307bce6380f6b89a21d574c50cc88b83bc5816c6f8c0c1"} Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.586957 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-h9vll" Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.588040 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-bsv4c" event={"ID":"608c505a-eb24-4cd9-aa6d-e07d4d8edc0c","Type":"ContainerStarted","Data":"0517b161f709a4f2346547d9cbf2918ef4509d916e7f60f9f7cb0befbabc5efb"} Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.588357 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-bsv4c" Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.589738 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-pvw7d" event={"ID":"29d3abf6-9968-4603-9c38-d3c1809b7275","Type":"ContainerStarted","Data":"edee227f56258481dab0169e8171bdd870c2f93d282ab0a78246f36961a25692"} Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.596600 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-pvw7d" Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.627921 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-6jt4g" event={"ID":"9046fce8-e07a-4cc6-9577-4ab38ecac0f3","Type":"ContainerStarted","Data":"a84b1456aa8f5bdfea5ccc9005d3a6331f49ac93aae1f998855427d8f44af97c"} Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.628147 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-6jt4g" Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.631327 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-74c4796899rqqwt" event={"ID":"5a502f8b-719d-4f52-812d-c559a8772461","Type":"ContainerStarted","Data":"f236452c57ee7d7c0ff26ef46b54c2e00b933934438d2addcd2e19c8f0c3274e"} Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.631872 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-74c4796899rqqwt" Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.661705 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-kckr9" podStartSLOduration=4.640938767 podStartE2EDuration="34.661686545s" podCreationTimestamp="2026-03-18 20:22:50 +0000 UTC" firstStartedPulling="2026-03-18 20:22:53.245794566 +0000 UTC m=+986.486636434" lastFinishedPulling="2026-03-18 20:23:23.266542334 +0000 UTC m=+1016.507384212" observedRunningTime="2026-03-18 20:23:24.625274558 +0000 UTC m=+1017.866116426" watchObservedRunningTime="2026-03-18 20:23:24.661686545 +0000 UTC m=+1017.902528413" Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.693704 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr"] Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.849208 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-5h94g" podStartSLOduration=4.173761287 podStartE2EDuration="33.84917303s" podCreationTimestamp="2026-03-18 20:22:51 +0000 UTC" firstStartedPulling="2026-03-18 20:22:53.471490647 +0000 UTC m=+986.712332515" lastFinishedPulling="2026-03-18 20:23:23.14690238 +0000 UTC m=+1016.387744258" observedRunningTime="2026-03-18 20:23:24.754669997 +0000 UTC m=+1017.995511865" watchObservedRunningTime="2026-03-18 20:23:24.84917303 +0000 UTC m=+1018.090014898" Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.853909 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5784578c99-zcjsh" podStartSLOduration=4.216047698 podStartE2EDuration="33.853897955s" podCreationTimestamp="2026-03-18 20:22:51 +0000 UTC" firstStartedPulling="2026-03-18 20:22:53.530962839 +0000 UTC m=+986.771804707" lastFinishedPulling="2026-03-18 20:23:23.168813086 +0000 UTC m=+1016.409654964" observedRunningTime="2026-03-18 20:23:24.841352795 +0000 UTC m=+1018.082194663" watchObservedRunningTime="2026-03-18 20:23:24.853897955 +0000 UTC m=+1018.094739823" Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.894944 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7vkgn" podStartSLOduration=4.189332786 podStartE2EDuration="33.894929933s" podCreationTimestamp="2026-03-18 20:22:51 +0000 UTC" firstStartedPulling="2026-03-18 20:22:53.541739373 +0000 UTC m=+986.782581241" lastFinishedPulling="2026-03-18 20:23:23.24733652 +0000 UTC m=+1016.488178388" observedRunningTime="2026-03-18 20:23:24.893043623 +0000 UTC m=+1018.133885491" watchObservedRunningTime="2026-03-18 20:23:24.894929933 +0000 UTC m=+1018.135771801" Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.952628 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-bsv4c" podStartSLOduration=4.210182924 podStartE2EDuration="33.952611179s" podCreationTimestamp="2026-03-18 20:22:51 +0000 UTC" firstStartedPulling="2026-03-18 20:22:53.521793808 +0000 UTC m=+986.762635676" lastFinishedPulling="2026-03-18 20:23:23.264222053 +0000 UTC m=+1016.505063931" observedRunningTime="2026-03-18 20:23:24.951162371 +0000 UTC m=+1018.192004239" watchObservedRunningTime="2026-03-18 20:23:24.952611179 +0000 UTC m=+1018.193453037" Mar 18 20:23:24 crc kubenswrapper[4950]: I0318 20:23:24.983706 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-5v45z" podStartSLOduration=4.800267292 podStartE2EDuration="34.983690895s" podCreationTimestamp="2026-03-18 20:22:50 +0000 UTC" firstStartedPulling="2026-03-18 20:22:53.058540555 +0000 UTC m=+986.299382423" lastFinishedPulling="2026-03-18 20:23:23.241964128 +0000 UTC m=+1016.482806026" observedRunningTime="2026-03-18 20:23:24.979778442 +0000 UTC m=+1018.220620310" watchObservedRunningTime="2026-03-18 20:23:24.983690895 +0000 UTC m=+1018.224532763" Mar 18 20:23:25 crc kubenswrapper[4950]: I0318 20:23:25.061903 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-6cg5j" podStartSLOduration=21.775957494 podStartE2EDuration="35.06188883s" podCreationTimestamp="2026-03-18 20:22:50 +0000 UTC" firstStartedPulling="2026-03-18 20:23:09.979320644 +0000 UTC m=+1003.220162512" lastFinishedPulling="2026-03-18 20:23:23.26525197 +0000 UTC m=+1016.506093848" observedRunningTime="2026-03-18 20:23:25.018251303 +0000 UTC m=+1018.259093171" watchObservedRunningTime="2026-03-18 20:23:25.06188883 +0000 UTC m=+1018.302730698" Mar 18 20:23:25 crc kubenswrapper[4950]: I0318 20:23:25.102712 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-8k97x" podStartSLOduration=4.502163416 podStartE2EDuration="34.102697562s" podCreationTimestamp="2026-03-18 20:22:51 +0000 UTC" firstStartedPulling="2026-03-18 20:22:53.54771924 +0000 UTC m=+986.788561108" lastFinishedPulling="2026-03-18 20:23:23.148253346 +0000 UTC m=+1016.389095254" observedRunningTime="2026-03-18 20:23:25.097554707 +0000 UTC m=+1018.338396575" watchObservedRunningTime="2026-03-18 20:23:25.102697562 +0000 UTC m=+1018.343539430" Mar 18 20:23:25 crc kubenswrapper[4950]: I0318 20:23:25.103009 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-dls6n" podStartSLOduration=4.180212098 podStartE2EDuration="35.10300518s" podCreationTimestamp="2026-03-18 20:22:50 +0000 UTC" firstStartedPulling="2026-03-18 20:22:53.055608008 +0000 UTC m=+986.296449886" lastFinishedPulling="2026-03-18 20:23:23.9784011 +0000 UTC m=+1017.219242968" observedRunningTime="2026-03-18 20:23:25.065716061 +0000 UTC m=+1018.306557929" watchObservedRunningTime="2026-03-18 20:23:25.10300518 +0000 UTC m=+1018.343847048" Mar 18 20:23:25 crc kubenswrapper[4950]: I0318 20:23:25.175367 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-c674c5965-tkbrk" podStartSLOduration=4.450635612 podStartE2EDuration="34.175350072s" podCreationTimestamp="2026-03-18 20:22:51 +0000 UTC" firstStartedPulling="2026-03-18 20:22:53.505961282 +0000 UTC m=+986.746803150" lastFinishedPulling="2026-03-18 20:23:23.230675732 +0000 UTC m=+1016.471517610" observedRunningTime="2026-03-18 20:23:25.172397114 +0000 UTC m=+1018.413238982" watchObservedRunningTime="2026-03-18 20:23:25.175350072 +0000 UTC m=+1018.416191940" Mar 18 20:23:25 crc kubenswrapper[4950]: I0318 20:23:25.177366 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-pvw7d" podStartSLOduration=6.440832009 podStartE2EDuration="34.177361194s" podCreationTimestamp="2026-03-18 20:22:51 +0000 UTC" firstStartedPulling="2026-03-18 20:22:53.497205802 +0000 UTC m=+986.738047670" lastFinishedPulling="2026-03-18 20:23:21.233734977 +0000 UTC m=+1014.474576855" observedRunningTime="2026-03-18 20:23:25.140059564 +0000 UTC m=+1018.380901432" watchObservedRunningTime="2026-03-18 20:23:25.177361194 +0000 UTC m=+1018.418203062" Mar 18 20:23:25 crc kubenswrapper[4950]: I0318 20:23:25.199249 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-h9vll" podStartSLOduration=4.906316218 podStartE2EDuration="35.199233869s" podCreationTimestamp="2026-03-18 20:22:50 +0000 UTC" firstStartedPulling="2026-03-18 20:22:52.956627567 +0000 UTC m=+986.197469435" lastFinishedPulling="2026-03-18 20:23:23.249545178 +0000 UTC m=+1016.490387086" observedRunningTime="2026-03-18 20:23:25.197069122 +0000 UTC m=+1018.437910990" watchObservedRunningTime="2026-03-18 20:23:25.199233869 +0000 UTC m=+1018.440075737" Mar 18 20:23:25 crc kubenswrapper[4950]: I0318 20:23:25.300348 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-6jt4g" podStartSLOduration=5.106649123 podStartE2EDuration="35.300333306s" podCreationTimestamp="2026-03-18 20:22:50 +0000 UTC" firstStartedPulling="2026-03-18 20:22:53.056545413 +0000 UTC m=+986.297387281" lastFinishedPulling="2026-03-18 20:23:23.250229586 +0000 UTC m=+1016.491071464" observedRunningTime="2026-03-18 20:23:25.297005048 +0000 UTC m=+1018.537846916" watchObservedRunningTime="2026-03-18 20:23:25.300333306 +0000 UTC m=+1018.541175174" Mar 18 20:23:25 crc kubenswrapper[4950]: I0318 20:23:25.303159 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-74c4796899rqqwt" podStartSLOduration=21.023837297 podStartE2EDuration="34.30315367s" podCreationTimestamp="2026-03-18 20:22:51 +0000 UTC" firstStartedPulling="2026-03-18 20:23:09.986616785 +0000 UTC m=+1003.227458653" lastFinishedPulling="2026-03-18 20:23:23.265933148 +0000 UTC m=+1016.506775026" observedRunningTime="2026-03-18 20:23:25.250535027 +0000 UTC m=+1018.491376895" watchObservedRunningTime="2026-03-18 20:23:25.30315367 +0000 UTC m=+1018.543995538" Mar 18 20:23:25 crc kubenswrapper[4950]: I0318 20:23:25.639215 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" event={"ID":"151d2334-a121-46f8-947b-3b17a21abcaf","Type":"ContainerStarted","Data":"1d17400f87f4bc7cea185fd7de5cace863ec79942da40f3ded11ceec814a1aac"} Mar 18 20:23:25 crc kubenswrapper[4950]: I0318 20:23:25.639252 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" event={"ID":"151d2334-a121-46f8-947b-3b17a21abcaf","Type":"ContainerStarted","Data":"e2cc90f18e90af18cab61d402c1b4ba23ab59a9cdeb27b817224269c982c7f49"} Mar 18 20:23:25 crc kubenswrapper[4950]: I0318 20:23:25.642747 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:23:25 crc kubenswrapper[4950]: I0318 20:23:25.675515 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" podStartSLOduration=34.675486384 podStartE2EDuration="34.675486384s" podCreationTimestamp="2026-03-18 20:22:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:23:25.672254159 +0000 UTC m=+1018.913096027" watchObservedRunningTime="2026-03-18 20:23:25.675486384 +0000 UTC m=+1018.916328252" Mar 18 20:23:31 crc kubenswrapper[4950]: I0318 20:23:31.103212 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-h9vll" Mar 18 20:23:31 crc kubenswrapper[4950]: I0318 20:23:31.389631 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-kckr9" Mar 18 20:23:31 crc kubenswrapper[4950]: I0318 20:23:31.491801 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-6jt4g" Mar 18 20:23:31 crc kubenswrapper[4950]: I0318 20:23:31.492155 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-5v45z" Mar 18 20:23:31 crc kubenswrapper[4950]: I0318 20:23:31.518390 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-dls6n" Mar 18 20:23:31 crc kubenswrapper[4950]: I0318 20:23:31.604270 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-5h94g" Mar 18 20:23:31 crc kubenswrapper[4950]: I0318 20:23:31.703345 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-pvw7d" Mar 18 20:23:31 crc kubenswrapper[4950]: I0318 20:23:31.724803 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-8k97x" Mar 18 20:23:31 crc kubenswrapper[4950]: I0318 20:23:31.815058 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5784578c99-zcjsh" Mar 18 20:23:31 crc kubenswrapper[4950]: I0318 20:23:31.967398 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-bsv4c" Mar 18 20:23:32 crc kubenswrapper[4950]: I0318 20:23:32.138368 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-c674c5965-tkbrk" Mar 18 20:23:34 crc kubenswrapper[4950]: I0318 20:23:34.081907 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-86bd8996f6-fxpwr" Mar 18 20:23:36 crc kubenswrapper[4950]: I0318 20:23:36.980687 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-6cg5j" Mar 18 20:23:37 crc kubenswrapper[4950]: I0318 20:23:37.322791 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-74c4796899rqqwt" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.589670 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5448ff6dc7-l75zc"] Mar 18 20:23:55 crc kubenswrapper[4950]: E0318 20:23:55.590440 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3792146f-d91d-436e-b8af-fb90133a58b4" containerName="extract-utilities" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.590453 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="3792146f-d91d-436e-b8af-fb90133a58b4" containerName="extract-utilities" Mar 18 20:23:55 crc kubenswrapper[4950]: E0318 20:23:55.590466 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3792146f-d91d-436e-b8af-fb90133a58b4" containerName="extract-content" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.590473 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="3792146f-d91d-436e-b8af-fb90133a58b4" containerName="extract-content" Mar 18 20:23:55 crc kubenswrapper[4950]: E0318 20:23:55.590489 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3792146f-d91d-436e-b8af-fb90133a58b4" containerName="registry-server" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.590495 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="3792146f-d91d-436e-b8af-fb90133a58b4" containerName="registry-server" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.590622 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="3792146f-d91d-436e-b8af-fb90133a58b4" containerName="registry-server" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.591263 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5448ff6dc7-l75zc" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.593857 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.594047 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.594306 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.596632 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-bqt85" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.606445 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5448ff6dc7-l75zc"] Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.641215 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64696987c5-nvh4b"] Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.642237 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64696987c5-nvh4b" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.648399 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.661601 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64696987c5-nvh4b"] Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.750648 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkdhs\" (UniqueName: \"kubernetes.io/projected/03bf57fc-e31e-4e47-8ac5-774fdb8cfb50-kube-api-access-rkdhs\") pod \"dnsmasq-dns-64696987c5-nvh4b\" (UID: \"03bf57fc-e31e-4e47-8ac5-774fdb8cfb50\") " pod="openstack/dnsmasq-dns-64696987c5-nvh4b" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.750704 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79c0ef32-5f60-4e97-b2ce-f2bfab2936ea-config\") pod \"dnsmasq-dns-5448ff6dc7-l75zc\" (UID: \"79c0ef32-5f60-4e97-b2ce-f2bfab2936ea\") " pod="openstack/dnsmasq-dns-5448ff6dc7-l75zc" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.750733 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03bf57fc-e31e-4e47-8ac5-774fdb8cfb50-config\") pod \"dnsmasq-dns-64696987c5-nvh4b\" (UID: \"03bf57fc-e31e-4e47-8ac5-774fdb8cfb50\") " pod="openstack/dnsmasq-dns-64696987c5-nvh4b" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.750752 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfjtp\" (UniqueName: \"kubernetes.io/projected/79c0ef32-5f60-4e97-b2ce-f2bfab2936ea-kube-api-access-lfjtp\") pod \"dnsmasq-dns-5448ff6dc7-l75zc\" (UID: \"79c0ef32-5f60-4e97-b2ce-f2bfab2936ea\") " pod="openstack/dnsmasq-dns-5448ff6dc7-l75zc" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.750772 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03bf57fc-e31e-4e47-8ac5-774fdb8cfb50-dns-svc\") pod \"dnsmasq-dns-64696987c5-nvh4b\" (UID: \"03bf57fc-e31e-4e47-8ac5-774fdb8cfb50\") " pod="openstack/dnsmasq-dns-64696987c5-nvh4b" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.852118 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkdhs\" (UniqueName: \"kubernetes.io/projected/03bf57fc-e31e-4e47-8ac5-774fdb8cfb50-kube-api-access-rkdhs\") pod \"dnsmasq-dns-64696987c5-nvh4b\" (UID: \"03bf57fc-e31e-4e47-8ac5-774fdb8cfb50\") " pod="openstack/dnsmasq-dns-64696987c5-nvh4b" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.852169 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79c0ef32-5f60-4e97-b2ce-f2bfab2936ea-config\") pod \"dnsmasq-dns-5448ff6dc7-l75zc\" (UID: \"79c0ef32-5f60-4e97-b2ce-f2bfab2936ea\") " pod="openstack/dnsmasq-dns-5448ff6dc7-l75zc" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.852199 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03bf57fc-e31e-4e47-8ac5-774fdb8cfb50-config\") pod \"dnsmasq-dns-64696987c5-nvh4b\" (UID: \"03bf57fc-e31e-4e47-8ac5-774fdb8cfb50\") " pod="openstack/dnsmasq-dns-64696987c5-nvh4b" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.852581 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfjtp\" (UniqueName: \"kubernetes.io/projected/79c0ef32-5f60-4e97-b2ce-f2bfab2936ea-kube-api-access-lfjtp\") pod \"dnsmasq-dns-5448ff6dc7-l75zc\" (UID: \"79c0ef32-5f60-4e97-b2ce-f2bfab2936ea\") " pod="openstack/dnsmasq-dns-5448ff6dc7-l75zc" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.853005 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03bf57fc-e31e-4e47-8ac5-774fdb8cfb50-dns-svc\") pod \"dnsmasq-dns-64696987c5-nvh4b\" (UID: \"03bf57fc-e31e-4e47-8ac5-774fdb8cfb50\") " pod="openstack/dnsmasq-dns-64696987c5-nvh4b" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.853290 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03bf57fc-e31e-4e47-8ac5-774fdb8cfb50-config\") pod \"dnsmasq-dns-64696987c5-nvh4b\" (UID: \"03bf57fc-e31e-4e47-8ac5-774fdb8cfb50\") " pod="openstack/dnsmasq-dns-64696987c5-nvh4b" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.853341 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79c0ef32-5f60-4e97-b2ce-f2bfab2936ea-config\") pod \"dnsmasq-dns-5448ff6dc7-l75zc\" (UID: \"79c0ef32-5f60-4e97-b2ce-f2bfab2936ea\") " pod="openstack/dnsmasq-dns-5448ff6dc7-l75zc" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.853742 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03bf57fc-e31e-4e47-8ac5-774fdb8cfb50-dns-svc\") pod \"dnsmasq-dns-64696987c5-nvh4b\" (UID: \"03bf57fc-e31e-4e47-8ac5-774fdb8cfb50\") " pod="openstack/dnsmasq-dns-64696987c5-nvh4b" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.877577 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkdhs\" (UniqueName: \"kubernetes.io/projected/03bf57fc-e31e-4e47-8ac5-774fdb8cfb50-kube-api-access-rkdhs\") pod \"dnsmasq-dns-64696987c5-nvh4b\" (UID: \"03bf57fc-e31e-4e47-8ac5-774fdb8cfb50\") " pod="openstack/dnsmasq-dns-64696987c5-nvh4b" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.893781 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfjtp\" (UniqueName: \"kubernetes.io/projected/79c0ef32-5f60-4e97-b2ce-f2bfab2936ea-kube-api-access-lfjtp\") pod \"dnsmasq-dns-5448ff6dc7-l75zc\" (UID: \"79c0ef32-5f60-4e97-b2ce-f2bfab2936ea\") " pod="openstack/dnsmasq-dns-5448ff6dc7-l75zc" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.906863 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5448ff6dc7-l75zc" Mar 18 20:23:55 crc kubenswrapper[4950]: I0318 20:23:55.962696 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64696987c5-nvh4b" Mar 18 20:23:56 crc kubenswrapper[4950]: I0318 20:23:56.233207 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64696987c5-nvh4b"] Mar 18 20:23:56 crc kubenswrapper[4950]: I0318 20:23:56.355647 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5448ff6dc7-l75zc"] Mar 18 20:23:56 crc kubenswrapper[4950]: W0318 20:23:56.363046 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79c0ef32_5f60_4e97_b2ce_f2bfab2936ea.slice/crio-935c313155c8117804dc62fc468d418c8e8bce76e98187be1548b77844296de1 WatchSource:0}: Error finding container 935c313155c8117804dc62fc468d418c8e8bce76e98187be1548b77844296de1: Status 404 returned error can't find the container with id 935c313155c8117804dc62fc468d418c8e8bce76e98187be1548b77844296de1 Mar 18 20:23:56 crc kubenswrapper[4950]: I0318 20:23:56.885024 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64696987c5-nvh4b" event={"ID":"03bf57fc-e31e-4e47-8ac5-774fdb8cfb50","Type":"ContainerStarted","Data":"b5a50b89076cfa0284f61b933bc296e04554cb1d5bf9349c99147042b55786d1"} Mar 18 20:23:56 crc kubenswrapper[4950]: I0318 20:23:56.886037 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5448ff6dc7-l75zc" event={"ID":"79c0ef32-5f60-4e97-b2ce-f2bfab2936ea","Type":"ContainerStarted","Data":"935c313155c8117804dc62fc468d418c8e8bce76e98187be1548b77844296de1"} Mar 18 20:23:58 crc kubenswrapper[4950]: I0318 20:23:58.417222 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5448ff6dc7-l75zc"] Mar 18 20:23:58 crc kubenswrapper[4950]: I0318 20:23:58.450188 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-658f55c9f5-4qhbd"] Mar 18 20:23:58 crc kubenswrapper[4950]: I0318 20:23:58.451517 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658f55c9f5-4qhbd" Mar 18 20:23:58 crc kubenswrapper[4950]: I0318 20:23:58.464048 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-658f55c9f5-4qhbd"] Mar 18 20:23:58 crc kubenswrapper[4950]: I0318 20:23:58.611268 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e5ae697-1746-49b5-ac19-d0b5bf951ce9-config\") pod \"dnsmasq-dns-658f55c9f5-4qhbd\" (UID: \"0e5ae697-1746-49b5-ac19-d0b5bf951ce9\") " pod="openstack/dnsmasq-dns-658f55c9f5-4qhbd" Mar 18 20:23:58 crc kubenswrapper[4950]: I0318 20:23:58.611379 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8tnl\" (UniqueName: \"kubernetes.io/projected/0e5ae697-1746-49b5-ac19-d0b5bf951ce9-kube-api-access-m8tnl\") pod \"dnsmasq-dns-658f55c9f5-4qhbd\" (UID: \"0e5ae697-1746-49b5-ac19-d0b5bf951ce9\") " pod="openstack/dnsmasq-dns-658f55c9f5-4qhbd" Mar 18 20:23:58 crc kubenswrapper[4950]: I0318 20:23:58.611438 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e5ae697-1746-49b5-ac19-d0b5bf951ce9-dns-svc\") pod \"dnsmasq-dns-658f55c9f5-4qhbd\" (UID: \"0e5ae697-1746-49b5-ac19-d0b5bf951ce9\") " pod="openstack/dnsmasq-dns-658f55c9f5-4qhbd" Mar 18 20:23:58 crc kubenswrapper[4950]: I0318 20:23:58.712373 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8tnl\" (UniqueName: \"kubernetes.io/projected/0e5ae697-1746-49b5-ac19-d0b5bf951ce9-kube-api-access-m8tnl\") pod \"dnsmasq-dns-658f55c9f5-4qhbd\" (UID: \"0e5ae697-1746-49b5-ac19-d0b5bf951ce9\") " pod="openstack/dnsmasq-dns-658f55c9f5-4qhbd" Mar 18 20:23:58 crc kubenswrapper[4950]: I0318 20:23:58.712438 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e5ae697-1746-49b5-ac19-d0b5bf951ce9-dns-svc\") pod \"dnsmasq-dns-658f55c9f5-4qhbd\" (UID: \"0e5ae697-1746-49b5-ac19-d0b5bf951ce9\") " pod="openstack/dnsmasq-dns-658f55c9f5-4qhbd" Mar 18 20:23:58 crc kubenswrapper[4950]: I0318 20:23:58.712472 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e5ae697-1746-49b5-ac19-d0b5bf951ce9-config\") pod \"dnsmasq-dns-658f55c9f5-4qhbd\" (UID: \"0e5ae697-1746-49b5-ac19-d0b5bf951ce9\") " pod="openstack/dnsmasq-dns-658f55c9f5-4qhbd" Mar 18 20:23:58 crc kubenswrapper[4950]: I0318 20:23:58.713767 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e5ae697-1746-49b5-ac19-d0b5bf951ce9-config\") pod \"dnsmasq-dns-658f55c9f5-4qhbd\" (UID: \"0e5ae697-1746-49b5-ac19-d0b5bf951ce9\") " pod="openstack/dnsmasq-dns-658f55c9f5-4qhbd" Mar 18 20:23:58 crc kubenswrapper[4950]: I0318 20:23:58.714331 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e5ae697-1746-49b5-ac19-d0b5bf951ce9-dns-svc\") pod \"dnsmasq-dns-658f55c9f5-4qhbd\" (UID: \"0e5ae697-1746-49b5-ac19-d0b5bf951ce9\") " pod="openstack/dnsmasq-dns-658f55c9f5-4qhbd" Mar 18 20:23:58 crc kubenswrapper[4950]: I0318 20:23:58.755739 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8tnl\" (UniqueName: \"kubernetes.io/projected/0e5ae697-1746-49b5-ac19-d0b5bf951ce9-kube-api-access-m8tnl\") pod \"dnsmasq-dns-658f55c9f5-4qhbd\" (UID: \"0e5ae697-1746-49b5-ac19-d0b5bf951ce9\") " pod="openstack/dnsmasq-dns-658f55c9f5-4qhbd" Mar 18 20:23:58 crc kubenswrapper[4950]: I0318 20:23:58.759361 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64696987c5-nvh4b"] Mar 18 20:23:58 crc kubenswrapper[4950]: I0318 20:23:58.774529 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658f55c9f5-4qhbd" Mar 18 20:23:58 crc kubenswrapper[4950]: I0318 20:23:58.778313 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54b5dffb47-9wgpv"] Mar 18 20:23:58 crc kubenswrapper[4950]: I0318 20:23:58.781303 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54b5dffb47-9wgpv" Mar 18 20:23:58 crc kubenswrapper[4950]: I0318 20:23:58.785463 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54b5dffb47-9wgpv"] Mar 18 20:23:58 crc kubenswrapper[4950]: I0318 20:23:58.914478 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf74b193-b765-42b9-883e-951eb9f12a40-dns-svc\") pod \"dnsmasq-dns-54b5dffb47-9wgpv\" (UID: \"cf74b193-b765-42b9-883e-951eb9f12a40\") " pod="openstack/dnsmasq-dns-54b5dffb47-9wgpv" Mar 18 20:23:58 crc kubenswrapper[4950]: I0318 20:23:58.914838 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf74b193-b765-42b9-883e-951eb9f12a40-config\") pod \"dnsmasq-dns-54b5dffb47-9wgpv\" (UID: \"cf74b193-b765-42b9-883e-951eb9f12a40\") " pod="openstack/dnsmasq-dns-54b5dffb47-9wgpv" Mar 18 20:23:58 crc kubenswrapper[4950]: I0318 20:23:58.914869 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8qxq\" (UniqueName: \"kubernetes.io/projected/cf74b193-b765-42b9-883e-951eb9f12a40-kube-api-access-b8qxq\") pod \"dnsmasq-dns-54b5dffb47-9wgpv\" (UID: \"cf74b193-b765-42b9-883e-951eb9f12a40\") " pod="openstack/dnsmasq-dns-54b5dffb47-9wgpv" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.017319 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf74b193-b765-42b9-883e-951eb9f12a40-dns-svc\") pod \"dnsmasq-dns-54b5dffb47-9wgpv\" (UID: \"cf74b193-b765-42b9-883e-951eb9f12a40\") " pod="openstack/dnsmasq-dns-54b5dffb47-9wgpv" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.017401 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf74b193-b765-42b9-883e-951eb9f12a40-config\") pod \"dnsmasq-dns-54b5dffb47-9wgpv\" (UID: \"cf74b193-b765-42b9-883e-951eb9f12a40\") " pod="openstack/dnsmasq-dns-54b5dffb47-9wgpv" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.017444 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8qxq\" (UniqueName: \"kubernetes.io/projected/cf74b193-b765-42b9-883e-951eb9f12a40-kube-api-access-b8qxq\") pod \"dnsmasq-dns-54b5dffb47-9wgpv\" (UID: \"cf74b193-b765-42b9-883e-951eb9f12a40\") " pod="openstack/dnsmasq-dns-54b5dffb47-9wgpv" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.018549 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf74b193-b765-42b9-883e-951eb9f12a40-dns-svc\") pod \"dnsmasq-dns-54b5dffb47-9wgpv\" (UID: \"cf74b193-b765-42b9-883e-951eb9f12a40\") " pod="openstack/dnsmasq-dns-54b5dffb47-9wgpv" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.019224 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf74b193-b765-42b9-883e-951eb9f12a40-config\") pod \"dnsmasq-dns-54b5dffb47-9wgpv\" (UID: \"cf74b193-b765-42b9-883e-951eb9f12a40\") " pod="openstack/dnsmasq-dns-54b5dffb47-9wgpv" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.036884 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8qxq\" (UniqueName: \"kubernetes.io/projected/cf74b193-b765-42b9-883e-951eb9f12a40-kube-api-access-b8qxq\") pod \"dnsmasq-dns-54b5dffb47-9wgpv\" (UID: \"cf74b193-b765-42b9-883e-951eb9f12a40\") " pod="openstack/dnsmasq-dns-54b5dffb47-9wgpv" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.125618 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54b5dffb47-9wgpv" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.398196 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-658f55c9f5-4qhbd"] Mar 18 20:23:59 crc kubenswrapper[4950]: W0318 20:23:59.407180 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e5ae697_1746_49b5_ac19_d0b5bf951ce9.slice/crio-295f99fd18e8102be9c12b437eaca3684d52971979e0973a4bc4dcc1c9dd917e WatchSource:0}: Error finding container 295f99fd18e8102be9c12b437eaca3684d52971979e0973a4bc4dcc1c9dd917e: Status 404 returned error can't find the container with id 295f99fd18e8102be9c12b437eaca3684d52971979e0973a4bc4dcc1c9dd917e Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.620852 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.622135 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.623470 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.623661 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.623808 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.624020 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.624172 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.628141 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-bjdw7" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.629581 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.632008 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.696534 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54b5dffb47-9wgpv"] Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.728728 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.728774 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.728869 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7502f5ea-31ca-4349-bd79-2219bf3903c0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.728900 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtmb7\" (UniqueName: \"kubernetes.io/projected/7502f5ea-31ca-4349-bd79-2219bf3903c0-kube-api-access-gtmb7\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.728923 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7502f5ea-31ca-4349-bd79-2219bf3903c0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.728947 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.728964 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.728998 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7502f5ea-31ca-4349-bd79-2219bf3903c0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.729053 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7502f5ea-31ca-4349-bd79-2219bf3903c0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.729087 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7502f5ea-31ca-4349-bd79-2219bf3903c0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.729119 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.831502 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7502f5ea-31ca-4349-bd79-2219bf3903c0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.835382 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtmb7\" (UniqueName: \"kubernetes.io/projected/7502f5ea-31ca-4349-bd79-2219bf3903c0-kube-api-access-gtmb7\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.835426 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7502f5ea-31ca-4349-bd79-2219bf3903c0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.835467 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.835486 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.835521 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7502f5ea-31ca-4349-bd79-2219bf3903c0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.835567 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7502f5ea-31ca-4349-bd79-2219bf3903c0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.835614 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7502f5ea-31ca-4349-bd79-2219bf3903c0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.835672 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.835728 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.835754 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.850156 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7502f5ea-31ca-4349-bd79-2219bf3903c0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.851222 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7502f5ea-31ca-4349-bd79-2219bf3903c0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.851509 4950 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.853310 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.853609 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.853812 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.855063 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7502f5ea-31ca-4349-bd79-2219bf3903c0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.857224 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7502f5ea-31ca-4349-bd79-2219bf3903c0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.866968 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7502f5ea-31ca-4349-bd79-2219bf3903c0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.870835 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtmb7\" (UniqueName: \"kubernetes.io/projected/7502f5ea-31ca-4349-bd79-2219bf3903c0-kube-api-access-gtmb7\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.878380 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.882176 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.944752 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54b5dffb47-9wgpv" event={"ID":"cf74b193-b765-42b9-883e-951eb9f12a40","Type":"ContainerStarted","Data":"14f3e2b99481903a82f697779da288400af99b1cd657bddc5ce4a6a5b0ff9ef7"} Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.945841 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658f55c9f5-4qhbd" event={"ID":"0e5ae697-1746-49b5-ac19-d0b5bf951ce9","Type":"ContainerStarted","Data":"295f99fd18e8102be9c12b437eaca3684d52971979e0973a4bc4dcc1c9dd917e"} Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.948966 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.958569 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.970955 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.971056 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.979956 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.980146 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.980255 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.980348 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.980473 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.982818 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Mar 18 20:23:59 crc kubenswrapper[4950]: I0318 20:23:59.988629 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-xm9qz" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.135027 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564424-q58sw"] Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.135890 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564424-q58sw" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.138890 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.142833 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.148967 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.162478 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564424-q58sw"] Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.170032 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.170081 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b82f7668-ae1a-4bfe-be27-435cca5df467-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.170101 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5gd4\" (UniqueName: \"kubernetes.io/projected/b82f7668-ae1a-4bfe-be27-435cca5df467-kube-api-access-g5gd4\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.170147 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b82f7668-ae1a-4bfe-be27-435cca5df467-config-data\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.170232 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.170246 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b82f7668-ae1a-4bfe-be27-435cca5df467-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.170261 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.170352 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.170377 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.170400 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b82f7668-ae1a-4bfe-be27-435cca5df467-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.170428 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b82f7668-ae1a-4bfe-be27-435cca5df467-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.271270 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.271316 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.271340 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b82f7668-ae1a-4bfe-be27-435cca5df467-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.271359 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b82f7668-ae1a-4bfe-be27-435cca5df467-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.271386 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.271400 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b82f7668-ae1a-4bfe-be27-435cca5df467-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.271433 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5gd4\" (UniqueName: \"kubernetes.io/projected/b82f7668-ae1a-4bfe-be27-435cca5df467-kube-api-access-g5gd4\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.271459 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hc4pp\" (UniqueName: \"kubernetes.io/projected/2ff7fb3b-ac2f-4718-8c93-65b9a6291a20-kube-api-access-hc4pp\") pod \"auto-csr-approver-29564424-q58sw\" (UID: \"2ff7fb3b-ac2f-4718-8c93-65b9a6291a20\") " pod="openshift-infra/auto-csr-approver-29564424-q58sw" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.271477 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b82f7668-ae1a-4bfe-be27-435cca5df467-config-data\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.271523 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.271540 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.271557 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b82f7668-ae1a-4bfe-be27-435cca5df467-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.271919 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.272979 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b82f7668-ae1a-4bfe-be27-435cca5df467-config-data\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.273988 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b82f7668-ae1a-4bfe-be27-435cca5df467-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.274743 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b82f7668-ae1a-4bfe-be27-435cca5df467-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.278722 4950 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.279806 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.285026 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b82f7668-ae1a-4bfe-be27-435cca5df467-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.299876 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b82f7668-ae1a-4bfe-be27-435cca5df467-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.326880 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.331296 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5gd4\" (UniqueName: \"kubernetes.io/projected/b82f7668-ae1a-4bfe-be27-435cca5df467-kube-api-access-g5gd4\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.338602 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.345653 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.396124 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hc4pp\" (UniqueName: \"kubernetes.io/projected/2ff7fb3b-ac2f-4718-8c93-65b9a6291a20-kube-api-access-hc4pp\") pod \"auto-csr-approver-29564424-q58sw\" (UID: \"2ff7fb3b-ac2f-4718-8c93-65b9a6291a20\") " pod="openshift-infra/auto-csr-approver-29564424-q58sw" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.415210 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hc4pp\" (UniqueName: \"kubernetes.io/projected/2ff7fb3b-ac2f-4718-8c93-65b9a6291a20-kube-api-access-hc4pp\") pod \"auto-csr-approver-29564424-q58sw\" (UID: \"2ff7fb3b-ac2f-4718-8c93-65b9a6291a20\") " pod="openshift-infra/auto-csr-approver-29564424-q58sw" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.458395 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564424-q58sw" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.605508 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.865951 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 18 20:24:00 crc kubenswrapper[4950]: I0318 20:24:00.997102 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7502f5ea-31ca-4349-bd79-2219bf3903c0","Type":"ContainerStarted","Data":"499d73cbee8abb443a1869c77ddd26c53518424c0040f187778794b996492a62"} Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.104756 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.112981 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.122387 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.122859 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-xrvp5" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.123022 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.123191 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.127738 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.175763 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.188923 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.245202 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564424-q58sw"] Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.316242 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.316289 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.316322 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3-kolla-config\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.316343 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5754z\" (UniqueName: \"kubernetes.io/projected/e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3-kube-api-access-5754z\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.316369 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.316430 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.316445 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3-config-data-default\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.316475 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.428707 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.428742 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3-config-data-default\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.428778 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.428808 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.428831 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.428861 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3-kolla-config\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.428879 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5754z\" (UniqueName: \"kubernetes.io/projected/e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3-kube-api-access-5754z\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.428904 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.429154 4950 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.441051 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3-config-data-default\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.443706 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.443887 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3-kolla-config\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.445272 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.445670 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.456981 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.461102 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5754z\" (UniqueName: \"kubernetes.io/projected/e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3-kube-api-access-5754z\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.475811 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3\") " pod="openstack/openstack-galera-0" Mar 18 20:24:01 crc kubenswrapper[4950]: I0318 20:24:01.765730 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.011275 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b82f7668-ae1a-4bfe-be27-435cca5df467","Type":"ContainerStarted","Data":"87ba0f1995bc9ca4f94deefc7d8d1428ea35680dcfdc49bf02fc667ea2a2d3d0"} Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.012596 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564424-q58sw" event={"ID":"2ff7fb3b-ac2f-4718-8c93-65b9a6291a20","Type":"ContainerStarted","Data":"33ec77fd8b0f1ab05517b3a63dcf121a39d0b869160b08140103069ba2ade4b2"} Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.425988 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.427660 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.433166 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.433389 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.433814 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.442734 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-8fhhw" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.463105 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.474083 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.550678 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4c546\" (UniqueName: \"kubernetes.io/projected/aedf1977-d498-4b4f-a92b-f84be888045d-kube-api-access-4c546\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.550720 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.550752 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/aedf1977-d498-4b4f-a92b-f84be888045d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.550823 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/aedf1977-d498-4b4f-a92b-f84be888045d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.550859 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aedf1977-d498-4b4f-a92b-f84be888045d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.550914 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aedf1977-d498-4b4f-a92b-f84be888045d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.550935 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/aedf1977-d498-4b4f-a92b-f84be888045d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.551069 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/aedf1977-d498-4b4f-a92b-f84be888045d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.653688 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aedf1977-d498-4b4f-a92b-f84be888045d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.653734 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aedf1977-d498-4b4f-a92b-f84be888045d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.653768 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/aedf1977-d498-4b4f-a92b-f84be888045d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.653835 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/aedf1977-d498-4b4f-a92b-f84be888045d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.653878 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4c546\" (UniqueName: \"kubernetes.io/projected/aedf1977-d498-4b4f-a92b-f84be888045d-kube-api-access-4c546\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.653906 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.653931 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/aedf1977-d498-4b4f-a92b-f84be888045d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.653947 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/aedf1977-d498-4b4f-a92b-f84be888045d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.654834 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/aedf1977-d498-4b4f-a92b-f84be888045d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.655190 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aedf1977-d498-4b4f-a92b-f84be888045d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.658117 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/aedf1977-d498-4b4f-a92b-f84be888045d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.659629 4950 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.681600 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/aedf1977-d498-4b4f-a92b-f84be888045d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.694893 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/aedf1977-d498-4b4f-a92b-f84be888045d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.695099 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aedf1977-d498-4b4f-a92b-f84be888045d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.705991 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4c546\" (UniqueName: \"kubernetes.io/projected/aedf1977-d498-4b4f-a92b-f84be888045d-kube-api-access-4c546\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.733753 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"aedf1977-d498-4b4f-a92b-f84be888045d\") " pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.741853 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.743793 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.745104 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.746647 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.747247 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-lx7vh" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.748915 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.751538 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.860225 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pkqx\" (UniqueName: \"kubernetes.io/projected/18eb6f19-11f0-4023-93ac-fc03ecdb9486-kube-api-access-9pkqx\") pod \"memcached-0\" (UID: \"18eb6f19-11f0-4023-93ac-fc03ecdb9486\") " pod="openstack/memcached-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.860301 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/18eb6f19-11f0-4023-93ac-fc03ecdb9486-kolla-config\") pod \"memcached-0\" (UID: \"18eb6f19-11f0-4023-93ac-fc03ecdb9486\") " pod="openstack/memcached-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.860325 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18eb6f19-11f0-4023-93ac-fc03ecdb9486-config-data\") pod \"memcached-0\" (UID: \"18eb6f19-11f0-4023-93ac-fc03ecdb9486\") " pod="openstack/memcached-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.860339 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/18eb6f19-11f0-4023-93ac-fc03ecdb9486-memcached-tls-certs\") pod \"memcached-0\" (UID: \"18eb6f19-11f0-4023-93ac-fc03ecdb9486\") " pod="openstack/memcached-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.860356 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18eb6f19-11f0-4023-93ac-fc03ecdb9486-combined-ca-bundle\") pod \"memcached-0\" (UID: \"18eb6f19-11f0-4023-93ac-fc03ecdb9486\") " pod="openstack/memcached-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.964356 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pkqx\" (UniqueName: \"kubernetes.io/projected/18eb6f19-11f0-4023-93ac-fc03ecdb9486-kube-api-access-9pkqx\") pod \"memcached-0\" (UID: \"18eb6f19-11f0-4023-93ac-fc03ecdb9486\") " pod="openstack/memcached-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.964441 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/18eb6f19-11f0-4023-93ac-fc03ecdb9486-kolla-config\") pod \"memcached-0\" (UID: \"18eb6f19-11f0-4023-93ac-fc03ecdb9486\") " pod="openstack/memcached-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.964467 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18eb6f19-11f0-4023-93ac-fc03ecdb9486-config-data\") pod \"memcached-0\" (UID: \"18eb6f19-11f0-4023-93ac-fc03ecdb9486\") " pod="openstack/memcached-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.964481 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/18eb6f19-11f0-4023-93ac-fc03ecdb9486-memcached-tls-certs\") pod \"memcached-0\" (UID: \"18eb6f19-11f0-4023-93ac-fc03ecdb9486\") " pod="openstack/memcached-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.964498 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18eb6f19-11f0-4023-93ac-fc03ecdb9486-combined-ca-bundle\") pod \"memcached-0\" (UID: \"18eb6f19-11f0-4023-93ac-fc03ecdb9486\") " pod="openstack/memcached-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.966683 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/18eb6f19-11f0-4023-93ac-fc03ecdb9486-kolla-config\") pod \"memcached-0\" (UID: \"18eb6f19-11f0-4023-93ac-fc03ecdb9486\") " pod="openstack/memcached-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.968999 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18eb6f19-11f0-4023-93ac-fc03ecdb9486-combined-ca-bundle\") pod \"memcached-0\" (UID: \"18eb6f19-11f0-4023-93ac-fc03ecdb9486\") " pod="openstack/memcached-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.973183 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18eb6f19-11f0-4023-93ac-fc03ecdb9486-config-data\") pod \"memcached-0\" (UID: \"18eb6f19-11f0-4023-93ac-fc03ecdb9486\") " pod="openstack/memcached-0" Mar 18 20:24:02 crc kubenswrapper[4950]: I0318 20:24:02.973895 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/18eb6f19-11f0-4023-93ac-fc03ecdb9486-memcached-tls-certs\") pod \"memcached-0\" (UID: \"18eb6f19-11f0-4023-93ac-fc03ecdb9486\") " pod="openstack/memcached-0" Mar 18 20:24:03 crc kubenswrapper[4950]: I0318 20:24:02.997922 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pkqx\" (UniqueName: \"kubernetes.io/projected/18eb6f19-11f0-4023-93ac-fc03ecdb9486-kube-api-access-9pkqx\") pod \"memcached-0\" (UID: \"18eb6f19-11f0-4023-93ac-fc03ecdb9486\") " pod="openstack/memcached-0" Mar 18 20:24:03 crc kubenswrapper[4950]: I0318 20:24:03.037602 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3","Type":"ContainerStarted","Data":"1af5c62d72a47cc8b8588814ecddeae07696910a5370e1274ef02362fb458b31"} Mar 18 20:24:03 crc kubenswrapper[4950]: I0318 20:24:03.170357 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 18 20:24:03 crc kubenswrapper[4950]: I0318 20:24:03.342694 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 18 20:24:03 crc kubenswrapper[4950]: W0318 20:24:03.359405 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaedf1977_d498_4b4f_a92b_f84be888045d.slice/crio-1ecf5dfbc8bc303708e421b30e856beb03a0f162248280ed23045921f3a7acad WatchSource:0}: Error finding container 1ecf5dfbc8bc303708e421b30e856beb03a0f162248280ed23045921f3a7acad: Status 404 returned error can't find the container with id 1ecf5dfbc8bc303708e421b30e856beb03a0f162248280ed23045921f3a7acad Mar 18 20:24:03 crc kubenswrapper[4950]: I0318 20:24:03.632034 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 18 20:24:03 crc kubenswrapper[4950]: W0318 20:24:03.651623 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18eb6f19_11f0_4023_93ac_fc03ecdb9486.slice/crio-a909e505b1fe08f744362b2274b6b6b9fd0c997d92e573235dd151827e3460a8 WatchSource:0}: Error finding container a909e505b1fe08f744362b2274b6b6b9fd0c997d92e573235dd151827e3460a8: Status 404 returned error can't find the container with id a909e505b1fe08f744362b2274b6b6b9fd0c997d92e573235dd151827e3460a8 Mar 18 20:24:04 crc kubenswrapper[4950]: I0318 20:24:04.071631 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"aedf1977-d498-4b4f-a92b-f84be888045d","Type":"ContainerStarted","Data":"1ecf5dfbc8bc303708e421b30e856beb03a0f162248280ed23045921f3a7acad"} Mar 18 20:24:04 crc kubenswrapper[4950]: I0318 20:24:04.074167 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"18eb6f19-11f0-4023-93ac-fc03ecdb9486","Type":"ContainerStarted","Data":"a909e505b1fe08f744362b2274b6b6b9fd0c997d92e573235dd151827e3460a8"} Mar 18 20:24:04 crc kubenswrapper[4950]: I0318 20:24:04.916459 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 18 20:24:04 crc kubenswrapper[4950]: I0318 20:24:04.917744 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 18 20:24:04 crc kubenswrapper[4950]: I0318 20:24:04.934676 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-5vbz6" Mar 18 20:24:04 crc kubenswrapper[4950]: I0318 20:24:04.945199 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 18 20:24:05 crc kubenswrapper[4950]: I0318 20:24:05.016078 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t8qg\" (UniqueName: \"kubernetes.io/projected/859fa182-cb03-41ab-b127-20b3564cde97-kube-api-access-7t8qg\") pod \"kube-state-metrics-0\" (UID: \"859fa182-cb03-41ab-b127-20b3564cde97\") " pod="openstack/kube-state-metrics-0" Mar 18 20:24:05 crc kubenswrapper[4950]: I0318 20:24:05.098603 4950 generic.go:334] "Generic (PLEG): container finished" podID="2ff7fb3b-ac2f-4718-8c93-65b9a6291a20" containerID="df6c20cda351b2931e39b19a9a5094c1f74accc06feb7285ed7e0985dee7e485" exitCode=0 Mar 18 20:24:05 crc kubenswrapper[4950]: I0318 20:24:05.098644 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564424-q58sw" event={"ID":"2ff7fb3b-ac2f-4718-8c93-65b9a6291a20","Type":"ContainerDied","Data":"df6c20cda351b2931e39b19a9a5094c1f74accc06feb7285ed7e0985dee7e485"} Mar 18 20:24:05 crc kubenswrapper[4950]: I0318 20:24:05.117287 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t8qg\" (UniqueName: \"kubernetes.io/projected/859fa182-cb03-41ab-b127-20b3564cde97-kube-api-access-7t8qg\") pod \"kube-state-metrics-0\" (UID: \"859fa182-cb03-41ab-b127-20b3564cde97\") " pod="openstack/kube-state-metrics-0" Mar 18 20:24:05 crc kubenswrapper[4950]: I0318 20:24:05.141743 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t8qg\" (UniqueName: \"kubernetes.io/projected/859fa182-cb03-41ab-b127-20b3564cde97-kube-api-access-7t8qg\") pod \"kube-state-metrics-0\" (UID: \"859fa182-cb03-41ab-b127-20b3564cde97\") " pod="openstack/kube-state-metrics-0" Mar 18 20:24:05 crc kubenswrapper[4950]: I0318 20:24:05.280513 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 18 20:24:05 crc kubenswrapper[4950]: I0318 20:24:05.820981 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.469274 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zntd5"] Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.474387 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zntd5" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.478757 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zntd5"] Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.484167 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-t8w5c" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.484488 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.484782 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.498739 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-lvf7g"] Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.500150 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.534557 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-lvf7g"] Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.617911 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b72c4ff1-1234-41b7-9faa-b5fd15749084-combined-ca-bundle\") pod \"ovn-controller-zntd5\" (UID: \"b72c4ff1-1234-41b7-9faa-b5fd15749084\") " pod="openstack/ovn-controller-zntd5" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.618176 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b72c4ff1-1234-41b7-9faa-b5fd15749084-var-log-ovn\") pod \"ovn-controller-zntd5\" (UID: \"b72c4ff1-1234-41b7-9faa-b5fd15749084\") " pod="openstack/ovn-controller-zntd5" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.618440 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/21aec953-2e02-48d3-85ff-872d48c4d661-etc-ovs\") pod \"ovn-controller-ovs-lvf7g\" (UID: \"21aec953-2e02-48d3-85ff-872d48c4d661\") " pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.618515 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/b72c4ff1-1234-41b7-9faa-b5fd15749084-ovn-controller-tls-certs\") pod \"ovn-controller-zntd5\" (UID: \"b72c4ff1-1234-41b7-9faa-b5fd15749084\") " pod="openstack/ovn-controller-zntd5" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.618592 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/21aec953-2e02-48d3-85ff-872d48c4d661-var-run\") pod \"ovn-controller-ovs-lvf7g\" (UID: \"21aec953-2e02-48d3-85ff-872d48c4d661\") " pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.618757 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xphf8\" (UniqueName: \"kubernetes.io/projected/b72c4ff1-1234-41b7-9faa-b5fd15749084-kube-api-access-xphf8\") pod \"ovn-controller-zntd5\" (UID: \"b72c4ff1-1234-41b7-9faa-b5fd15749084\") " pod="openstack/ovn-controller-zntd5" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.619115 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b72c4ff1-1234-41b7-9faa-b5fd15749084-var-run\") pod \"ovn-controller-zntd5\" (UID: \"b72c4ff1-1234-41b7-9faa-b5fd15749084\") " pod="openstack/ovn-controller-zntd5" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.619333 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwfm7\" (UniqueName: \"kubernetes.io/projected/21aec953-2e02-48d3-85ff-872d48c4d661-kube-api-access-jwfm7\") pod \"ovn-controller-ovs-lvf7g\" (UID: \"21aec953-2e02-48d3-85ff-872d48c4d661\") " pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.619380 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/21aec953-2e02-48d3-85ff-872d48c4d661-var-lib\") pod \"ovn-controller-ovs-lvf7g\" (UID: \"21aec953-2e02-48d3-85ff-872d48c4d661\") " pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.619507 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/21aec953-2e02-48d3-85ff-872d48c4d661-scripts\") pod \"ovn-controller-ovs-lvf7g\" (UID: \"21aec953-2e02-48d3-85ff-872d48c4d661\") " pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.619536 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b72c4ff1-1234-41b7-9faa-b5fd15749084-scripts\") pod \"ovn-controller-zntd5\" (UID: \"b72c4ff1-1234-41b7-9faa-b5fd15749084\") " pod="openstack/ovn-controller-zntd5" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.619556 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/21aec953-2e02-48d3-85ff-872d48c4d661-var-log\") pod \"ovn-controller-ovs-lvf7g\" (UID: \"21aec953-2e02-48d3-85ff-872d48c4d661\") " pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.619589 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b72c4ff1-1234-41b7-9faa-b5fd15749084-var-run-ovn\") pod \"ovn-controller-zntd5\" (UID: \"b72c4ff1-1234-41b7-9faa-b5fd15749084\") " pod="openstack/ovn-controller-zntd5" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.720937 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwfm7\" (UniqueName: \"kubernetes.io/projected/21aec953-2e02-48d3-85ff-872d48c4d661-kube-api-access-jwfm7\") pod \"ovn-controller-ovs-lvf7g\" (UID: \"21aec953-2e02-48d3-85ff-872d48c4d661\") " pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.720982 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/21aec953-2e02-48d3-85ff-872d48c4d661-var-lib\") pod \"ovn-controller-ovs-lvf7g\" (UID: \"21aec953-2e02-48d3-85ff-872d48c4d661\") " pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.721004 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/21aec953-2e02-48d3-85ff-872d48c4d661-scripts\") pod \"ovn-controller-ovs-lvf7g\" (UID: \"21aec953-2e02-48d3-85ff-872d48c4d661\") " pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.721021 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b72c4ff1-1234-41b7-9faa-b5fd15749084-scripts\") pod \"ovn-controller-zntd5\" (UID: \"b72c4ff1-1234-41b7-9faa-b5fd15749084\") " pod="openstack/ovn-controller-zntd5" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.721038 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/21aec953-2e02-48d3-85ff-872d48c4d661-var-log\") pod \"ovn-controller-ovs-lvf7g\" (UID: \"21aec953-2e02-48d3-85ff-872d48c4d661\") " pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.721052 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b72c4ff1-1234-41b7-9faa-b5fd15749084-var-run-ovn\") pod \"ovn-controller-zntd5\" (UID: \"b72c4ff1-1234-41b7-9faa-b5fd15749084\") " pod="openstack/ovn-controller-zntd5" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.721082 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b72c4ff1-1234-41b7-9faa-b5fd15749084-combined-ca-bundle\") pod \"ovn-controller-zntd5\" (UID: \"b72c4ff1-1234-41b7-9faa-b5fd15749084\") " pod="openstack/ovn-controller-zntd5" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.721101 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b72c4ff1-1234-41b7-9faa-b5fd15749084-var-log-ovn\") pod \"ovn-controller-zntd5\" (UID: \"b72c4ff1-1234-41b7-9faa-b5fd15749084\") " pod="openstack/ovn-controller-zntd5" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.721117 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/21aec953-2e02-48d3-85ff-872d48c4d661-etc-ovs\") pod \"ovn-controller-ovs-lvf7g\" (UID: \"21aec953-2e02-48d3-85ff-872d48c4d661\") " pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.721132 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/b72c4ff1-1234-41b7-9faa-b5fd15749084-ovn-controller-tls-certs\") pod \"ovn-controller-zntd5\" (UID: \"b72c4ff1-1234-41b7-9faa-b5fd15749084\") " pod="openstack/ovn-controller-zntd5" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.721146 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/21aec953-2e02-48d3-85ff-872d48c4d661-var-run\") pod \"ovn-controller-ovs-lvf7g\" (UID: \"21aec953-2e02-48d3-85ff-872d48c4d661\") " pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.721173 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xphf8\" (UniqueName: \"kubernetes.io/projected/b72c4ff1-1234-41b7-9faa-b5fd15749084-kube-api-access-xphf8\") pod \"ovn-controller-zntd5\" (UID: \"b72c4ff1-1234-41b7-9faa-b5fd15749084\") " pod="openstack/ovn-controller-zntd5" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.721205 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b72c4ff1-1234-41b7-9faa-b5fd15749084-var-run\") pod \"ovn-controller-zntd5\" (UID: \"b72c4ff1-1234-41b7-9faa-b5fd15749084\") " pod="openstack/ovn-controller-zntd5" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.721656 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/21aec953-2e02-48d3-85ff-872d48c4d661-etc-ovs\") pod \"ovn-controller-ovs-lvf7g\" (UID: \"21aec953-2e02-48d3-85ff-872d48c4d661\") " pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.721866 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/21aec953-2e02-48d3-85ff-872d48c4d661-var-lib\") pod \"ovn-controller-ovs-lvf7g\" (UID: \"21aec953-2e02-48d3-85ff-872d48c4d661\") " pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.721908 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b72c4ff1-1234-41b7-9faa-b5fd15749084-var-run-ovn\") pod \"ovn-controller-zntd5\" (UID: \"b72c4ff1-1234-41b7-9faa-b5fd15749084\") " pod="openstack/ovn-controller-zntd5" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.721971 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/21aec953-2e02-48d3-85ff-872d48c4d661-var-run\") pod \"ovn-controller-ovs-lvf7g\" (UID: \"21aec953-2e02-48d3-85ff-872d48c4d661\") " pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.723227 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/21aec953-2e02-48d3-85ff-872d48c4d661-var-log\") pod \"ovn-controller-ovs-lvf7g\" (UID: \"21aec953-2e02-48d3-85ff-872d48c4d661\") " pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.724373 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b72c4ff1-1234-41b7-9faa-b5fd15749084-var-run\") pod \"ovn-controller-zntd5\" (UID: \"b72c4ff1-1234-41b7-9faa-b5fd15749084\") " pod="openstack/ovn-controller-zntd5" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.724424 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b72c4ff1-1234-41b7-9faa-b5fd15749084-var-log-ovn\") pod \"ovn-controller-zntd5\" (UID: \"b72c4ff1-1234-41b7-9faa-b5fd15749084\") " pod="openstack/ovn-controller-zntd5" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.724892 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b72c4ff1-1234-41b7-9faa-b5fd15749084-scripts\") pod \"ovn-controller-zntd5\" (UID: \"b72c4ff1-1234-41b7-9faa-b5fd15749084\") " pod="openstack/ovn-controller-zntd5" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.729175 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/b72c4ff1-1234-41b7-9faa-b5fd15749084-ovn-controller-tls-certs\") pod \"ovn-controller-zntd5\" (UID: \"b72c4ff1-1234-41b7-9faa-b5fd15749084\") " pod="openstack/ovn-controller-zntd5" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.740080 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/21aec953-2e02-48d3-85ff-872d48c4d661-scripts\") pod \"ovn-controller-ovs-lvf7g\" (UID: \"21aec953-2e02-48d3-85ff-872d48c4d661\") " pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.742703 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwfm7\" (UniqueName: \"kubernetes.io/projected/21aec953-2e02-48d3-85ff-872d48c4d661-kube-api-access-jwfm7\") pod \"ovn-controller-ovs-lvf7g\" (UID: \"21aec953-2e02-48d3-85ff-872d48c4d661\") " pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.771576 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xphf8\" (UniqueName: \"kubernetes.io/projected/b72c4ff1-1234-41b7-9faa-b5fd15749084-kube-api-access-xphf8\") pod \"ovn-controller-zntd5\" (UID: \"b72c4ff1-1234-41b7-9faa-b5fd15749084\") " pod="openstack/ovn-controller-zntd5" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.774692 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b72c4ff1-1234-41b7-9faa-b5fd15749084-combined-ca-bundle\") pod \"ovn-controller-zntd5\" (UID: \"b72c4ff1-1234-41b7-9faa-b5fd15749084\") " pod="openstack/ovn-controller-zntd5" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.807552 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zntd5" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.927488 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.975142 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.976619 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.979171 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.979594 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.980127 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.984319 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.985067 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-xmhfs" Mar 18 20:24:08 crc kubenswrapper[4950]: I0318 20:24:08.992646 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.127461 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49c49971-1a60-453f-8283-922ea50cc323-config\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.127506 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49c49971-1a60-453f-8283-922ea50cc323-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.127531 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49c49971-1a60-453f-8283-922ea50cc323-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.127657 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/49c49971-1a60-453f-8283-922ea50cc323-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.127759 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/49c49971-1a60-453f-8283-922ea50cc323-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.127815 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfqqc\" (UniqueName: \"kubernetes.io/projected/49c49971-1a60-453f-8283-922ea50cc323-kube-api-access-mfqqc\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.127959 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.127999 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/49c49971-1a60-453f-8283-922ea50cc323-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.229327 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.229393 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/49c49971-1a60-453f-8283-922ea50cc323-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.229514 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49c49971-1a60-453f-8283-922ea50cc323-config\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.229540 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49c49971-1a60-453f-8283-922ea50cc323-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.229830 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49c49971-1a60-453f-8283-922ea50cc323-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.229853 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/49c49971-1a60-453f-8283-922ea50cc323-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.229917 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/49c49971-1a60-453f-8283-922ea50cc323-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.229745 4950 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.229968 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfqqc\" (UniqueName: \"kubernetes.io/projected/49c49971-1a60-453f-8283-922ea50cc323-kube-api-access-mfqqc\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.230129 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/49c49971-1a60-453f-8283-922ea50cc323-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.231483 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49c49971-1a60-453f-8283-922ea50cc323-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.232020 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49c49971-1a60-453f-8283-922ea50cc323-config\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.240061 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49c49971-1a60-453f-8283-922ea50cc323-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.247160 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/49c49971-1a60-453f-8283-922ea50cc323-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.247570 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/49c49971-1a60-453f-8283-922ea50cc323-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.250308 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfqqc\" (UniqueName: \"kubernetes.io/projected/49c49971-1a60-453f-8283-922ea50cc323-kube-api-access-mfqqc\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.259147 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"49c49971-1a60-453f-8283-922ea50cc323\") " pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:09 crc kubenswrapper[4950]: I0318 20:24:09.338671 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.714227 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.717770 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.722111 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.722493 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-n9vhs" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.722807 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.723046 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.758151 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.881391 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwczg\" (UniqueName: \"kubernetes.io/projected/f37c3f8b-786f-42f6-8c5f-4db367edf6a5-kube-api-access-gwczg\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.881546 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.881738 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f37c3f8b-786f-42f6-8c5f-4db367edf6a5-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.881780 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f37c3f8b-786f-42f6-8c5f-4db367edf6a5-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.881819 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f37c3f8b-786f-42f6-8c5f-4db367edf6a5-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.881840 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f37c3f8b-786f-42f6-8c5f-4db367edf6a5-config\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.881879 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f37c3f8b-786f-42f6-8c5f-4db367edf6a5-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.882067 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f37c3f8b-786f-42f6-8c5f-4db367edf6a5-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.984041 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f37c3f8b-786f-42f6-8c5f-4db367edf6a5-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.984950 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f37c3f8b-786f-42f6-8c5f-4db367edf6a5-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.984982 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f37c3f8b-786f-42f6-8c5f-4db367edf6a5-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.984997 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f37c3f8b-786f-42f6-8c5f-4db367edf6a5-config\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.985014 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f37c3f8b-786f-42f6-8c5f-4db367edf6a5-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.985041 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f37c3f8b-786f-42f6-8c5f-4db367edf6a5-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.985097 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwczg\" (UniqueName: \"kubernetes.io/projected/f37c3f8b-786f-42f6-8c5f-4db367edf6a5-kube-api-access-gwczg\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.985124 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.985403 4950 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.985541 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f37c3f8b-786f-42f6-8c5f-4db367edf6a5-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.986450 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f37c3f8b-786f-42f6-8c5f-4db367edf6a5-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:11 crc kubenswrapper[4950]: I0318 20:24:11.987173 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f37c3f8b-786f-42f6-8c5f-4db367edf6a5-config\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:12 crc kubenswrapper[4950]: I0318 20:24:12.001475 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f37c3f8b-786f-42f6-8c5f-4db367edf6a5-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:12 crc kubenswrapper[4950]: I0318 20:24:12.002154 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f37c3f8b-786f-42f6-8c5f-4db367edf6a5-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:12 crc kubenswrapper[4950]: I0318 20:24:12.009581 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwczg\" (UniqueName: \"kubernetes.io/projected/f37c3f8b-786f-42f6-8c5f-4db367edf6a5-kube-api-access-gwczg\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:12 crc kubenswrapper[4950]: I0318 20:24:12.010984 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f37c3f8b-786f-42f6-8c5f-4db367edf6a5-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:12 crc kubenswrapper[4950]: I0318 20:24:12.024205 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f37c3f8b-786f-42f6-8c5f-4db367edf6a5\") " pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:12 crc kubenswrapper[4950]: I0318 20:24:12.038586 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:12 crc kubenswrapper[4950]: W0318 20:24:12.332389 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod859fa182_cb03_41ab_b127_20b3564cde97.slice/crio-e202347533c299d2380522657ce7347670815f7e94f83241af7284fb98a18268 WatchSource:0}: Error finding container e202347533c299d2380522657ce7347670815f7e94f83241af7284fb98a18268: Status 404 returned error can't find the container with id e202347533c299d2380522657ce7347670815f7e94f83241af7284fb98a18268 Mar 18 20:24:13 crc kubenswrapper[4950]: I0318 20:24:13.192225 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"859fa182-cb03-41ab-b127-20b3564cde97","Type":"ContainerStarted","Data":"e202347533c299d2380522657ce7347670815f7e94f83241af7284fb98a18268"} Mar 18 20:24:20 crc kubenswrapper[4950]: I0318 20:24:20.824032 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564424-q58sw" Mar 18 20:24:20 crc kubenswrapper[4950]: I0318 20:24:20.939303 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hc4pp\" (UniqueName: \"kubernetes.io/projected/2ff7fb3b-ac2f-4718-8c93-65b9a6291a20-kube-api-access-hc4pp\") pod \"2ff7fb3b-ac2f-4718-8c93-65b9a6291a20\" (UID: \"2ff7fb3b-ac2f-4718-8c93-65b9a6291a20\") " Mar 18 20:24:20 crc kubenswrapper[4950]: I0318 20:24:20.961584 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ff7fb3b-ac2f-4718-8c93-65b9a6291a20-kube-api-access-hc4pp" (OuterVolumeSpecName: "kube-api-access-hc4pp") pod "2ff7fb3b-ac2f-4718-8c93-65b9a6291a20" (UID: "2ff7fb3b-ac2f-4718-8c93-65b9a6291a20"). InnerVolumeSpecName "kube-api-access-hc4pp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:24:21 crc kubenswrapper[4950]: I0318 20:24:21.041442 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hc4pp\" (UniqueName: \"kubernetes.io/projected/2ff7fb3b-ac2f-4718-8c93-65b9a6291a20-kube-api-access-hc4pp\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:21 crc kubenswrapper[4950]: I0318 20:24:21.250724 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564424-q58sw" event={"ID":"2ff7fb3b-ac2f-4718-8c93-65b9a6291a20","Type":"ContainerDied","Data":"33ec77fd8b0f1ab05517b3a63dcf121a39d0b869160b08140103069ba2ade4b2"} Mar 18 20:24:21 crc kubenswrapper[4950]: I0318 20:24:21.250765 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33ec77fd8b0f1ab05517b3a63dcf121a39d0b869160b08140103069ba2ade4b2" Mar 18 20:24:21 crc kubenswrapper[4950]: I0318 20:24:21.250821 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564424-q58sw" Mar 18 20:24:21 crc kubenswrapper[4950]: E0318 20:24:21.386555 4950 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ff7fb3b_ac2f_4718_8c93_65b9a6291a20.slice/crio-33ec77fd8b0f1ab05517b3a63dcf121a39d0b869160b08140103069ba2ade4b2\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ff7fb3b_ac2f_4718_8c93_65b9a6291a20.slice\": RecentStats: unable to find data in memory cache]" Mar 18 20:24:21 crc kubenswrapper[4950]: I0318 20:24:21.909742 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564418-bh82f"] Mar 18 20:24:21 crc kubenswrapper[4950]: I0318 20:24:21.918543 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564418-bh82f"] Mar 18 20:24:23 crc kubenswrapper[4950]: I0318 20:24:23.497297 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f375d3f-628f-47a7-b2e5-d8afe6859aa8" path="/var/lib/kubelet/pods/9f375d3f-628f-47a7-b2e5-d8afe6859aa8/volumes" Mar 18 20:24:23 crc kubenswrapper[4950]: E0318 20:24:23.662928 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:2087a09e7ea9f1dbadd433366bb46cc93dd5460ac9606b65f430460f4c2ee18d" Mar 18 20:24:23 crc kubenswrapper[4950]: E0318 20:24:23.667695 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:2087a09e7ea9f1dbadd433366bb46cc93dd5460ac9606b65f430460f4c2ee18d" Mar 18 20:24:23 crc kubenswrapper[4950]: E0318 20:24:23.667938 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:2087a09e7ea9f1dbadd433366bb46cc93dd5460ac9606b65f430460f4c2ee18d,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gtmb7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(7502f5ea-31ca-4349-bd79-2219bf3903c0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 18 20:24:23 crc kubenswrapper[4950]: E0318 20:24:23.669154 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="7502f5ea-31ca-4349-bd79-2219bf3903c0" Mar 18 20:24:23 crc kubenswrapper[4950]: E0318 20:24:23.663617 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:2087a09e7ea9f1dbadd433366bb46cc93dd5460ac9606b65f430460f4c2ee18d,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g5gd4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(b82f7668-ae1a-4bfe-be27-435cca5df467): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 18 20:24:23 crc kubenswrapper[4950]: E0318 20:24:23.675814 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="b82f7668-ae1a-4bfe-be27-435cca5df467" Mar 18 20:24:24 crc kubenswrapper[4950]: E0318 20:24:24.272131 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:2087a09e7ea9f1dbadd433366bb46cc93dd5460ac9606b65f430460f4c2ee18d\\\"\"" pod="openstack/rabbitmq-server-0" podUID="b82f7668-ae1a-4bfe-be27-435cca5df467" Mar 18 20:24:24 crc kubenswrapper[4950]: E0318 20:24:24.272638 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:2087a09e7ea9f1dbadd433366bb46cc93dd5460ac9606b65f430460f4c2ee18d\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="7502f5ea-31ca-4349-bd79-2219bf3903c0" Mar 18 20:24:29 crc kubenswrapper[4950]: E0318 20:24:29.137696 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached@sha256:0d759b31e4da88b3fa1b823ab634d982fd713e81ed648626de1d8ec40ae7cad4" Mar 18 20:24:29 crc kubenswrapper[4950]: E0318 20:24:29.138194 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached@sha256:0d759b31e4da88b3fa1b823ab634d982fd713e81ed648626de1d8ec40ae7cad4,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n57chd5h54h5c4h695h585h554h96h58bh554h57h645hb9h5c7h79h658h655h5b6h696h689h569h5c4h78hfdhf8h596h569h4h5b5h5d5h695h545q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9pkqx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(18eb6f19-11f0-4023-93ac-fc03ecdb9486): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 18 20:24:29 crc kubenswrapper[4950]: E0318 20:24:29.141250 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="18eb6f19-11f0-4023-93ac-fc03ecdb9486" Mar 18 20:24:29 crc kubenswrapper[4950]: E0318 20:24:29.306528 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached@sha256:0d759b31e4da88b3fa1b823ab634d982fd713e81ed648626de1d8ec40ae7cad4\\\"\"" pod="openstack/memcached-0" podUID="18eb6f19-11f0-4023-93ac-fc03ecdb9486" Mar 18 20:24:29 crc kubenswrapper[4950]: E0318 20:24:29.923839 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" Mar 18 20:24:29 crc kubenswrapper[4950]: E0318 20:24:29.924395 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rkdhs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-64696987c5-nvh4b_openstack(03bf57fc-e31e-4e47-8ac5-774fdb8cfb50): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 18 20:24:29 crc kubenswrapper[4950]: E0318 20:24:29.925642 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-64696987c5-nvh4b" podUID="03bf57fc-e31e-4e47-8ac5-774fdb8cfb50" Mar 18 20:24:29 crc kubenswrapper[4950]: E0318 20:24:29.961106 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" Mar 18 20:24:29 crc kubenswrapper[4950]: E0318 20:24:29.961296 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m8tnl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-658f55c9f5-4qhbd_openstack(0e5ae697-1746-49b5-ac19-d0b5bf951ce9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 18 20:24:29 crc kubenswrapper[4950]: E0318 20:24:29.962948 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-658f55c9f5-4qhbd" podUID="0e5ae697-1746-49b5-ac19-d0b5bf951ce9" Mar 18 20:24:29 crc kubenswrapper[4950]: E0318 20:24:29.980126 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" Mar 18 20:24:29 crc kubenswrapper[4950]: E0318 20:24:29.980258 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b8qxq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-54b5dffb47-9wgpv_openstack(cf74b193-b765-42b9-883e-951eb9f12a40): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 18 20:24:29 crc kubenswrapper[4950]: E0318 20:24:29.981506 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-54b5dffb47-9wgpv" podUID="cf74b193-b765-42b9-883e-951eb9f12a40" Mar 18 20:24:30 crc kubenswrapper[4950]: E0318 20:24:30.047658 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" Mar 18 20:24:30 crc kubenswrapper[4950]: E0318 20:24:30.047899 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lfjtp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5448ff6dc7-l75zc_openstack(79c0ef32-5f60-4e97-b2ce-f2bfab2936ea): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 18 20:24:30 crc kubenswrapper[4950]: E0318 20:24:30.050846 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5448ff6dc7-l75zc" podUID="79c0ef32-5f60-4e97-b2ce-f2bfab2936ea" Mar 18 20:24:30 crc kubenswrapper[4950]: E0318 20:24:30.317708 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51\\\"\"" pod="openstack/dnsmasq-dns-54b5dffb47-9wgpv" podUID="cf74b193-b765-42b9-883e-951eb9f12a40" Mar 18 20:24:30 crc kubenswrapper[4950]: E0318 20:24:30.317708 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51\\\"\"" pod="openstack/dnsmasq-dns-658f55c9f5-4qhbd" podUID="0e5ae697-1746-49b5-ac19-d0b5bf951ce9" Mar 18 20:24:30 crc kubenswrapper[4950]: I0318 20:24:30.564761 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zntd5"] Mar 18 20:24:30 crc kubenswrapper[4950]: I0318 20:24:30.710984 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 18 20:24:30 crc kubenswrapper[4950]: W0318 20:24:30.745798 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb72c4ff1_1234_41b7_9faa_b5fd15749084.slice/crio-3dcc96ac842b383bad3c75f1375d27020ce260808ea4e00e8fbfe7a8d01d20f7 WatchSource:0}: Error finding container 3dcc96ac842b383bad3c75f1375d27020ce260808ea4e00e8fbfe7a8d01d20f7: Status 404 returned error can't find the container with id 3dcc96ac842b383bad3c75f1375d27020ce260808ea4e00e8fbfe7a8d01d20f7 Mar 18 20:24:30 crc kubenswrapper[4950]: W0318 20:24:30.748948 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf37c3f8b_786f_42f6_8c5f_4db367edf6a5.slice/crio-bc9fa6a16017e84e269aa3bc1b0728100831ccd5f0990a19e35942fb16e089ac WatchSource:0}: Error finding container bc9fa6a16017e84e269aa3bc1b0728100831ccd5f0990a19e35942fb16e089ac: Status 404 returned error can't find the container with id bc9fa6a16017e84e269aa3bc1b0728100831ccd5f0990a19e35942fb16e089ac Mar 18 20:24:30 crc kubenswrapper[4950]: I0318 20:24:30.808030 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5448ff6dc7-l75zc" Mar 18 20:24:30 crc kubenswrapper[4950]: I0318 20:24:30.815237 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64696987c5-nvh4b" Mar 18 20:24:30 crc kubenswrapper[4950]: I0318 20:24:30.914331 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfjtp\" (UniqueName: \"kubernetes.io/projected/79c0ef32-5f60-4e97-b2ce-f2bfab2936ea-kube-api-access-lfjtp\") pod \"79c0ef32-5f60-4e97-b2ce-f2bfab2936ea\" (UID: \"79c0ef32-5f60-4e97-b2ce-f2bfab2936ea\") " Mar 18 20:24:30 crc kubenswrapper[4950]: I0318 20:24:30.914478 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03bf57fc-e31e-4e47-8ac5-774fdb8cfb50-config\") pod \"03bf57fc-e31e-4e47-8ac5-774fdb8cfb50\" (UID: \"03bf57fc-e31e-4e47-8ac5-774fdb8cfb50\") " Mar 18 20:24:30 crc kubenswrapper[4950]: I0318 20:24:30.914510 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03bf57fc-e31e-4e47-8ac5-774fdb8cfb50-dns-svc\") pod \"03bf57fc-e31e-4e47-8ac5-774fdb8cfb50\" (UID: \"03bf57fc-e31e-4e47-8ac5-774fdb8cfb50\") " Mar 18 20:24:30 crc kubenswrapper[4950]: I0318 20:24:30.914605 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkdhs\" (UniqueName: \"kubernetes.io/projected/03bf57fc-e31e-4e47-8ac5-774fdb8cfb50-kube-api-access-rkdhs\") pod \"03bf57fc-e31e-4e47-8ac5-774fdb8cfb50\" (UID: \"03bf57fc-e31e-4e47-8ac5-774fdb8cfb50\") " Mar 18 20:24:30 crc kubenswrapper[4950]: I0318 20:24:30.915159 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79c0ef32-5f60-4e97-b2ce-f2bfab2936ea-config" (OuterVolumeSpecName: "config") pod "79c0ef32-5f60-4e97-b2ce-f2bfab2936ea" (UID: "79c0ef32-5f60-4e97-b2ce-f2bfab2936ea"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:24:30 crc kubenswrapper[4950]: I0318 20:24:30.915375 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03bf57fc-e31e-4e47-8ac5-774fdb8cfb50-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "03bf57fc-e31e-4e47-8ac5-774fdb8cfb50" (UID: "03bf57fc-e31e-4e47-8ac5-774fdb8cfb50"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:24:30 crc kubenswrapper[4950]: I0318 20:24:30.915439 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79c0ef32-5f60-4e97-b2ce-f2bfab2936ea-config\") pod \"79c0ef32-5f60-4e97-b2ce-f2bfab2936ea\" (UID: \"79c0ef32-5f60-4e97-b2ce-f2bfab2936ea\") " Mar 18 20:24:30 crc kubenswrapper[4950]: I0318 20:24:30.915490 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03bf57fc-e31e-4e47-8ac5-774fdb8cfb50-config" (OuterVolumeSpecName: "config") pod "03bf57fc-e31e-4e47-8ac5-774fdb8cfb50" (UID: "03bf57fc-e31e-4e47-8ac5-774fdb8cfb50"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:24:30 crc kubenswrapper[4950]: I0318 20:24:30.915826 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03bf57fc-e31e-4e47-8ac5-774fdb8cfb50-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:30 crc kubenswrapper[4950]: I0318 20:24:30.915852 4950 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03bf57fc-e31e-4e47-8ac5-774fdb8cfb50-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:30 crc kubenswrapper[4950]: I0318 20:24:30.915866 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79c0ef32-5f60-4e97-b2ce-f2bfab2936ea-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:30 crc kubenswrapper[4950]: I0318 20:24:30.920625 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03bf57fc-e31e-4e47-8ac5-774fdb8cfb50-kube-api-access-rkdhs" (OuterVolumeSpecName: "kube-api-access-rkdhs") pod "03bf57fc-e31e-4e47-8ac5-774fdb8cfb50" (UID: "03bf57fc-e31e-4e47-8ac5-774fdb8cfb50"). InnerVolumeSpecName "kube-api-access-rkdhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:24:30 crc kubenswrapper[4950]: I0318 20:24:30.924535 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79c0ef32-5f60-4e97-b2ce-f2bfab2936ea-kube-api-access-lfjtp" (OuterVolumeSpecName: "kube-api-access-lfjtp") pod "79c0ef32-5f60-4e97-b2ce-f2bfab2936ea" (UID: "79c0ef32-5f60-4e97-b2ce-f2bfab2936ea"). InnerVolumeSpecName "kube-api-access-lfjtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:24:31 crc kubenswrapper[4950]: I0318 20:24:31.017565 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkdhs\" (UniqueName: \"kubernetes.io/projected/03bf57fc-e31e-4e47-8ac5-774fdb8cfb50-kube-api-access-rkdhs\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:31 crc kubenswrapper[4950]: I0318 20:24:31.017597 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfjtp\" (UniqueName: \"kubernetes.io/projected/79c0ef32-5f60-4e97-b2ce-f2bfab2936ea-kube-api-access-lfjtp\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:31 crc kubenswrapper[4950]: I0318 20:24:31.325281 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f37c3f8b-786f-42f6-8c5f-4db367edf6a5","Type":"ContainerStarted","Data":"bc9fa6a16017e84e269aa3bc1b0728100831ccd5f0990a19e35942fb16e089ac"} Mar 18 20:24:31 crc kubenswrapper[4950]: I0318 20:24:31.327204 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zntd5" event={"ID":"b72c4ff1-1234-41b7-9faa-b5fd15749084","Type":"ContainerStarted","Data":"3dcc96ac842b383bad3c75f1375d27020ce260808ea4e00e8fbfe7a8d01d20f7"} Mar 18 20:24:31 crc kubenswrapper[4950]: I0318 20:24:31.328460 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5448ff6dc7-l75zc" event={"ID":"79c0ef32-5f60-4e97-b2ce-f2bfab2936ea","Type":"ContainerDied","Data":"935c313155c8117804dc62fc468d418c8e8bce76e98187be1548b77844296de1"} Mar 18 20:24:31 crc kubenswrapper[4950]: I0318 20:24:31.328532 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5448ff6dc7-l75zc" Mar 18 20:24:31 crc kubenswrapper[4950]: I0318 20:24:31.333207 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3","Type":"ContainerStarted","Data":"a668689848894a29fbbfb0315c973810c2cc0dfaed6cd0f93544b271255ce719"} Mar 18 20:24:31 crc kubenswrapper[4950]: I0318 20:24:31.335342 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64696987c5-nvh4b" event={"ID":"03bf57fc-e31e-4e47-8ac5-774fdb8cfb50","Type":"ContainerDied","Data":"b5a50b89076cfa0284f61b933bc296e04554cb1d5bf9349c99147042b55786d1"} Mar 18 20:24:31 crc kubenswrapper[4950]: I0318 20:24:31.335374 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64696987c5-nvh4b" Mar 18 20:24:31 crc kubenswrapper[4950]: I0318 20:24:31.401246 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5448ff6dc7-l75zc"] Mar 18 20:24:31 crc kubenswrapper[4950]: I0318 20:24:31.412130 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5448ff6dc7-l75zc"] Mar 18 20:24:31 crc kubenswrapper[4950]: I0318 20:24:31.451125 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64696987c5-nvh4b"] Mar 18 20:24:31 crc kubenswrapper[4950]: I0318 20:24:31.457792 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64696987c5-nvh4b"] Mar 18 20:24:31 crc kubenswrapper[4950]: I0318 20:24:31.500448 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03bf57fc-e31e-4e47-8ac5-774fdb8cfb50" path="/var/lib/kubelet/pods/03bf57fc-e31e-4e47-8ac5-774fdb8cfb50/volumes" Mar 18 20:24:31 crc kubenswrapper[4950]: I0318 20:24:31.500990 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79c0ef32-5f60-4e97-b2ce-f2bfab2936ea" path="/var/lib/kubelet/pods/79c0ef32-5f60-4e97-b2ce-f2bfab2936ea/volumes" Mar 18 20:24:31 crc kubenswrapper[4950]: I0318 20:24:31.542213 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 18 20:24:32 crc kubenswrapper[4950]: I0318 20:24:32.127661 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-lvf7g"] Mar 18 20:24:32 crc kubenswrapper[4950]: I0318 20:24:32.343198 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"49c49971-1a60-453f-8283-922ea50cc323","Type":"ContainerStarted","Data":"1559398315a6a7fa489d0f905a16f275dff55ca3411372c27fe96d692891c59e"} Mar 18 20:24:32 crc kubenswrapper[4950]: I0318 20:24:32.346918 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-lvf7g" event={"ID":"21aec953-2e02-48d3-85ff-872d48c4d661","Type":"ContainerStarted","Data":"b501d171ddc816bd5297d2862143041692bcd70670fa56e88af78e861523d128"} Mar 18 20:24:32 crc kubenswrapper[4950]: I0318 20:24:32.348203 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"aedf1977-d498-4b4f-a92b-f84be888045d","Type":"ContainerStarted","Data":"1ce3a9927845e138fe5c99aa6d522eb7678631f9f1ac96aa9560e373616a40f0"} Mar 18 20:24:32 crc kubenswrapper[4950]: I0318 20:24:32.350366 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"859fa182-cb03-41ab-b127-20b3564cde97","Type":"ContainerStarted","Data":"cc57df497094b1e0391644a07ee2893fe6ea32b3d25a2060022a9233de18a1f3"} Mar 18 20:24:32 crc kubenswrapper[4950]: I0318 20:24:32.351264 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 18 20:24:32 crc kubenswrapper[4950]: I0318 20:24:32.383525 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=8.646752381 podStartE2EDuration="28.383503655s" podCreationTimestamp="2026-03-18 20:24:04 +0000 UTC" firstStartedPulling="2026-03-18 20:24:12.334924227 +0000 UTC m=+1065.575766095" lastFinishedPulling="2026-03-18 20:24:32.071675511 +0000 UTC m=+1085.312517369" observedRunningTime="2026-03-18 20:24:32.377639391 +0000 UTC m=+1085.618481259" watchObservedRunningTime="2026-03-18 20:24:32.383503655 +0000 UTC m=+1085.624345523" Mar 18 20:24:33 crc kubenswrapper[4950]: I0318 20:24:33.836499 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:24:33 crc kubenswrapper[4950]: I0318 20:24:33.836549 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:24:34 crc kubenswrapper[4950]: I0318 20:24:34.364401 4950 generic.go:334] "Generic (PLEG): container finished" podID="e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3" containerID="a668689848894a29fbbfb0315c973810c2cc0dfaed6cd0f93544b271255ce719" exitCode=0 Mar 18 20:24:34 crc kubenswrapper[4950]: I0318 20:24:34.364461 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3","Type":"ContainerDied","Data":"a668689848894a29fbbfb0315c973810c2cc0dfaed6cd0f93544b271255ce719"} Mar 18 20:24:35 crc kubenswrapper[4950]: I0318 20:24:35.374217 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3","Type":"ContainerStarted","Data":"cc96b901ae51dfb05cd23aa5016794f07a4f3c63465c78e5ea1c19c5b34b10e4"} Mar 18 20:24:35 crc kubenswrapper[4950]: I0318 20:24:35.376369 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"49c49971-1a60-453f-8283-922ea50cc323","Type":"ContainerStarted","Data":"20bf236104604ffcbe92c89d62dd75de67750582c9e1b1506b17cfaeaa51e597"} Mar 18 20:24:35 crc kubenswrapper[4950]: I0318 20:24:35.377853 4950 generic.go:334] "Generic (PLEG): container finished" podID="21aec953-2e02-48d3-85ff-872d48c4d661" containerID="62aa1a9353c0e5abb94e57a02b1bc53ddb8b3819f36f64a18af20f869fa48ffe" exitCode=0 Mar 18 20:24:35 crc kubenswrapper[4950]: I0318 20:24:35.377922 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-lvf7g" event={"ID":"21aec953-2e02-48d3-85ff-872d48c4d661","Type":"ContainerDied","Data":"62aa1a9353c0e5abb94e57a02b1bc53ddb8b3819f36f64a18af20f869fa48ffe"} Mar 18 20:24:35 crc kubenswrapper[4950]: I0318 20:24:35.380960 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f37c3f8b-786f-42f6-8c5f-4db367edf6a5","Type":"ContainerStarted","Data":"84a18eb22fee98b46f3ef6be6ef47bb3c1c25a54366f1e2a63558dc4c2288375"} Mar 18 20:24:35 crc kubenswrapper[4950]: I0318 20:24:35.382315 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zntd5" event={"ID":"b72c4ff1-1234-41b7-9faa-b5fd15749084","Type":"ContainerStarted","Data":"1eb916a417d9cc8dbb609b8df23bd329aad770108736f21701f72673248ff799"} Mar 18 20:24:35 crc kubenswrapper[4950]: I0318 20:24:35.382667 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-zntd5" Mar 18 20:24:35 crc kubenswrapper[4950]: I0318 20:24:35.400264 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=8.779416349 podStartE2EDuration="35.400244386s" podCreationTimestamp="2026-03-18 20:24:00 +0000 UTC" firstStartedPulling="2026-03-18 20:24:02.503901407 +0000 UTC m=+1055.744743265" lastFinishedPulling="2026-03-18 20:24:29.124729434 +0000 UTC m=+1082.365571302" observedRunningTime="2026-03-18 20:24:35.399849355 +0000 UTC m=+1088.640691223" watchObservedRunningTime="2026-03-18 20:24:35.400244386 +0000 UTC m=+1088.641086254" Mar 18 20:24:35 crc kubenswrapper[4950]: I0318 20:24:35.427545 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-zntd5" podStartSLOduration=23.498999514 podStartE2EDuration="27.427529273s" podCreationTimestamp="2026-03-18 20:24:08 +0000 UTC" firstStartedPulling="2026-03-18 20:24:30.749149159 +0000 UTC m=+1083.989991027" lastFinishedPulling="2026-03-18 20:24:34.677678918 +0000 UTC m=+1087.918520786" observedRunningTime="2026-03-18 20:24:35.416329078 +0000 UTC m=+1088.657170966" watchObservedRunningTime="2026-03-18 20:24:35.427529273 +0000 UTC m=+1088.668371141" Mar 18 20:24:36 crc kubenswrapper[4950]: I0318 20:24:36.397294 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-lvf7g" event={"ID":"21aec953-2e02-48d3-85ff-872d48c4d661","Type":"ContainerStarted","Data":"5b8228c811621f6f890f78f07bb92ef949efef2814307a830f03b6b92004fa36"} Mar 18 20:24:36 crc kubenswrapper[4950]: I0318 20:24:36.397654 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-lvf7g" event={"ID":"21aec953-2e02-48d3-85ff-872d48c4d661","Type":"ContainerStarted","Data":"97ba2ee126f4439c363f8145d9050ebf51eb9901f25e0d64fb03e953bf3ce5d4"} Mar 18 20:24:36 crc kubenswrapper[4950]: I0318 20:24:36.397672 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:24:36 crc kubenswrapper[4950]: I0318 20:24:36.400567 4950 generic.go:334] "Generic (PLEG): container finished" podID="aedf1977-d498-4b4f-a92b-f84be888045d" containerID="1ce3a9927845e138fe5c99aa6d522eb7678631f9f1ac96aa9560e373616a40f0" exitCode=0 Mar 18 20:24:36 crc kubenswrapper[4950]: I0318 20:24:36.400856 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"aedf1977-d498-4b4f-a92b-f84be888045d","Type":"ContainerDied","Data":"1ce3a9927845e138fe5c99aa6d522eb7678631f9f1ac96aa9560e373616a40f0"} Mar 18 20:24:36 crc kubenswrapper[4950]: I0318 20:24:36.420663 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-lvf7g" podStartSLOduration=25.891494592 podStartE2EDuration="28.420642758s" podCreationTimestamp="2026-03-18 20:24:08 +0000 UTC" firstStartedPulling="2026-03-18 20:24:32.143833458 +0000 UTC m=+1085.384675326" lastFinishedPulling="2026-03-18 20:24:34.672981614 +0000 UTC m=+1087.913823492" observedRunningTime="2026-03-18 20:24:36.415511153 +0000 UTC m=+1089.656353041" watchObservedRunningTime="2026-03-18 20:24:36.420642758 +0000 UTC m=+1089.661484626" Mar 18 20:24:37 crc kubenswrapper[4950]: I0318 20:24:37.409794 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:24:39 crc kubenswrapper[4950]: I0318 20:24:39.427224 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f37c3f8b-786f-42f6-8c5f-4db367edf6a5","Type":"ContainerStarted","Data":"308e2bb80d0d6af520f620132850819c8ee1c02d355c77855dc4918e0d7c573f"} Mar 18 20:24:39 crc kubenswrapper[4950]: I0318 20:24:39.433439 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"49c49971-1a60-453f-8283-922ea50cc323","Type":"ContainerStarted","Data":"058a7f24345eda7e0b546205160f0bb4b3c19f22951da08bbe1c2337e497da73"} Mar 18 20:24:39 crc kubenswrapper[4950]: I0318 20:24:39.438478 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"aedf1977-d498-4b4f-a92b-f84be888045d","Type":"ContainerStarted","Data":"dbbec1650b11fa433295caf97fd1dcbadda65eda06859550b7dee042dce5c6d1"} Mar 18 20:24:39 crc kubenswrapper[4950]: I0318 20:24:39.454699 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=21.403148904 podStartE2EDuration="29.454684295s" podCreationTimestamp="2026-03-18 20:24:10 +0000 UTC" firstStartedPulling="2026-03-18 20:24:30.760016554 +0000 UTC m=+1084.000858422" lastFinishedPulling="2026-03-18 20:24:38.811551945 +0000 UTC m=+1092.052393813" observedRunningTime="2026-03-18 20:24:39.450097854 +0000 UTC m=+1092.690939722" watchObservedRunningTime="2026-03-18 20:24:39.454684295 +0000 UTC m=+1092.695526163" Mar 18 20:24:39 crc kubenswrapper[4950]: I0318 20:24:39.481152 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=11.920112865 podStartE2EDuration="38.48113462s" podCreationTimestamp="2026-03-18 20:24:01 +0000 UTC" firstStartedPulling="2026-03-18 20:24:03.384203299 +0000 UTC m=+1056.625045167" lastFinishedPulling="2026-03-18 20:24:29.945225034 +0000 UTC m=+1083.186066922" observedRunningTime="2026-03-18 20:24:39.475151853 +0000 UTC m=+1092.715993721" watchObservedRunningTime="2026-03-18 20:24:39.48113462 +0000 UTC m=+1092.721976488" Mar 18 20:24:39 crc kubenswrapper[4950]: I0318 20:24:39.515279 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=25.717750029 podStartE2EDuration="32.515253476s" podCreationTimestamp="2026-03-18 20:24:07 +0000 UTC" firstStartedPulling="2026-03-18 20:24:32.011894921 +0000 UTC m=+1085.252736789" lastFinishedPulling="2026-03-18 20:24:38.809398368 +0000 UTC m=+1092.050240236" observedRunningTime="2026-03-18 20:24:39.511391095 +0000 UTC m=+1092.752233003" watchObservedRunningTime="2026-03-18 20:24:39.515253476 +0000 UTC m=+1092.756095344" Mar 18 20:24:40 crc kubenswrapper[4950]: I0318 20:24:40.467880 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b82f7668-ae1a-4bfe-be27-435cca5df467","Type":"ContainerStarted","Data":"a9215184422c6bfb281f115311fa646f8d17ed53337255e639a33ab2c4583605"} Mar 18 20:24:40 crc kubenswrapper[4950]: I0318 20:24:40.471128 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7502f5ea-31ca-4349-bd79-2219bf3903c0","Type":"ContainerStarted","Data":"c87e384b986ec489013da0da075d69803e2eecf7ea2534fbb8b55233e4c8990e"} Mar 18 20:24:41 crc kubenswrapper[4950]: I0318 20:24:41.766554 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Mar 18 20:24:41 crc kubenswrapper[4950]: I0318 20:24:41.766621 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Mar 18 20:24:41 crc kubenswrapper[4950]: I0318 20:24:41.831715 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.038878 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.039320 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.103660 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.339949 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.398104 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.495518 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.555405 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.559298 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.636395 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.746334 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.746405 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.788295 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54b5dffb47-9wgpv"] Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.813219 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ccf85c649-grdqp"] Mar 18 20:24:42 crc kubenswrapper[4950]: E0318 20:24:42.813512 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff7fb3b-ac2f-4718-8c93-65b9a6291a20" containerName="oc" Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.813528 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff7fb3b-ac2f-4718-8c93-65b9a6291a20" containerName="oc" Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.813670 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff7fb3b-ac2f-4718-8c93-65b9a6291a20" containerName="oc" Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.816584 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.827128 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.844800 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ccf85c649-grdqp"] Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.916923 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbb60200-f508-4614-baa8-5ca52e7d0127-config\") pod \"dnsmasq-dns-7ccf85c649-grdqp\" (UID: \"fbb60200-f508-4614-baa8-5ca52e7d0127\") " pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.916968 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fbb60200-f508-4614-baa8-5ca52e7d0127-ovsdbserver-sb\") pod \"dnsmasq-dns-7ccf85c649-grdqp\" (UID: \"fbb60200-f508-4614-baa8-5ca52e7d0127\") " pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.916987 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fbb60200-f508-4614-baa8-5ca52e7d0127-dns-svc\") pod \"dnsmasq-dns-7ccf85c649-grdqp\" (UID: \"fbb60200-f508-4614-baa8-5ca52e7d0127\") " pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.917058 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48zjx\" (UniqueName: \"kubernetes.io/projected/fbb60200-f508-4614-baa8-5ca52e7d0127-kube-api-access-48zjx\") pod \"dnsmasq-dns-7ccf85c649-grdqp\" (UID: \"fbb60200-f508-4614-baa8-5ca52e7d0127\") " pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.944534 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-9wnnb"] Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.949183 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-9wnnb" Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.954219 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Mar 18 20:24:42 crc kubenswrapper[4950]: I0318 20:24:42.997978 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-9wnnb"] Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.018299 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48zjx\" (UniqueName: \"kubernetes.io/projected/fbb60200-f508-4614-baa8-5ca52e7d0127-kube-api-access-48zjx\") pod \"dnsmasq-dns-7ccf85c649-grdqp\" (UID: \"fbb60200-f508-4614-baa8-5ca52e7d0127\") " pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.018358 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f70fe2d-f600-42da-8a3d-a4f814374ca7-combined-ca-bundle\") pod \"ovn-controller-metrics-9wnnb\" (UID: \"3f70fe2d-f600-42da-8a3d-a4f814374ca7\") " pod="openstack/ovn-controller-metrics-9wnnb" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.018389 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f70fe2d-f600-42da-8a3d-a4f814374ca7-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-9wnnb\" (UID: \"3f70fe2d-f600-42da-8a3d-a4f814374ca7\") " pod="openstack/ovn-controller-metrics-9wnnb" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.018463 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f70fe2d-f600-42da-8a3d-a4f814374ca7-config\") pod \"ovn-controller-metrics-9wnnb\" (UID: \"3f70fe2d-f600-42da-8a3d-a4f814374ca7\") " pod="openstack/ovn-controller-metrics-9wnnb" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.018491 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3f70fe2d-f600-42da-8a3d-a4f814374ca7-ovs-rundir\") pod \"ovn-controller-metrics-9wnnb\" (UID: \"3f70fe2d-f600-42da-8a3d-a4f814374ca7\") " pod="openstack/ovn-controller-metrics-9wnnb" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.018517 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbb60200-f508-4614-baa8-5ca52e7d0127-config\") pod \"dnsmasq-dns-7ccf85c649-grdqp\" (UID: \"fbb60200-f508-4614-baa8-5ca52e7d0127\") " pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.018533 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fbb60200-f508-4614-baa8-5ca52e7d0127-ovsdbserver-sb\") pod \"dnsmasq-dns-7ccf85c649-grdqp\" (UID: \"fbb60200-f508-4614-baa8-5ca52e7d0127\") " pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.018548 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3f70fe2d-f600-42da-8a3d-a4f814374ca7-ovn-rundir\") pod \"ovn-controller-metrics-9wnnb\" (UID: \"3f70fe2d-f600-42da-8a3d-a4f814374ca7\") " pod="openstack/ovn-controller-metrics-9wnnb" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.018562 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fbb60200-f508-4614-baa8-5ca52e7d0127-dns-svc\") pod \"dnsmasq-dns-7ccf85c649-grdqp\" (UID: \"fbb60200-f508-4614-baa8-5ca52e7d0127\") " pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.018587 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wmtx\" (UniqueName: \"kubernetes.io/projected/3f70fe2d-f600-42da-8a3d-a4f814374ca7-kube-api-access-8wmtx\") pod \"ovn-controller-metrics-9wnnb\" (UID: \"3f70fe2d-f600-42da-8a3d-a4f814374ca7\") " pod="openstack/ovn-controller-metrics-9wnnb" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.020048 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbb60200-f508-4614-baa8-5ca52e7d0127-config\") pod \"dnsmasq-dns-7ccf85c649-grdqp\" (UID: \"fbb60200-f508-4614-baa8-5ca52e7d0127\") " pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.020370 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fbb60200-f508-4614-baa8-5ca52e7d0127-ovsdbserver-sb\") pod \"dnsmasq-dns-7ccf85c649-grdqp\" (UID: \"fbb60200-f508-4614-baa8-5ca52e7d0127\") " pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.026105 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fbb60200-f508-4614-baa8-5ca52e7d0127-dns-svc\") pod \"dnsmasq-dns-7ccf85c649-grdqp\" (UID: \"fbb60200-f508-4614-baa8-5ca52e7d0127\") " pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.074277 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48zjx\" (UniqueName: \"kubernetes.io/projected/fbb60200-f508-4614-baa8-5ca52e7d0127-kube-api-access-48zjx\") pod \"dnsmasq-dns-7ccf85c649-grdqp\" (UID: \"fbb60200-f508-4614-baa8-5ca52e7d0127\") " pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.117067 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-658f55c9f5-4qhbd"] Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.120265 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f70fe2d-f600-42da-8a3d-a4f814374ca7-config\") pod \"ovn-controller-metrics-9wnnb\" (UID: \"3f70fe2d-f600-42da-8a3d-a4f814374ca7\") " pod="openstack/ovn-controller-metrics-9wnnb" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.120329 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3f70fe2d-f600-42da-8a3d-a4f814374ca7-ovs-rundir\") pod \"ovn-controller-metrics-9wnnb\" (UID: \"3f70fe2d-f600-42da-8a3d-a4f814374ca7\") " pod="openstack/ovn-controller-metrics-9wnnb" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.120366 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3f70fe2d-f600-42da-8a3d-a4f814374ca7-ovn-rundir\") pod \"ovn-controller-metrics-9wnnb\" (UID: \"3f70fe2d-f600-42da-8a3d-a4f814374ca7\") " pod="openstack/ovn-controller-metrics-9wnnb" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.120399 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wmtx\" (UniqueName: \"kubernetes.io/projected/3f70fe2d-f600-42da-8a3d-a4f814374ca7-kube-api-access-8wmtx\") pod \"ovn-controller-metrics-9wnnb\" (UID: \"3f70fe2d-f600-42da-8a3d-a4f814374ca7\") " pod="openstack/ovn-controller-metrics-9wnnb" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.120485 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f70fe2d-f600-42da-8a3d-a4f814374ca7-combined-ca-bundle\") pod \"ovn-controller-metrics-9wnnb\" (UID: \"3f70fe2d-f600-42da-8a3d-a4f814374ca7\") " pod="openstack/ovn-controller-metrics-9wnnb" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.120529 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f70fe2d-f600-42da-8a3d-a4f814374ca7-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-9wnnb\" (UID: \"3f70fe2d-f600-42da-8a3d-a4f814374ca7\") " pod="openstack/ovn-controller-metrics-9wnnb" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.121314 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3f70fe2d-f600-42da-8a3d-a4f814374ca7-ovn-rundir\") pod \"ovn-controller-metrics-9wnnb\" (UID: \"3f70fe2d-f600-42da-8a3d-a4f814374ca7\") " pod="openstack/ovn-controller-metrics-9wnnb" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.122147 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f70fe2d-f600-42da-8a3d-a4f814374ca7-config\") pod \"ovn-controller-metrics-9wnnb\" (UID: \"3f70fe2d-f600-42da-8a3d-a4f814374ca7\") " pod="openstack/ovn-controller-metrics-9wnnb" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.122197 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3f70fe2d-f600-42da-8a3d-a4f814374ca7-ovs-rundir\") pod \"ovn-controller-metrics-9wnnb\" (UID: \"3f70fe2d-f600-42da-8a3d-a4f814374ca7\") " pod="openstack/ovn-controller-metrics-9wnnb" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.128968 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f70fe2d-f600-42da-8a3d-a4f814374ca7-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-9wnnb\" (UID: \"3f70fe2d-f600-42da-8a3d-a4f814374ca7\") " pod="openstack/ovn-controller-metrics-9wnnb" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.139989 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.141509 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.141908 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.150158 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f70fe2d-f600-42da-8a3d-a4f814374ca7-combined-ca-bundle\") pod \"ovn-controller-metrics-9wnnb\" (UID: \"3f70fe2d-f600-42da-8a3d-a4f814374ca7\") " pod="openstack/ovn-controller-metrics-9wnnb" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.155867 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.156318 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.156461 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-dk88k" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.156580 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.181820 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.201611 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wmtx\" (UniqueName: \"kubernetes.io/projected/3f70fe2d-f600-42da-8a3d-a4f814374ca7-kube-api-access-8wmtx\") pod \"ovn-controller-metrics-9wnnb\" (UID: \"3f70fe2d-f600-42da-8a3d-a4f814374ca7\") " pod="openstack/ovn-controller-metrics-9wnnb" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.201682 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f697c8bff-wvwhn"] Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.213973 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f697c8bff-wvwhn"] Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.214074 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.217108 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.225651 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8cql\" (UniqueName: \"kubernetes.io/projected/a3a71276-b0fa-481f-a91e-551d017c6462-kube-api-access-s8cql\") pod \"ovn-northd-0\" (UID: \"a3a71276-b0fa-481f-a91e-551d017c6462\") " pod="openstack/ovn-northd-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.225721 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3a71276-b0fa-481f-a91e-551d017c6462-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a3a71276-b0fa-481f-a91e-551d017c6462\") " pod="openstack/ovn-northd-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.225758 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3a71276-b0fa-481f-a91e-551d017c6462-config\") pod \"ovn-northd-0\" (UID: \"a3a71276-b0fa-481f-a91e-551d017c6462\") " pod="openstack/ovn-northd-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.225790 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a3a71276-b0fa-481f-a91e-551d017c6462-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a3a71276-b0fa-481f-a91e-551d017c6462\") " pod="openstack/ovn-northd-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.225810 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3a71276-b0fa-481f-a91e-551d017c6462-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a3a71276-b0fa-481f-a91e-551d017c6462\") " pod="openstack/ovn-northd-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.225830 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3a71276-b0fa-481f-a91e-551d017c6462-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a3a71276-b0fa-481f-a91e-551d017c6462\") " pod="openstack/ovn-northd-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.225898 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a3a71276-b0fa-481f-a91e-551d017c6462-scripts\") pod \"ovn-northd-0\" (UID: \"a3a71276-b0fa-481f-a91e-551d017c6462\") " pod="openstack/ovn-northd-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.327686 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3a71276-b0fa-481f-a91e-551d017c6462-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a3a71276-b0fa-481f-a91e-551d017c6462\") " pod="openstack/ovn-northd-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.327756 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-dns-svc\") pod \"dnsmasq-dns-f697c8bff-wvwhn\" (UID: \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\") " pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.327797 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a3a71276-b0fa-481f-a91e-551d017c6462-scripts\") pod \"ovn-northd-0\" (UID: \"a3a71276-b0fa-481f-a91e-551d017c6462\") " pod="openstack/ovn-northd-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.327836 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-ovsdbserver-sb\") pod \"dnsmasq-dns-f697c8bff-wvwhn\" (UID: \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\") " pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.327855 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8cql\" (UniqueName: \"kubernetes.io/projected/a3a71276-b0fa-481f-a91e-551d017c6462-kube-api-access-s8cql\") pod \"ovn-northd-0\" (UID: \"a3a71276-b0fa-481f-a91e-551d017c6462\") " pod="openstack/ovn-northd-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.327903 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3a71276-b0fa-481f-a91e-551d017c6462-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a3a71276-b0fa-481f-a91e-551d017c6462\") " pod="openstack/ovn-northd-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.327928 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3a71276-b0fa-481f-a91e-551d017c6462-config\") pod \"ovn-northd-0\" (UID: \"a3a71276-b0fa-481f-a91e-551d017c6462\") " pod="openstack/ovn-northd-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.327949 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djjt6\" (UniqueName: \"kubernetes.io/projected/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-kube-api-access-djjt6\") pod \"dnsmasq-dns-f697c8bff-wvwhn\" (UID: \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\") " pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.327977 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-ovsdbserver-nb\") pod \"dnsmasq-dns-f697c8bff-wvwhn\" (UID: \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\") " pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.327995 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a3a71276-b0fa-481f-a91e-551d017c6462-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a3a71276-b0fa-481f-a91e-551d017c6462\") " pod="openstack/ovn-northd-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.328010 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-config\") pod \"dnsmasq-dns-f697c8bff-wvwhn\" (UID: \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\") " pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.328093 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3a71276-b0fa-481f-a91e-551d017c6462-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a3a71276-b0fa-481f-a91e-551d017c6462\") " pod="openstack/ovn-northd-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.329281 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a3a71276-b0fa-481f-a91e-551d017c6462-scripts\") pod \"ovn-northd-0\" (UID: \"a3a71276-b0fa-481f-a91e-551d017c6462\") " pod="openstack/ovn-northd-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.329603 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a3a71276-b0fa-481f-a91e-551d017c6462-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a3a71276-b0fa-481f-a91e-551d017c6462\") " pod="openstack/ovn-northd-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.329785 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3a71276-b0fa-481f-a91e-551d017c6462-config\") pod \"ovn-northd-0\" (UID: \"a3a71276-b0fa-481f-a91e-551d017c6462\") " pod="openstack/ovn-northd-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.334498 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3a71276-b0fa-481f-a91e-551d017c6462-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a3a71276-b0fa-481f-a91e-551d017c6462\") " pod="openstack/ovn-northd-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.336003 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3a71276-b0fa-481f-a91e-551d017c6462-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a3a71276-b0fa-481f-a91e-551d017c6462\") " pod="openstack/ovn-northd-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.339746 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3a71276-b0fa-481f-a91e-551d017c6462-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a3a71276-b0fa-481f-a91e-551d017c6462\") " pod="openstack/ovn-northd-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.352961 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8cql\" (UniqueName: \"kubernetes.io/projected/a3a71276-b0fa-481f-a91e-551d017c6462-kube-api-access-s8cql\") pod \"ovn-northd-0\" (UID: \"a3a71276-b0fa-481f-a91e-551d017c6462\") " pod="openstack/ovn-northd-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.401077 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-9wnnb" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.420942 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54b5dffb47-9wgpv" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.431000 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-ovsdbserver-sb\") pod \"dnsmasq-dns-f697c8bff-wvwhn\" (UID: \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\") " pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.431119 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djjt6\" (UniqueName: \"kubernetes.io/projected/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-kube-api-access-djjt6\") pod \"dnsmasq-dns-f697c8bff-wvwhn\" (UID: \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\") " pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.431338 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-ovsdbserver-nb\") pod \"dnsmasq-dns-f697c8bff-wvwhn\" (UID: \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\") " pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.431362 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-config\") pod \"dnsmasq-dns-f697c8bff-wvwhn\" (UID: \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\") " pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.431450 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-dns-svc\") pod \"dnsmasq-dns-f697c8bff-wvwhn\" (UID: \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\") " pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.434765 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-ovsdbserver-nb\") pod \"dnsmasq-dns-f697c8bff-wvwhn\" (UID: \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\") " pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.435346 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-config\") pod \"dnsmasq-dns-f697c8bff-wvwhn\" (UID: \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\") " pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.439116 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-dns-svc\") pod \"dnsmasq-dns-f697c8bff-wvwhn\" (UID: \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\") " pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.442211 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-ovsdbserver-sb\") pod \"dnsmasq-dns-f697c8bff-wvwhn\" (UID: \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\") " pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.455311 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djjt6\" (UniqueName: \"kubernetes.io/projected/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-kube-api-access-djjt6\") pod \"dnsmasq-dns-f697c8bff-wvwhn\" (UID: \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\") " pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.509824 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54b5dffb47-9wgpv" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.532802 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf74b193-b765-42b9-883e-951eb9f12a40-dns-svc\") pod \"cf74b193-b765-42b9-883e-951eb9f12a40\" (UID: \"cf74b193-b765-42b9-883e-951eb9f12a40\") " Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.532853 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf74b193-b765-42b9-883e-951eb9f12a40-config\") pod \"cf74b193-b765-42b9-883e-951eb9f12a40\" (UID: \"cf74b193-b765-42b9-883e-951eb9f12a40\") " Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.532873 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8qxq\" (UniqueName: \"kubernetes.io/projected/cf74b193-b765-42b9-883e-951eb9f12a40-kube-api-access-b8qxq\") pod \"cf74b193-b765-42b9-883e-951eb9f12a40\" (UID: \"cf74b193-b765-42b9-883e-951eb9f12a40\") " Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.537843 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.538699 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf74b193-b765-42b9-883e-951eb9f12a40-config" (OuterVolumeSpecName: "config") pod "cf74b193-b765-42b9-883e-951eb9f12a40" (UID: "cf74b193-b765-42b9-883e-951eb9f12a40"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.540652 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf74b193-b765-42b9-883e-951eb9f12a40-kube-api-access-b8qxq" (OuterVolumeSpecName: "kube-api-access-b8qxq") pod "cf74b193-b765-42b9-883e-951eb9f12a40" (UID: "cf74b193-b765-42b9-883e-951eb9f12a40"). InnerVolumeSpecName "kube-api-access-b8qxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.540928 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54b5dffb47-9wgpv" event={"ID":"cf74b193-b765-42b9-883e-951eb9f12a40","Type":"ContainerDied","Data":"14f3e2b99481903a82f697779da288400af99b1cd657bddc5ce4a6a5b0ff9ef7"} Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.540953 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"18eb6f19-11f0-4023-93ac-fc03ecdb9486","Type":"ContainerStarted","Data":"080eb7a6fc7400017fd681cb3b2544c4bf8c8b720e371af58b299fc68b740f70"} Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.542513 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.542718 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf74b193-b765-42b9-883e-951eb9f12a40-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cf74b193-b765-42b9-883e-951eb9f12a40" (UID: "cf74b193-b765-42b9-883e-951eb9f12a40"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.565328 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.596079 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.299067879 podStartE2EDuration="41.596057567s" podCreationTimestamp="2026-03-18 20:24:02 +0000 UTC" firstStartedPulling="2026-03-18 20:24:03.65747362 +0000 UTC m=+1056.898315488" lastFinishedPulling="2026-03-18 20:24:42.954463308 +0000 UTC m=+1096.195305176" observedRunningTime="2026-03-18 20:24:43.590520531 +0000 UTC m=+1096.831362399" watchObservedRunningTime="2026-03-18 20:24:43.596057567 +0000 UTC m=+1096.836899445" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.635773 4950 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf74b193-b765-42b9-883e-951eb9f12a40-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.635805 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf74b193-b765-42b9-883e-951eb9f12a40-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.635814 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8qxq\" (UniqueName: \"kubernetes.io/projected/cf74b193-b765-42b9-883e-951eb9f12a40-kube-api-access-b8qxq\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.636718 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658f55c9f5-4qhbd" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.742258 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e5ae697-1746-49b5-ac19-d0b5bf951ce9-dns-svc\") pod \"0e5ae697-1746-49b5-ac19-d0b5bf951ce9\" (UID: \"0e5ae697-1746-49b5-ac19-d0b5bf951ce9\") " Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.745373 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8tnl\" (UniqueName: \"kubernetes.io/projected/0e5ae697-1746-49b5-ac19-d0b5bf951ce9-kube-api-access-m8tnl\") pod \"0e5ae697-1746-49b5-ac19-d0b5bf951ce9\" (UID: \"0e5ae697-1746-49b5-ac19-d0b5bf951ce9\") " Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.745448 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e5ae697-1746-49b5-ac19-d0b5bf951ce9-config\") pod \"0e5ae697-1746-49b5-ac19-d0b5bf951ce9\" (UID: \"0e5ae697-1746-49b5-ac19-d0b5bf951ce9\") " Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.745861 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e5ae697-1746-49b5-ac19-d0b5bf951ce9-config" (OuterVolumeSpecName: "config") pod "0e5ae697-1746-49b5-ac19-d0b5bf951ce9" (UID: "0e5ae697-1746-49b5-ac19-d0b5bf951ce9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.746079 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e5ae697-1746-49b5-ac19-d0b5bf951ce9-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.750326 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e5ae697-1746-49b5-ac19-d0b5bf951ce9-kube-api-access-m8tnl" (OuterVolumeSpecName: "kube-api-access-m8tnl") pod "0e5ae697-1746-49b5-ac19-d0b5bf951ce9" (UID: "0e5ae697-1746-49b5-ac19-d0b5bf951ce9"). InnerVolumeSpecName "kube-api-access-m8tnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.750654 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e5ae697-1746-49b5-ac19-d0b5bf951ce9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0e5ae697-1746-49b5-ac19-d0b5bf951ce9" (UID: "0e5ae697-1746-49b5-ac19-d0b5bf951ce9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.854040 4950 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e5ae697-1746-49b5-ac19-d0b5bf951ce9-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.854072 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8tnl\" (UniqueName: \"kubernetes.io/projected/0e5ae697-1746-49b5-ac19-d0b5bf951ce9-kube-api-access-m8tnl\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.889628 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54b5dffb47-9wgpv"] Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.902999 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54b5dffb47-9wgpv"] Mar 18 20:24:43 crc kubenswrapper[4950]: I0318 20:24:43.918811 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ccf85c649-grdqp"] Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.046242 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-9wnnb"] Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.156517 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.211069 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-n55x2"] Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.212108 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-n55x2" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.219890 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-f5c3-account-create-update-m5jc8"] Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.220926 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f5c3-account-create-update-m5jc8" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.222851 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.233217 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-n55x2"] Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.247553 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f5c3-account-create-update-m5jc8"] Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.265302 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8f7b4ae-81da-49ac-978c-d6e0ff3c839e-operator-scripts\") pod \"keystone-db-create-n55x2\" (UID: \"e8f7b4ae-81da-49ac-978c-d6e0ff3c839e\") " pod="openstack/keystone-db-create-n55x2" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.265355 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px89b\" (UniqueName: \"kubernetes.io/projected/e8f7b4ae-81da-49ac-978c-d6e0ff3c839e-kube-api-access-px89b\") pod \"keystone-db-create-n55x2\" (UID: \"e8f7b4ae-81da-49ac-978c-d6e0ff3c839e\") " pod="openstack/keystone-db-create-n55x2" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.268778 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f697c8bff-wvwhn"] Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.293744 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-mz5gw"] Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.299039 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mz5gw" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.318726 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-mz5gw"] Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.369578 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01a1bd17-8d39-4435-abe7-15617b6566ec-operator-scripts\") pod \"placement-db-create-mz5gw\" (UID: \"01a1bd17-8d39-4435-abe7-15617b6566ec\") " pod="openstack/placement-db-create-mz5gw" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.369661 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8f7b4ae-81da-49ac-978c-d6e0ff3c839e-operator-scripts\") pod \"keystone-db-create-n55x2\" (UID: \"e8f7b4ae-81da-49ac-978c-d6e0ff3c839e\") " pod="openstack/keystone-db-create-n55x2" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.369714 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px89b\" (UniqueName: \"kubernetes.io/projected/e8f7b4ae-81da-49ac-978c-d6e0ff3c839e-kube-api-access-px89b\") pod \"keystone-db-create-n55x2\" (UID: \"e8f7b4ae-81da-49ac-978c-d6e0ff3c839e\") " pod="openstack/keystone-db-create-n55x2" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.370422 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsdc6\" (UniqueName: \"kubernetes.io/projected/01a1bd17-8d39-4435-abe7-15617b6566ec-kube-api-access-tsdc6\") pod \"placement-db-create-mz5gw\" (UID: \"01a1bd17-8d39-4435-abe7-15617b6566ec\") " pod="openstack/placement-db-create-mz5gw" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.370646 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58d9f568-de35-40a7-be22-4ab9e138791f-operator-scripts\") pod \"keystone-f5c3-account-create-update-m5jc8\" (UID: \"58d9f568-de35-40a7-be22-4ab9e138791f\") " pod="openstack/keystone-f5c3-account-create-update-m5jc8" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.370716 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d8bz\" (UniqueName: \"kubernetes.io/projected/58d9f568-de35-40a7-be22-4ab9e138791f-kube-api-access-6d8bz\") pod \"keystone-f5c3-account-create-update-m5jc8\" (UID: \"58d9f568-de35-40a7-be22-4ab9e138791f\") " pod="openstack/keystone-f5c3-account-create-update-m5jc8" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.374103 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8f7b4ae-81da-49ac-978c-d6e0ff3c839e-operator-scripts\") pod \"keystone-db-create-n55x2\" (UID: \"e8f7b4ae-81da-49ac-978c-d6e0ff3c839e\") " pod="openstack/keystone-db-create-n55x2" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.392109 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px89b\" (UniqueName: \"kubernetes.io/projected/e8f7b4ae-81da-49ac-978c-d6e0ff3c839e-kube-api-access-px89b\") pod \"keystone-db-create-n55x2\" (UID: \"e8f7b4ae-81da-49ac-978c-d6e0ff3c839e\") " pod="openstack/keystone-db-create-n55x2" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.399966 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-cefd-account-create-update-ppvt2"] Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.401010 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-cefd-account-create-update-ppvt2" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.407753 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.409524 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-cefd-account-create-update-ppvt2"] Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.471703 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d8bz\" (UniqueName: \"kubernetes.io/projected/58d9f568-de35-40a7-be22-4ab9e138791f-kube-api-access-6d8bz\") pod \"keystone-f5c3-account-create-update-m5jc8\" (UID: \"58d9f568-de35-40a7-be22-4ab9e138791f\") " pod="openstack/keystone-f5c3-account-create-update-m5jc8" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.471752 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01a1bd17-8d39-4435-abe7-15617b6566ec-operator-scripts\") pod \"placement-db-create-mz5gw\" (UID: \"01a1bd17-8d39-4435-abe7-15617b6566ec\") " pod="openstack/placement-db-create-mz5gw" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.471830 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsdc6\" (UniqueName: \"kubernetes.io/projected/01a1bd17-8d39-4435-abe7-15617b6566ec-kube-api-access-tsdc6\") pod \"placement-db-create-mz5gw\" (UID: \"01a1bd17-8d39-4435-abe7-15617b6566ec\") " pod="openstack/placement-db-create-mz5gw" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.471863 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9dkw\" (UniqueName: \"kubernetes.io/projected/44e46c00-4ccb-4739-89c2-df4872915c1f-kube-api-access-g9dkw\") pod \"placement-cefd-account-create-update-ppvt2\" (UID: \"44e46c00-4ccb-4739-89c2-df4872915c1f\") " pod="openstack/placement-cefd-account-create-update-ppvt2" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.471889 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44e46c00-4ccb-4739-89c2-df4872915c1f-operator-scripts\") pod \"placement-cefd-account-create-update-ppvt2\" (UID: \"44e46c00-4ccb-4739-89c2-df4872915c1f\") " pod="openstack/placement-cefd-account-create-update-ppvt2" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.471912 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58d9f568-de35-40a7-be22-4ab9e138791f-operator-scripts\") pod \"keystone-f5c3-account-create-update-m5jc8\" (UID: \"58d9f568-de35-40a7-be22-4ab9e138791f\") " pod="openstack/keystone-f5c3-account-create-update-m5jc8" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.472715 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58d9f568-de35-40a7-be22-4ab9e138791f-operator-scripts\") pod \"keystone-f5c3-account-create-update-m5jc8\" (UID: \"58d9f568-de35-40a7-be22-4ab9e138791f\") " pod="openstack/keystone-f5c3-account-create-update-m5jc8" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.473353 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01a1bd17-8d39-4435-abe7-15617b6566ec-operator-scripts\") pod \"placement-db-create-mz5gw\" (UID: \"01a1bd17-8d39-4435-abe7-15617b6566ec\") " pod="openstack/placement-db-create-mz5gw" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.486430 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsdc6\" (UniqueName: \"kubernetes.io/projected/01a1bd17-8d39-4435-abe7-15617b6566ec-kube-api-access-tsdc6\") pod \"placement-db-create-mz5gw\" (UID: \"01a1bd17-8d39-4435-abe7-15617b6566ec\") " pod="openstack/placement-db-create-mz5gw" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.490665 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d8bz\" (UniqueName: \"kubernetes.io/projected/58d9f568-de35-40a7-be22-4ab9e138791f-kube-api-access-6d8bz\") pod \"keystone-f5c3-account-create-update-m5jc8\" (UID: \"58d9f568-de35-40a7-be22-4ab9e138791f\") " pod="openstack/keystone-f5c3-account-create-update-m5jc8" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.526276 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a3a71276-b0fa-481f-a91e-551d017c6462","Type":"ContainerStarted","Data":"31a17504ce8b0965d811d4899a4f5e2d1e78666042fa43811625558d4b2a52c2"} Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.527767 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-9wnnb" event={"ID":"3f70fe2d-f600-42da-8a3d-a4f814374ca7","Type":"ContainerStarted","Data":"3617084f56757f539a9f05a73550a528fb7faa8ccf87a09297d09e510f29022a"} Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.529129 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" event={"ID":"c4f9eed1-5dc4-4e60-b35a-812bc273acc6","Type":"ContainerStarted","Data":"9db2c9aed0821ee092af95f77e7005ba1c97396fe3a65ad17f620a7e1ee880ac"} Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.530103 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658f55c9f5-4qhbd" event={"ID":"0e5ae697-1746-49b5-ac19-d0b5bf951ce9","Type":"ContainerDied","Data":"295f99fd18e8102be9c12b437eaca3684d52971979e0973a4bc4dcc1c9dd917e"} Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.530185 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658f55c9f5-4qhbd" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.531965 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" event={"ID":"fbb60200-f508-4614-baa8-5ca52e7d0127","Type":"ContainerStarted","Data":"33a345e55df35557abfd092f184e6bbc77304e7b572d8f30220f41dd25980d90"} Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.548958 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-n55x2" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.562484 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f5c3-account-create-update-m5jc8" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.572857 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9dkw\" (UniqueName: \"kubernetes.io/projected/44e46c00-4ccb-4739-89c2-df4872915c1f-kube-api-access-g9dkw\") pod \"placement-cefd-account-create-update-ppvt2\" (UID: \"44e46c00-4ccb-4739-89c2-df4872915c1f\") " pod="openstack/placement-cefd-account-create-update-ppvt2" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.572908 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44e46c00-4ccb-4739-89c2-df4872915c1f-operator-scripts\") pod \"placement-cefd-account-create-update-ppvt2\" (UID: \"44e46c00-4ccb-4739-89c2-df4872915c1f\") " pod="openstack/placement-cefd-account-create-update-ppvt2" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.574131 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44e46c00-4ccb-4739-89c2-df4872915c1f-operator-scripts\") pod \"placement-cefd-account-create-update-ppvt2\" (UID: \"44e46c00-4ccb-4739-89c2-df4872915c1f\") " pod="openstack/placement-cefd-account-create-update-ppvt2" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.585999 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-658f55c9f5-4qhbd"] Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.599492 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-658f55c9f5-4qhbd"] Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.599683 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9dkw\" (UniqueName: \"kubernetes.io/projected/44e46c00-4ccb-4739-89c2-df4872915c1f-kube-api-access-g9dkw\") pod \"placement-cefd-account-create-update-ppvt2\" (UID: \"44e46c00-4ccb-4739-89c2-df4872915c1f\") " pod="openstack/placement-cefd-account-create-update-ppvt2" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.619666 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mz5gw" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.718270 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-cefd-account-create-update-ppvt2" Mar 18 20:24:44 crc kubenswrapper[4950]: I0318 20:24:44.724278 4950 scope.go:117] "RemoveContainer" containerID="b9e558de791450ddb096532acb3ce036f8ba7ae1f3018bd784ce268015e46df9" Mar 18 20:24:45 crc kubenswrapper[4950]: I0318 20:24:45.041507 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-n55x2"] Mar 18 20:24:45 crc kubenswrapper[4950]: I0318 20:24:45.118093 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-mz5gw"] Mar 18 20:24:45 crc kubenswrapper[4950]: I0318 20:24:45.131483 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f5c3-account-create-update-m5jc8"] Mar 18 20:24:45 crc kubenswrapper[4950]: I0318 20:24:45.182779 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-cefd-account-create-update-ppvt2"] Mar 18 20:24:45 crc kubenswrapper[4950]: W0318 20:24:45.191481 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod44e46c00_4ccb_4739_89c2_df4872915c1f.slice/crio-8eb075c96cd92e2d927b35196e14d6aadb6eb3b4a771f5fc5c52a949d09ecd23 WatchSource:0}: Error finding container 8eb075c96cd92e2d927b35196e14d6aadb6eb3b4a771f5fc5c52a949d09ecd23: Status 404 returned error can't find the container with id 8eb075c96cd92e2d927b35196e14d6aadb6eb3b4a771f5fc5c52a949d09ecd23 Mar 18 20:24:45 crc kubenswrapper[4950]: I0318 20:24:45.284299 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 18 20:24:45 crc kubenswrapper[4950]: I0318 20:24:45.491342 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e5ae697-1746-49b5-ac19-d0b5bf951ce9" path="/var/lib/kubelet/pods/0e5ae697-1746-49b5-ac19-d0b5bf951ce9/volumes" Mar 18 20:24:45 crc kubenswrapper[4950]: I0318 20:24:45.491919 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf74b193-b765-42b9-883e-951eb9f12a40" path="/var/lib/kubelet/pods/cf74b193-b765-42b9-883e-951eb9f12a40/volumes" Mar 18 20:24:45 crc kubenswrapper[4950]: I0318 20:24:45.540074 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-mz5gw" event={"ID":"01a1bd17-8d39-4435-abe7-15617b6566ec","Type":"ContainerStarted","Data":"db0c0cf8e8d75390a9bff38069b6e5f3d7b88fad4a378503459b1981e56fd793"} Mar 18 20:24:45 crc kubenswrapper[4950]: I0318 20:24:45.541198 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-n55x2" event={"ID":"e8f7b4ae-81da-49ac-978c-d6e0ff3c839e","Type":"ContainerStarted","Data":"b8a22a7fa3a1ba283ee903e9a352bcec89338d6cb05dedabbfae2b4f60149abd"} Mar 18 20:24:45 crc kubenswrapper[4950]: I0318 20:24:45.543130 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-cefd-account-create-update-ppvt2" event={"ID":"44e46c00-4ccb-4739-89c2-df4872915c1f","Type":"ContainerStarted","Data":"8eb075c96cd92e2d927b35196e14d6aadb6eb3b4a771f5fc5c52a949d09ecd23"} Mar 18 20:24:45 crc kubenswrapper[4950]: I0318 20:24:45.544197 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f5c3-account-create-update-m5jc8" event={"ID":"58d9f568-de35-40a7-be22-4ab9e138791f","Type":"ContainerStarted","Data":"9c9cc0e540bee7786d6341c309c51a55b6e55bb02a31b0eb3caacf98a9935004"} Mar 18 20:24:45 crc kubenswrapper[4950]: I0318 20:24:45.779593 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:45 crc kubenswrapper[4950]: I0318 20:24:45.882587 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Mar 18 20:24:46 crc kubenswrapper[4950]: I0318 20:24:46.559220 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-9wnnb" event={"ID":"3f70fe2d-f600-42da-8a3d-a4f814374ca7","Type":"ContainerStarted","Data":"c7e1c3fab25dd3a45eabd8d1cac3b373f8b22ca4d581d29983b6f5119a4a446f"} Mar 18 20:24:46 crc kubenswrapper[4950]: I0318 20:24:46.584232 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-9wnnb" podStartSLOduration=4.584209997 podStartE2EDuration="4.584209997s" podCreationTimestamp="2026-03-18 20:24:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:24:46.579530934 +0000 UTC m=+1099.820372812" watchObservedRunningTime="2026-03-18 20:24:46.584209997 +0000 UTC m=+1099.825051865" Mar 18 20:24:47 crc kubenswrapper[4950]: I0318 20:24:47.569516 4950 generic.go:334] "Generic (PLEG): container finished" podID="fbb60200-f508-4614-baa8-5ca52e7d0127" containerID="c7135d56ed96b30b2c405ec1d8a4f772aedb0560266dee571056767b258309e0" exitCode=0 Mar 18 20:24:47 crc kubenswrapper[4950]: I0318 20:24:47.569955 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" event={"ID":"fbb60200-f508-4614-baa8-5ca52e7d0127","Type":"ContainerDied","Data":"c7135d56ed96b30b2c405ec1d8a4f772aedb0560266dee571056767b258309e0"} Mar 18 20:24:47 crc kubenswrapper[4950]: I0318 20:24:47.586728 4950 generic.go:334] "Generic (PLEG): container finished" podID="e8f7b4ae-81da-49ac-978c-d6e0ff3c839e" containerID="a9c4f36c3ddf85aa65683808380cd3ce1c2813d7594b3d5caf33ff0deb12b3fe" exitCode=0 Mar 18 20:24:47 crc kubenswrapper[4950]: I0318 20:24:47.586853 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-n55x2" event={"ID":"e8f7b4ae-81da-49ac-978c-d6e0ff3c839e","Type":"ContainerDied","Data":"a9c4f36c3ddf85aa65683808380cd3ce1c2813d7594b3d5caf33ff0deb12b3fe"} Mar 18 20:24:47 crc kubenswrapper[4950]: I0318 20:24:47.591999 4950 generic.go:334] "Generic (PLEG): container finished" podID="58d9f568-de35-40a7-be22-4ab9e138791f" containerID="c80f2b4f8af07b87c741978f2d3efdb10ca356b077fe4071a9bb0750ebb7b4ad" exitCode=0 Mar 18 20:24:47 crc kubenswrapper[4950]: I0318 20:24:47.592182 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f5c3-account-create-update-m5jc8" event={"ID":"58d9f568-de35-40a7-be22-4ab9e138791f","Type":"ContainerDied","Data":"c80f2b4f8af07b87c741978f2d3efdb10ca356b077fe4071a9bb0750ebb7b4ad"} Mar 18 20:24:47 crc kubenswrapper[4950]: I0318 20:24:47.600568 4950 generic.go:334] "Generic (PLEG): container finished" podID="c4f9eed1-5dc4-4e60-b35a-812bc273acc6" containerID="343707eb2fcc094bab69e01c5cf3a14e2e573896ed1c5f1fc16552c5d5afb6aa" exitCode=0 Mar 18 20:24:47 crc kubenswrapper[4950]: I0318 20:24:47.600690 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" event={"ID":"c4f9eed1-5dc4-4e60-b35a-812bc273acc6","Type":"ContainerDied","Data":"343707eb2fcc094bab69e01c5cf3a14e2e573896ed1c5f1fc16552c5d5afb6aa"} Mar 18 20:24:47 crc kubenswrapper[4950]: I0318 20:24:47.610748 4950 generic.go:334] "Generic (PLEG): container finished" podID="44e46c00-4ccb-4739-89c2-df4872915c1f" containerID="66e79885253efa65a0435682ed0a52ce59b093fa665af1665f584db420ccd084" exitCode=0 Mar 18 20:24:47 crc kubenswrapper[4950]: I0318 20:24:47.610803 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-cefd-account-create-update-ppvt2" event={"ID":"44e46c00-4ccb-4739-89c2-df4872915c1f","Type":"ContainerDied","Data":"66e79885253efa65a0435682ed0a52ce59b093fa665af1665f584db420ccd084"} Mar 18 20:24:47 crc kubenswrapper[4950]: I0318 20:24:47.613668 4950 generic.go:334] "Generic (PLEG): container finished" podID="01a1bd17-8d39-4435-abe7-15617b6566ec" containerID="3134bc10c0a641b70ab56d0f32141528cb9d14dc72b09ac5bed072548aee33b4" exitCode=0 Mar 18 20:24:47 crc kubenswrapper[4950]: I0318 20:24:47.613719 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-mz5gw" event={"ID":"01a1bd17-8d39-4435-abe7-15617b6566ec","Type":"ContainerDied","Data":"3134bc10c0a641b70ab56d0f32141528cb9d14dc72b09ac5bed072548aee33b4"} Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.172394 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.356746 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-2kh85"] Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.358613 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2kh85" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.379186 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-2kh85"] Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.444363 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9cxp\" (UniqueName: \"kubernetes.io/projected/c89c3ec6-2109-4bd7-92e4-5d872dc95cca-kube-api-access-c9cxp\") pod \"glance-db-create-2kh85\" (UID: \"c89c3ec6-2109-4bd7-92e4-5d872dc95cca\") " pod="openstack/glance-db-create-2kh85" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.444647 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c89c3ec6-2109-4bd7-92e4-5d872dc95cca-operator-scripts\") pod \"glance-db-create-2kh85\" (UID: \"c89c3ec6-2109-4bd7-92e4-5d872dc95cca\") " pod="openstack/glance-db-create-2kh85" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.447452 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-aa88-account-create-update-78r4v"] Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.448338 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-aa88-account-create-update-78r4v" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.450101 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.458442 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-aa88-account-create-update-78r4v"] Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.546871 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99f9366a-2413-4243-9196-95088cf66e87-operator-scripts\") pod \"glance-aa88-account-create-update-78r4v\" (UID: \"99f9366a-2413-4243-9196-95088cf66e87\") " pod="openstack/glance-aa88-account-create-update-78r4v" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.546953 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c89c3ec6-2109-4bd7-92e4-5d872dc95cca-operator-scripts\") pod \"glance-db-create-2kh85\" (UID: \"c89c3ec6-2109-4bd7-92e4-5d872dc95cca\") " pod="openstack/glance-db-create-2kh85" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.547184 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmjn9\" (UniqueName: \"kubernetes.io/projected/99f9366a-2413-4243-9196-95088cf66e87-kube-api-access-kmjn9\") pod \"glance-aa88-account-create-update-78r4v\" (UID: \"99f9366a-2413-4243-9196-95088cf66e87\") " pod="openstack/glance-aa88-account-create-update-78r4v" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.547394 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9cxp\" (UniqueName: \"kubernetes.io/projected/c89c3ec6-2109-4bd7-92e4-5d872dc95cca-kube-api-access-c9cxp\") pod \"glance-db-create-2kh85\" (UID: \"c89c3ec6-2109-4bd7-92e4-5d872dc95cca\") " pod="openstack/glance-db-create-2kh85" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.547802 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c89c3ec6-2109-4bd7-92e4-5d872dc95cca-operator-scripts\") pod \"glance-db-create-2kh85\" (UID: \"c89c3ec6-2109-4bd7-92e4-5d872dc95cca\") " pod="openstack/glance-db-create-2kh85" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.566025 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9cxp\" (UniqueName: \"kubernetes.io/projected/c89c3ec6-2109-4bd7-92e4-5d872dc95cca-kube-api-access-c9cxp\") pod \"glance-db-create-2kh85\" (UID: \"c89c3ec6-2109-4bd7-92e4-5d872dc95cca\") " pod="openstack/glance-db-create-2kh85" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.624370 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a3a71276-b0fa-481f-a91e-551d017c6462","Type":"ContainerStarted","Data":"9a89fd2cb21de57cb3e14767b531e67b79728634b77ede53ec3d669d713eca80"} Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.624450 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a3a71276-b0fa-481f-a91e-551d017c6462","Type":"ContainerStarted","Data":"be404e181cee179853aa294202da5db60ab8da3a4b815599ac50dd9b64440533"} Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.624578 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.627990 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" event={"ID":"c4f9eed1-5dc4-4e60-b35a-812bc273acc6","Type":"ContainerStarted","Data":"35ce87ad0e96b6e3da14255f5539d36606cf3e9e498844b9b1bea2ba280a43b6"} Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.628202 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.630098 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" event={"ID":"fbb60200-f508-4614-baa8-5ca52e7d0127","Type":"ContainerStarted","Data":"ce882643ed3e35fe169671ba1e98531fc1b2385502a138f584dbcb09c8719d65"} Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.630124 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.650664 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99f9366a-2413-4243-9196-95088cf66e87-operator-scripts\") pod \"glance-aa88-account-create-update-78r4v\" (UID: \"99f9366a-2413-4243-9196-95088cf66e87\") " pod="openstack/glance-aa88-account-create-update-78r4v" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.650852 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmjn9\" (UniqueName: \"kubernetes.io/projected/99f9366a-2413-4243-9196-95088cf66e87-kube-api-access-kmjn9\") pod \"glance-aa88-account-create-update-78r4v\" (UID: \"99f9366a-2413-4243-9196-95088cf66e87\") " pod="openstack/glance-aa88-account-create-update-78r4v" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.654260 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99f9366a-2413-4243-9196-95088cf66e87-operator-scripts\") pod \"glance-aa88-account-create-update-78r4v\" (UID: \"99f9366a-2413-4243-9196-95088cf66e87\") " pod="openstack/glance-aa88-account-create-update-78r4v" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.668325 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.288857399 podStartE2EDuration="5.668301691s" podCreationTimestamp="2026-03-18 20:24:43 +0000 UTC" firstStartedPulling="2026-03-18 20:24:44.161284069 +0000 UTC m=+1097.402125937" lastFinishedPulling="2026-03-18 20:24:47.540728351 +0000 UTC m=+1100.781570229" observedRunningTime="2026-03-18 20:24:48.644368592 +0000 UTC m=+1101.885210460" watchObservedRunningTime="2026-03-18 20:24:48.668301691 +0000 UTC m=+1101.909143559" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.673547 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmjn9\" (UniqueName: \"kubernetes.io/projected/99f9366a-2413-4243-9196-95088cf66e87-kube-api-access-kmjn9\") pod \"glance-aa88-account-create-update-78r4v\" (UID: \"99f9366a-2413-4243-9196-95088cf66e87\") " pod="openstack/glance-aa88-account-create-update-78r4v" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.675252 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2kh85" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.679896 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" podStartSLOduration=3.163956663 podStartE2EDuration="5.679884105s" podCreationTimestamp="2026-03-18 20:24:43 +0000 UTC" firstStartedPulling="2026-03-18 20:24:44.24844932 +0000 UTC m=+1097.489291178" lastFinishedPulling="2026-03-18 20:24:46.764376752 +0000 UTC m=+1100.005218620" observedRunningTime="2026-03-18 20:24:48.664583163 +0000 UTC m=+1101.905425031" watchObservedRunningTime="2026-03-18 20:24:48.679884105 +0000 UTC m=+1101.920725973" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.764806 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-aa88-account-create-update-78r4v" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.931434 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mz5gw" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.950269 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" podStartSLOduration=4.101206914 podStartE2EDuration="6.95025404s" podCreationTimestamp="2026-03-18 20:24:42 +0000 UTC" firstStartedPulling="2026-03-18 20:24:43.914227237 +0000 UTC m=+1097.155069105" lastFinishedPulling="2026-03-18 20:24:46.763274363 +0000 UTC m=+1100.004116231" observedRunningTime="2026-03-18 20:24:48.695360892 +0000 UTC m=+1101.936202760" watchObservedRunningTime="2026-03-18 20:24:48.95025404 +0000 UTC m=+1102.191095908" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.954309 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsdc6\" (UniqueName: \"kubernetes.io/projected/01a1bd17-8d39-4435-abe7-15617b6566ec-kube-api-access-tsdc6\") pod \"01a1bd17-8d39-4435-abe7-15617b6566ec\" (UID: \"01a1bd17-8d39-4435-abe7-15617b6566ec\") " Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.954396 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01a1bd17-8d39-4435-abe7-15617b6566ec-operator-scripts\") pod \"01a1bd17-8d39-4435-abe7-15617b6566ec\" (UID: \"01a1bd17-8d39-4435-abe7-15617b6566ec\") " Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.955319 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01a1bd17-8d39-4435-abe7-15617b6566ec-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "01a1bd17-8d39-4435-abe7-15617b6566ec" (UID: "01a1bd17-8d39-4435-abe7-15617b6566ec"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:24:48 crc kubenswrapper[4950]: I0318 20:24:48.969333 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01a1bd17-8d39-4435-abe7-15617b6566ec-kube-api-access-tsdc6" (OuterVolumeSpecName: "kube-api-access-tsdc6") pod "01a1bd17-8d39-4435-abe7-15617b6566ec" (UID: "01a1bd17-8d39-4435-abe7-15617b6566ec"). InnerVolumeSpecName "kube-api-access-tsdc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.056643 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsdc6\" (UniqueName: \"kubernetes.io/projected/01a1bd17-8d39-4435-abe7-15617b6566ec-kube-api-access-tsdc6\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.056672 4950 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01a1bd17-8d39-4435-abe7-15617b6566ec-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.100042 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f5c3-account-create-update-m5jc8" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.101286 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-n55x2" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.126825 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-cefd-account-create-update-ppvt2" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.158135 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9dkw\" (UniqueName: \"kubernetes.io/projected/44e46c00-4ccb-4739-89c2-df4872915c1f-kube-api-access-g9dkw\") pod \"44e46c00-4ccb-4739-89c2-df4872915c1f\" (UID: \"44e46c00-4ccb-4739-89c2-df4872915c1f\") " Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.158196 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58d9f568-de35-40a7-be22-4ab9e138791f-operator-scripts\") pod \"58d9f568-de35-40a7-be22-4ab9e138791f\" (UID: \"58d9f568-de35-40a7-be22-4ab9e138791f\") " Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.158230 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8f7b4ae-81da-49ac-978c-d6e0ff3c839e-operator-scripts\") pod \"e8f7b4ae-81da-49ac-978c-d6e0ff3c839e\" (UID: \"e8f7b4ae-81da-49ac-978c-d6e0ff3c839e\") " Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.158329 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px89b\" (UniqueName: \"kubernetes.io/projected/e8f7b4ae-81da-49ac-978c-d6e0ff3c839e-kube-api-access-px89b\") pod \"e8f7b4ae-81da-49ac-978c-d6e0ff3c839e\" (UID: \"e8f7b4ae-81da-49ac-978c-d6e0ff3c839e\") " Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.158393 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44e46c00-4ccb-4739-89c2-df4872915c1f-operator-scripts\") pod \"44e46c00-4ccb-4739-89c2-df4872915c1f\" (UID: \"44e46c00-4ccb-4739-89c2-df4872915c1f\") " Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.158447 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6d8bz\" (UniqueName: \"kubernetes.io/projected/58d9f568-de35-40a7-be22-4ab9e138791f-kube-api-access-6d8bz\") pod \"58d9f568-de35-40a7-be22-4ab9e138791f\" (UID: \"58d9f568-de35-40a7-be22-4ab9e138791f\") " Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.161555 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44e46c00-4ccb-4739-89c2-df4872915c1f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "44e46c00-4ccb-4739-89c2-df4872915c1f" (UID: "44e46c00-4ccb-4739-89c2-df4872915c1f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.161809 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58d9f568-de35-40a7-be22-4ab9e138791f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "58d9f568-de35-40a7-be22-4ab9e138791f" (UID: "58d9f568-de35-40a7-be22-4ab9e138791f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.162091 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8f7b4ae-81da-49ac-978c-d6e0ff3c839e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e8f7b4ae-81da-49ac-978c-d6e0ff3c839e" (UID: "e8f7b4ae-81da-49ac-978c-d6e0ff3c839e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.162646 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58d9f568-de35-40a7-be22-4ab9e138791f-kube-api-access-6d8bz" (OuterVolumeSpecName: "kube-api-access-6d8bz") pod "58d9f568-de35-40a7-be22-4ab9e138791f" (UID: "58d9f568-de35-40a7-be22-4ab9e138791f"). InnerVolumeSpecName "kube-api-access-6d8bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.164794 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44e46c00-4ccb-4739-89c2-df4872915c1f-kube-api-access-g9dkw" (OuterVolumeSpecName: "kube-api-access-g9dkw") pod "44e46c00-4ccb-4739-89c2-df4872915c1f" (UID: "44e46c00-4ccb-4739-89c2-df4872915c1f"). InnerVolumeSpecName "kube-api-access-g9dkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.166462 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8f7b4ae-81da-49ac-978c-d6e0ff3c839e-kube-api-access-px89b" (OuterVolumeSpecName: "kube-api-access-px89b") pod "e8f7b4ae-81da-49ac-978c-d6e0ff3c839e" (UID: "e8f7b4ae-81da-49ac-978c-d6e0ff3c839e"). InnerVolumeSpecName "kube-api-access-px89b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.261177 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6d8bz\" (UniqueName: \"kubernetes.io/projected/58d9f568-de35-40a7-be22-4ab9e138791f-kube-api-access-6d8bz\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.261493 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9dkw\" (UniqueName: \"kubernetes.io/projected/44e46c00-4ccb-4739-89c2-df4872915c1f-kube-api-access-g9dkw\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.261557 4950 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58d9f568-de35-40a7-be22-4ab9e138791f-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.261631 4950 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8f7b4ae-81da-49ac-978c-d6e0ff3c839e-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.261693 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px89b\" (UniqueName: \"kubernetes.io/projected/e8f7b4ae-81da-49ac-978c-d6e0ff3c839e-kube-api-access-px89b\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.261753 4950 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44e46c00-4ccb-4739-89c2-df4872915c1f-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.322640 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-aa88-account-create-update-78r4v"] Mar 18 20:24:49 crc kubenswrapper[4950]: W0318 20:24:49.326559 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99f9366a_2413_4243_9196_95088cf66e87.slice/crio-a1283889fcabf14984cf128157eaaf20ff893e88a08ecfd4b8762b11173c04e5 WatchSource:0}: Error finding container a1283889fcabf14984cf128157eaaf20ff893e88a08ecfd4b8762b11173c04e5: Status 404 returned error can't find the container with id a1283889fcabf14984cf128157eaaf20ff893e88a08ecfd4b8762b11173c04e5 Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.339238 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-2kh85"] Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.638050 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f5c3-account-create-update-m5jc8" event={"ID":"58d9f568-de35-40a7-be22-4ab9e138791f","Type":"ContainerDied","Data":"9c9cc0e540bee7786d6341c309c51a55b6e55bb02a31b0eb3caacf98a9935004"} Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.638308 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c9cc0e540bee7786d6341c309c51a55b6e55bb02a31b0eb3caacf98a9935004" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.638368 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f5c3-account-create-update-m5jc8" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.641168 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-aa88-account-create-update-78r4v" event={"ID":"99f9366a-2413-4243-9196-95088cf66e87","Type":"ContainerStarted","Data":"780cfeaa8a4e5f4288db988fa84b84f89bfc27c5e2b1268a00628c29be67c890"} Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.641191 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-aa88-account-create-update-78r4v" event={"ID":"99f9366a-2413-4243-9196-95088cf66e87","Type":"ContainerStarted","Data":"a1283889fcabf14984cf128157eaaf20ff893e88a08ecfd4b8762b11173c04e5"} Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.642623 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-2kh85" event={"ID":"c89c3ec6-2109-4bd7-92e4-5d872dc95cca","Type":"ContainerStarted","Data":"a302c9bdde190529bc585b2f5e9e741d7c3462d6c1e115bfbc5162fcfaf814e9"} Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.642660 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-2kh85" event={"ID":"c89c3ec6-2109-4bd7-92e4-5d872dc95cca","Type":"ContainerStarted","Data":"5b17a4426a191aa71c5a0029fb7d761078d87897ba5ebc09743c28218e6b6dc3"} Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.643948 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-cefd-account-create-update-ppvt2" event={"ID":"44e46c00-4ccb-4739-89c2-df4872915c1f","Type":"ContainerDied","Data":"8eb075c96cd92e2d927b35196e14d6aadb6eb3b4a771f5fc5c52a949d09ecd23"} Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.643971 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8eb075c96cd92e2d927b35196e14d6aadb6eb3b4a771f5fc5c52a949d09ecd23" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.644020 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-cefd-account-create-update-ppvt2" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.645603 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-mz5gw" event={"ID":"01a1bd17-8d39-4435-abe7-15617b6566ec","Type":"ContainerDied","Data":"db0c0cf8e8d75390a9bff38069b6e5f3d7b88fad4a378503459b1981e56fd793"} Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.645624 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db0c0cf8e8d75390a9bff38069b6e5f3d7b88fad4a378503459b1981e56fd793" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.645666 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mz5gw" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.647268 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-n55x2" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.647582 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-n55x2" event={"ID":"e8f7b4ae-81da-49ac-978c-d6e0ff3c839e","Type":"ContainerDied","Data":"b8a22a7fa3a1ba283ee903e9a352bcec89338d6cb05dedabbfae2b4f60149abd"} Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.647600 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8a22a7fa3a1ba283ee903e9a352bcec89338d6cb05dedabbfae2b4f60149abd" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.659802 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-aa88-account-create-update-78r4v" podStartSLOduration=1.659783344 podStartE2EDuration="1.659783344s" podCreationTimestamp="2026-03-18 20:24:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:24:49.653334598 +0000 UTC m=+1102.894176466" watchObservedRunningTime="2026-03-18 20:24:49.659783344 +0000 UTC m=+1102.900625212" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.966980 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-2kh85" podStartSLOduration=1.9669577230000002 podStartE2EDuration="1.966957723s" podCreationTimestamp="2026-03-18 20:24:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:24:49.678175617 +0000 UTC m=+1102.919017485" watchObservedRunningTime="2026-03-18 20:24:49.966957723 +0000 UTC m=+1103.207799591" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.969102 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-7bzvq"] Mar 18 20:24:49 crc kubenswrapper[4950]: E0318 20:24:49.969662 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8f7b4ae-81da-49ac-978c-d6e0ff3c839e" containerName="mariadb-database-create" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.969780 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8f7b4ae-81da-49ac-978c-d6e0ff3c839e" containerName="mariadb-database-create" Mar 18 20:24:49 crc kubenswrapper[4950]: E0318 20:24:49.969867 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01a1bd17-8d39-4435-abe7-15617b6566ec" containerName="mariadb-database-create" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.969938 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="01a1bd17-8d39-4435-abe7-15617b6566ec" containerName="mariadb-database-create" Mar 18 20:24:49 crc kubenswrapper[4950]: E0318 20:24:49.970188 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44e46c00-4ccb-4739-89c2-df4872915c1f" containerName="mariadb-account-create-update" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.970251 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="44e46c00-4ccb-4739-89c2-df4872915c1f" containerName="mariadb-account-create-update" Mar 18 20:24:49 crc kubenswrapper[4950]: E0318 20:24:49.970344 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58d9f568-de35-40a7-be22-4ab9e138791f" containerName="mariadb-account-create-update" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.970445 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="58d9f568-de35-40a7-be22-4ab9e138791f" containerName="mariadb-account-create-update" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.970729 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="01a1bd17-8d39-4435-abe7-15617b6566ec" containerName="mariadb-database-create" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.970824 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="44e46c00-4ccb-4739-89c2-df4872915c1f" containerName="mariadb-account-create-update" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.970898 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8f7b4ae-81da-49ac-978c-d6e0ff3c839e" containerName="mariadb-database-create" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.970973 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="58d9f568-de35-40a7-be22-4ab9e138791f" containerName="mariadb-account-create-update" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.971717 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-7bzvq" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.974241 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Mar 18 20:24:49 crc kubenswrapper[4950]: I0318 20:24:49.991167 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-7bzvq"] Mar 18 20:24:50 crc kubenswrapper[4950]: I0318 20:24:50.081123 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0c57ee2-7be6-43f9-b2ce-7d912a74df28-operator-scripts\") pod \"root-account-create-update-7bzvq\" (UID: \"b0c57ee2-7be6-43f9-b2ce-7d912a74df28\") " pod="openstack/root-account-create-update-7bzvq" Mar 18 20:24:50 crc kubenswrapper[4950]: I0318 20:24:50.081482 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxb66\" (UniqueName: \"kubernetes.io/projected/b0c57ee2-7be6-43f9-b2ce-7d912a74df28-kube-api-access-rxb66\") pod \"root-account-create-update-7bzvq\" (UID: \"b0c57ee2-7be6-43f9-b2ce-7d912a74df28\") " pod="openstack/root-account-create-update-7bzvq" Mar 18 20:24:50 crc kubenswrapper[4950]: I0318 20:24:50.183269 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0c57ee2-7be6-43f9-b2ce-7d912a74df28-operator-scripts\") pod \"root-account-create-update-7bzvq\" (UID: \"b0c57ee2-7be6-43f9-b2ce-7d912a74df28\") " pod="openstack/root-account-create-update-7bzvq" Mar 18 20:24:50 crc kubenswrapper[4950]: I0318 20:24:50.183351 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxb66\" (UniqueName: \"kubernetes.io/projected/b0c57ee2-7be6-43f9-b2ce-7d912a74df28-kube-api-access-rxb66\") pod \"root-account-create-update-7bzvq\" (UID: \"b0c57ee2-7be6-43f9-b2ce-7d912a74df28\") " pod="openstack/root-account-create-update-7bzvq" Mar 18 20:24:50 crc kubenswrapper[4950]: I0318 20:24:50.184182 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0c57ee2-7be6-43f9-b2ce-7d912a74df28-operator-scripts\") pod \"root-account-create-update-7bzvq\" (UID: \"b0c57ee2-7be6-43f9-b2ce-7d912a74df28\") " pod="openstack/root-account-create-update-7bzvq" Mar 18 20:24:50 crc kubenswrapper[4950]: I0318 20:24:50.203757 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxb66\" (UniqueName: \"kubernetes.io/projected/b0c57ee2-7be6-43f9-b2ce-7d912a74df28-kube-api-access-rxb66\") pod \"root-account-create-update-7bzvq\" (UID: \"b0c57ee2-7be6-43f9-b2ce-7d912a74df28\") " pod="openstack/root-account-create-update-7bzvq" Mar 18 20:24:50 crc kubenswrapper[4950]: I0318 20:24:50.295068 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-7bzvq" Mar 18 20:24:50 crc kubenswrapper[4950]: I0318 20:24:50.654725 4950 generic.go:334] "Generic (PLEG): container finished" podID="c89c3ec6-2109-4bd7-92e4-5d872dc95cca" containerID="a302c9bdde190529bc585b2f5e9e741d7c3462d6c1e115bfbc5162fcfaf814e9" exitCode=0 Mar 18 20:24:50 crc kubenswrapper[4950]: I0318 20:24:50.654858 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-2kh85" event={"ID":"c89c3ec6-2109-4bd7-92e4-5d872dc95cca","Type":"ContainerDied","Data":"a302c9bdde190529bc585b2f5e9e741d7c3462d6c1e115bfbc5162fcfaf814e9"} Mar 18 20:24:50 crc kubenswrapper[4950]: I0318 20:24:50.656223 4950 generic.go:334] "Generic (PLEG): container finished" podID="99f9366a-2413-4243-9196-95088cf66e87" containerID="780cfeaa8a4e5f4288db988fa84b84f89bfc27c5e2b1268a00628c29be67c890" exitCode=0 Mar 18 20:24:50 crc kubenswrapper[4950]: I0318 20:24:50.656272 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-aa88-account-create-update-78r4v" event={"ID":"99f9366a-2413-4243-9196-95088cf66e87","Type":"ContainerDied","Data":"780cfeaa8a4e5f4288db988fa84b84f89bfc27c5e2b1268a00628c29be67c890"} Mar 18 20:24:50 crc kubenswrapper[4950]: I0318 20:24:50.762597 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-7bzvq"] Mar 18 20:24:51 crc kubenswrapper[4950]: I0318 20:24:51.687821 4950 generic.go:334] "Generic (PLEG): container finished" podID="b0c57ee2-7be6-43f9-b2ce-7d912a74df28" containerID="0aad004978d3ce977decea721b811314e837f4da8975d2e599ca41ff0b3339b0" exitCode=0 Mar 18 20:24:51 crc kubenswrapper[4950]: I0318 20:24:51.688825 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-7bzvq" event={"ID":"b0c57ee2-7be6-43f9-b2ce-7d912a74df28","Type":"ContainerDied","Data":"0aad004978d3ce977decea721b811314e837f4da8975d2e599ca41ff0b3339b0"} Mar 18 20:24:51 crc kubenswrapper[4950]: I0318 20:24:51.688857 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-7bzvq" event={"ID":"b0c57ee2-7be6-43f9-b2ce-7d912a74df28","Type":"ContainerStarted","Data":"4c71b5336938814145532eaa5d66145368592a4243d3a6e9e87440728197a926"} Mar 18 20:24:52 crc kubenswrapper[4950]: I0318 20:24:52.063782 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2kh85" Mar 18 20:24:52 crc kubenswrapper[4950]: I0318 20:24:52.070328 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-aa88-account-create-update-78r4v" Mar 18 20:24:52 crc kubenswrapper[4950]: I0318 20:24:52.156387 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9cxp\" (UniqueName: \"kubernetes.io/projected/c89c3ec6-2109-4bd7-92e4-5d872dc95cca-kube-api-access-c9cxp\") pod \"c89c3ec6-2109-4bd7-92e4-5d872dc95cca\" (UID: \"c89c3ec6-2109-4bd7-92e4-5d872dc95cca\") " Mar 18 20:24:52 crc kubenswrapper[4950]: I0318 20:24:52.156711 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c89c3ec6-2109-4bd7-92e4-5d872dc95cca-operator-scripts\") pod \"c89c3ec6-2109-4bd7-92e4-5d872dc95cca\" (UID: \"c89c3ec6-2109-4bd7-92e4-5d872dc95cca\") " Mar 18 20:24:52 crc kubenswrapper[4950]: I0318 20:24:52.156752 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99f9366a-2413-4243-9196-95088cf66e87-operator-scripts\") pod \"99f9366a-2413-4243-9196-95088cf66e87\" (UID: \"99f9366a-2413-4243-9196-95088cf66e87\") " Mar 18 20:24:52 crc kubenswrapper[4950]: I0318 20:24:52.156771 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmjn9\" (UniqueName: \"kubernetes.io/projected/99f9366a-2413-4243-9196-95088cf66e87-kube-api-access-kmjn9\") pod \"99f9366a-2413-4243-9196-95088cf66e87\" (UID: \"99f9366a-2413-4243-9196-95088cf66e87\") " Mar 18 20:24:52 crc kubenswrapper[4950]: I0318 20:24:52.157304 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c89c3ec6-2109-4bd7-92e4-5d872dc95cca-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c89c3ec6-2109-4bd7-92e4-5d872dc95cca" (UID: "c89c3ec6-2109-4bd7-92e4-5d872dc95cca"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:24:52 crc kubenswrapper[4950]: I0318 20:24:52.157408 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99f9366a-2413-4243-9196-95088cf66e87-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "99f9366a-2413-4243-9196-95088cf66e87" (UID: "99f9366a-2413-4243-9196-95088cf66e87"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:24:52 crc kubenswrapper[4950]: I0318 20:24:52.162136 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c89c3ec6-2109-4bd7-92e4-5d872dc95cca-kube-api-access-c9cxp" (OuterVolumeSpecName: "kube-api-access-c9cxp") pod "c89c3ec6-2109-4bd7-92e4-5d872dc95cca" (UID: "c89c3ec6-2109-4bd7-92e4-5d872dc95cca"). InnerVolumeSpecName "kube-api-access-c9cxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:24:52 crc kubenswrapper[4950]: I0318 20:24:52.174147 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99f9366a-2413-4243-9196-95088cf66e87-kube-api-access-kmjn9" (OuterVolumeSpecName: "kube-api-access-kmjn9") pod "99f9366a-2413-4243-9196-95088cf66e87" (UID: "99f9366a-2413-4243-9196-95088cf66e87"). InnerVolumeSpecName "kube-api-access-kmjn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:24:52 crc kubenswrapper[4950]: I0318 20:24:52.258660 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9cxp\" (UniqueName: \"kubernetes.io/projected/c89c3ec6-2109-4bd7-92e4-5d872dc95cca-kube-api-access-c9cxp\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:52 crc kubenswrapper[4950]: I0318 20:24:52.258702 4950 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c89c3ec6-2109-4bd7-92e4-5d872dc95cca-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:52 crc kubenswrapper[4950]: I0318 20:24:52.258716 4950 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99f9366a-2413-4243-9196-95088cf66e87-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:52 crc kubenswrapper[4950]: I0318 20:24:52.258728 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmjn9\" (UniqueName: \"kubernetes.io/projected/99f9366a-2413-4243-9196-95088cf66e87-kube-api-access-kmjn9\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:52 crc kubenswrapper[4950]: I0318 20:24:52.698447 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-aa88-account-create-update-78r4v" event={"ID":"99f9366a-2413-4243-9196-95088cf66e87","Type":"ContainerDied","Data":"a1283889fcabf14984cf128157eaaf20ff893e88a08ecfd4b8762b11173c04e5"} Mar 18 20:24:52 crc kubenswrapper[4950]: I0318 20:24:52.698508 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1283889fcabf14984cf128157eaaf20ff893e88a08ecfd4b8762b11173c04e5" Mar 18 20:24:52 crc kubenswrapper[4950]: I0318 20:24:52.698545 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-aa88-account-create-update-78r4v" Mar 18 20:24:52 crc kubenswrapper[4950]: I0318 20:24:52.700189 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2kh85" Mar 18 20:24:52 crc kubenswrapper[4950]: I0318 20:24:52.700191 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-2kh85" event={"ID":"c89c3ec6-2109-4bd7-92e4-5d872dc95cca","Type":"ContainerDied","Data":"5b17a4426a191aa71c5a0029fb7d761078d87897ba5ebc09743c28218e6b6dc3"} Mar 18 20:24:52 crc kubenswrapper[4950]: I0318 20:24:52.700426 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b17a4426a191aa71c5a0029fb7d761078d87897ba5ebc09743c28218e6b6dc3" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.048666 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-7bzvq" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.145019 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.169468 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxb66\" (UniqueName: \"kubernetes.io/projected/b0c57ee2-7be6-43f9-b2ce-7d912a74df28-kube-api-access-rxb66\") pod \"b0c57ee2-7be6-43f9-b2ce-7d912a74df28\" (UID: \"b0c57ee2-7be6-43f9-b2ce-7d912a74df28\") " Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.169578 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0c57ee2-7be6-43f9-b2ce-7d912a74df28-operator-scripts\") pod \"b0c57ee2-7be6-43f9-b2ce-7d912a74df28\" (UID: \"b0c57ee2-7be6-43f9-b2ce-7d912a74df28\") " Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.170461 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0c57ee2-7be6-43f9-b2ce-7d912a74df28-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b0c57ee2-7be6-43f9-b2ce-7d912a74df28" (UID: "b0c57ee2-7be6-43f9-b2ce-7d912a74df28"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.171379 4950 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0c57ee2-7be6-43f9-b2ce-7d912a74df28-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.175296 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0c57ee2-7be6-43f9-b2ce-7d912a74df28-kube-api-access-rxb66" (OuterVolumeSpecName: "kube-api-access-rxb66") pod "b0c57ee2-7be6-43f9-b2ce-7d912a74df28" (UID: "b0c57ee2-7be6-43f9-b2ce-7d912a74df28"). InnerVolumeSpecName "kube-api-access-rxb66". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.273328 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxb66\" (UniqueName: \"kubernetes.io/projected/b0c57ee2-7be6-43f9-b2ce-7d912a74df28-kube-api-access-rxb66\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.566547 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.654510 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ccf85c649-grdqp"] Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.709037 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" podUID="fbb60200-f508-4614-baa8-5ca52e7d0127" containerName="dnsmasq-dns" containerID="cri-o://ce882643ed3e35fe169671ba1e98531fc1b2385502a138f584dbcb09c8719d65" gracePeriod=10 Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.709152 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-7bzvq" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.709503 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-7bzvq" event={"ID":"b0c57ee2-7be6-43f9-b2ce-7d912a74df28","Type":"ContainerDied","Data":"4c71b5336938814145532eaa5d66145368592a4243d3a6e9e87440728197a926"} Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.709532 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c71b5336938814145532eaa5d66145368592a4243d3a6e9e87440728197a926" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.739655 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-sr482"] Mar 18 20:24:53 crc kubenswrapper[4950]: E0318 20:24:53.740174 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0c57ee2-7be6-43f9-b2ce-7d912a74df28" containerName="mariadb-account-create-update" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.740257 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0c57ee2-7be6-43f9-b2ce-7d912a74df28" containerName="mariadb-account-create-update" Mar 18 20:24:53 crc kubenswrapper[4950]: E0318 20:24:53.740344 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99f9366a-2413-4243-9196-95088cf66e87" containerName="mariadb-account-create-update" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.740453 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="99f9366a-2413-4243-9196-95088cf66e87" containerName="mariadb-account-create-update" Mar 18 20:24:53 crc kubenswrapper[4950]: E0318 20:24:53.740570 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c89c3ec6-2109-4bd7-92e4-5d872dc95cca" containerName="mariadb-database-create" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.740668 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="c89c3ec6-2109-4bd7-92e4-5d872dc95cca" containerName="mariadb-database-create" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.740964 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="c89c3ec6-2109-4bd7-92e4-5d872dc95cca" containerName="mariadb-database-create" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.741048 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="99f9366a-2413-4243-9196-95088cf66e87" containerName="mariadb-account-create-update" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.741137 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0c57ee2-7be6-43f9-b2ce-7d912a74df28" containerName="mariadb-account-create-update" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.741922 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sr482" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.748796 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.749094 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2vwp6" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.756148 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-sr482"] Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.884389 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvhz4\" (UniqueName: \"kubernetes.io/projected/32f194c3-aab7-4bf3-ae91-2a7051efc958-kube-api-access-rvhz4\") pod \"glance-db-sync-sr482\" (UID: \"32f194c3-aab7-4bf3-ae91-2a7051efc958\") " pod="openstack/glance-db-sync-sr482" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.884539 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32f194c3-aab7-4bf3-ae91-2a7051efc958-config-data\") pod \"glance-db-sync-sr482\" (UID: \"32f194c3-aab7-4bf3-ae91-2a7051efc958\") " pod="openstack/glance-db-sync-sr482" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.884611 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32f194c3-aab7-4bf3-ae91-2a7051efc958-db-sync-config-data\") pod \"glance-db-sync-sr482\" (UID: \"32f194c3-aab7-4bf3-ae91-2a7051efc958\") " pod="openstack/glance-db-sync-sr482" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.884710 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32f194c3-aab7-4bf3-ae91-2a7051efc958-combined-ca-bundle\") pod \"glance-db-sync-sr482\" (UID: \"32f194c3-aab7-4bf3-ae91-2a7051efc958\") " pod="openstack/glance-db-sync-sr482" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.986484 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvhz4\" (UniqueName: \"kubernetes.io/projected/32f194c3-aab7-4bf3-ae91-2a7051efc958-kube-api-access-rvhz4\") pod \"glance-db-sync-sr482\" (UID: \"32f194c3-aab7-4bf3-ae91-2a7051efc958\") " pod="openstack/glance-db-sync-sr482" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.986565 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32f194c3-aab7-4bf3-ae91-2a7051efc958-config-data\") pod \"glance-db-sync-sr482\" (UID: \"32f194c3-aab7-4bf3-ae91-2a7051efc958\") " pod="openstack/glance-db-sync-sr482" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.986595 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32f194c3-aab7-4bf3-ae91-2a7051efc958-db-sync-config-data\") pod \"glance-db-sync-sr482\" (UID: \"32f194c3-aab7-4bf3-ae91-2a7051efc958\") " pod="openstack/glance-db-sync-sr482" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.986636 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32f194c3-aab7-4bf3-ae91-2a7051efc958-combined-ca-bundle\") pod \"glance-db-sync-sr482\" (UID: \"32f194c3-aab7-4bf3-ae91-2a7051efc958\") " pod="openstack/glance-db-sync-sr482" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.991759 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32f194c3-aab7-4bf3-ae91-2a7051efc958-config-data\") pod \"glance-db-sync-sr482\" (UID: \"32f194c3-aab7-4bf3-ae91-2a7051efc958\") " pod="openstack/glance-db-sync-sr482" Mar 18 20:24:53 crc kubenswrapper[4950]: I0318 20:24:53.992770 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32f194c3-aab7-4bf3-ae91-2a7051efc958-combined-ca-bundle\") pod \"glance-db-sync-sr482\" (UID: \"32f194c3-aab7-4bf3-ae91-2a7051efc958\") " pod="openstack/glance-db-sync-sr482" Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.005633 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvhz4\" (UniqueName: \"kubernetes.io/projected/32f194c3-aab7-4bf3-ae91-2a7051efc958-kube-api-access-rvhz4\") pod \"glance-db-sync-sr482\" (UID: \"32f194c3-aab7-4bf3-ae91-2a7051efc958\") " pod="openstack/glance-db-sync-sr482" Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.008094 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32f194c3-aab7-4bf3-ae91-2a7051efc958-db-sync-config-data\") pod \"glance-db-sync-sr482\" (UID: \"32f194c3-aab7-4bf3-ae91-2a7051efc958\") " pod="openstack/glance-db-sync-sr482" Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.064007 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sr482" Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.234550 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.295295 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbb60200-f508-4614-baa8-5ca52e7d0127-config\") pod \"fbb60200-f508-4614-baa8-5ca52e7d0127\" (UID: \"fbb60200-f508-4614-baa8-5ca52e7d0127\") " Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.295348 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48zjx\" (UniqueName: \"kubernetes.io/projected/fbb60200-f508-4614-baa8-5ca52e7d0127-kube-api-access-48zjx\") pod \"fbb60200-f508-4614-baa8-5ca52e7d0127\" (UID: \"fbb60200-f508-4614-baa8-5ca52e7d0127\") " Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.295466 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fbb60200-f508-4614-baa8-5ca52e7d0127-ovsdbserver-sb\") pod \"fbb60200-f508-4614-baa8-5ca52e7d0127\" (UID: \"fbb60200-f508-4614-baa8-5ca52e7d0127\") " Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.295503 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fbb60200-f508-4614-baa8-5ca52e7d0127-dns-svc\") pod \"fbb60200-f508-4614-baa8-5ca52e7d0127\" (UID: \"fbb60200-f508-4614-baa8-5ca52e7d0127\") " Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.306681 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbb60200-f508-4614-baa8-5ca52e7d0127-kube-api-access-48zjx" (OuterVolumeSpecName: "kube-api-access-48zjx") pod "fbb60200-f508-4614-baa8-5ca52e7d0127" (UID: "fbb60200-f508-4614-baa8-5ca52e7d0127"). InnerVolumeSpecName "kube-api-access-48zjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.340388 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fbb60200-f508-4614-baa8-5ca52e7d0127-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fbb60200-f508-4614-baa8-5ca52e7d0127" (UID: "fbb60200-f508-4614-baa8-5ca52e7d0127"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.366194 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fbb60200-f508-4614-baa8-5ca52e7d0127-config" (OuterVolumeSpecName: "config") pod "fbb60200-f508-4614-baa8-5ca52e7d0127" (UID: "fbb60200-f508-4614-baa8-5ca52e7d0127"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.368196 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fbb60200-f508-4614-baa8-5ca52e7d0127-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fbb60200-f508-4614-baa8-5ca52e7d0127" (UID: "fbb60200-f508-4614-baa8-5ca52e7d0127"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.397821 4950 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fbb60200-f508-4614-baa8-5ca52e7d0127-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.397862 4950 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fbb60200-f508-4614-baa8-5ca52e7d0127-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.397877 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbb60200-f508-4614-baa8-5ca52e7d0127-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.397898 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48zjx\" (UniqueName: \"kubernetes.io/projected/fbb60200-f508-4614-baa8-5ca52e7d0127-kube-api-access-48zjx\") on node \"crc\" DevicePath \"\"" Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.654674 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-sr482"] Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.718702 4950 generic.go:334] "Generic (PLEG): container finished" podID="fbb60200-f508-4614-baa8-5ca52e7d0127" containerID="ce882643ed3e35fe169671ba1e98531fc1b2385502a138f584dbcb09c8719d65" exitCode=0 Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.718720 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" event={"ID":"fbb60200-f508-4614-baa8-5ca52e7d0127","Type":"ContainerDied","Data":"ce882643ed3e35fe169671ba1e98531fc1b2385502a138f584dbcb09c8719d65"} Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.718764 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.718769 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ccf85c649-grdqp" event={"ID":"fbb60200-f508-4614-baa8-5ca52e7d0127","Type":"ContainerDied","Data":"33a345e55df35557abfd092f184e6bbc77304e7b572d8f30220f41dd25980d90"} Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.718789 4950 scope.go:117] "RemoveContainer" containerID="ce882643ed3e35fe169671ba1e98531fc1b2385502a138f584dbcb09c8719d65" Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.721589 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sr482" event={"ID":"32f194c3-aab7-4bf3-ae91-2a7051efc958","Type":"ContainerStarted","Data":"14151fe48c8de8deb97b4b631eb2419c1bd31a5d56ba2ccd67257d46a68c0d7a"} Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.744686 4950 scope.go:117] "RemoveContainer" containerID="c7135d56ed96b30b2c405ec1d8a4f772aedb0560266dee571056767b258309e0" Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.751079 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ccf85c649-grdqp"] Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.759143 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7ccf85c649-grdqp"] Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.761896 4950 scope.go:117] "RemoveContainer" containerID="ce882643ed3e35fe169671ba1e98531fc1b2385502a138f584dbcb09c8719d65" Mar 18 20:24:54 crc kubenswrapper[4950]: E0318 20:24:54.771965 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce882643ed3e35fe169671ba1e98531fc1b2385502a138f584dbcb09c8719d65\": container with ID starting with ce882643ed3e35fe169671ba1e98531fc1b2385502a138f584dbcb09c8719d65 not found: ID does not exist" containerID="ce882643ed3e35fe169671ba1e98531fc1b2385502a138f584dbcb09c8719d65" Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.772041 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce882643ed3e35fe169671ba1e98531fc1b2385502a138f584dbcb09c8719d65"} err="failed to get container status \"ce882643ed3e35fe169671ba1e98531fc1b2385502a138f584dbcb09c8719d65\": rpc error: code = NotFound desc = could not find container \"ce882643ed3e35fe169671ba1e98531fc1b2385502a138f584dbcb09c8719d65\": container with ID starting with ce882643ed3e35fe169671ba1e98531fc1b2385502a138f584dbcb09c8719d65 not found: ID does not exist" Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.772077 4950 scope.go:117] "RemoveContainer" containerID="c7135d56ed96b30b2c405ec1d8a4f772aedb0560266dee571056767b258309e0" Mar 18 20:24:54 crc kubenswrapper[4950]: E0318 20:24:54.772377 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7135d56ed96b30b2c405ec1d8a4f772aedb0560266dee571056767b258309e0\": container with ID starting with c7135d56ed96b30b2c405ec1d8a4f772aedb0560266dee571056767b258309e0 not found: ID does not exist" containerID="c7135d56ed96b30b2c405ec1d8a4f772aedb0560266dee571056767b258309e0" Mar 18 20:24:54 crc kubenswrapper[4950]: I0318 20:24:54.772428 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7135d56ed96b30b2c405ec1d8a4f772aedb0560266dee571056767b258309e0"} err="failed to get container status \"c7135d56ed96b30b2c405ec1d8a4f772aedb0560266dee571056767b258309e0\": rpc error: code = NotFound desc = could not find container \"c7135d56ed96b30b2c405ec1d8a4f772aedb0560266dee571056767b258309e0\": container with ID starting with c7135d56ed96b30b2c405ec1d8a4f772aedb0560266dee571056767b258309e0 not found: ID does not exist" Mar 18 20:24:55 crc kubenswrapper[4950]: I0318 20:24:55.488765 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbb60200-f508-4614-baa8-5ca52e7d0127" path="/var/lib/kubelet/pods/fbb60200-f508-4614-baa8-5ca52e7d0127/volumes" Mar 18 20:24:56 crc kubenswrapper[4950]: I0318 20:24:56.460998 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-7bzvq"] Mar 18 20:24:56 crc kubenswrapper[4950]: I0318 20:24:56.468904 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-7bzvq"] Mar 18 20:24:57 crc kubenswrapper[4950]: I0318 20:24:57.494615 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0c57ee2-7be6-43f9-b2ce-7d912a74df28" path="/var/lib/kubelet/pods/b0c57ee2-7be6-43f9-b2ce-7d912a74df28/volumes" Mar 18 20:25:01 crc kubenswrapper[4950]: I0318 20:25:01.470581 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-2n7qm"] Mar 18 20:25:01 crc kubenswrapper[4950]: E0318 20:25:01.471514 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbb60200-f508-4614-baa8-5ca52e7d0127" containerName="dnsmasq-dns" Mar 18 20:25:01 crc kubenswrapper[4950]: I0318 20:25:01.471535 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbb60200-f508-4614-baa8-5ca52e7d0127" containerName="dnsmasq-dns" Mar 18 20:25:01 crc kubenswrapper[4950]: E0318 20:25:01.471564 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbb60200-f508-4614-baa8-5ca52e7d0127" containerName="init" Mar 18 20:25:01 crc kubenswrapper[4950]: I0318 20:25:01.471571 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbb60200-f508-4614-baa8-5ca52e7d0127" containerName="init" Mar 18 20:25:01 crc kubenswrapper[4950]: I0318 20:25:01.471784 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbb60200-f508-4614-baa8-5ca52e7d0127" containerName="dnsmasq-dns" Mar 18 20:25:01 crc kubenswrapper[4950]: I0318 20:25:01.472947 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-2n7qm" Mar 18 20:25:01 crc kubenswrapper[4950]: I0318 20:25:01.474663 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Mar 18 20:25:01 crc kubenswrapper[4950]: I0318 20:25:01.491472 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-2n7qm"] Mar 18 20:25:01 crc kubenswrapper[4950]: I0318 20:25:01.538526 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npcrb\" (UniqueName: \"kubernetes.io/projected/884d39cd-355d-491b-ac02-b47295da532f-kube-api-access-npcrb\") pod \"root-account-create-update-2n7qm\" (UID: \"884d39cd-355d-491b-ac02-b47295da532f\") " pod="openstack/root-account-create-update-2n7qm" Mar 18 20:25:01 crc kubenswrapper[4950]: I0318 20:25:01.538581 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/884d39cd-355d-491b-ac02-b47295da532f-operator-scripts\") pod \"root-account-create-update-2n7qm\" (UID: \"884d39cd-355d-491b-ac02-b47295da532f\") " pod="openstack/root-account-create-update-2n7qm" Mar 18 20:25:01 crc kubenswrapper[4950]: I0318 20:25:01.639835 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npcrb\" (UniqueName: \"kubernetes.io/projected/884d39cd-355d-491b-ac02-b47295da532f-kube-api-access-npcrb\") pod \"root-account-create-update-2n7qm\" (UID: \"884d39cd-355d-491b-ac02-b47295da532f\") " pod="openstack/root-account-create-update-2n7qm" Mar 18 20:25:01 crc kubenswrapper[4950]: I0318 20:25:01.639886 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/884d39cd-355d-491b-ac02-b47295da532f-operator-scripts\") pod \"root-account-create-update-2n7qm\" (UID: \"884d39cd-355d-491b-ac02-b47295da532f\") " pod="openstack/root-account-create-update-2n7qm" Mar 18 20:25:01 crc kubenswrapper[4950]: I0318 20:25:01.640630 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/884d39cd-355d-491b-ac02-b47295da532f-operator-scripts\") pod \"root-account-create-update-2n7qm\" (UID: \"884d39cd-355d-491b-ac02-b47295da532f\") " pod="openstack/root-account-create-update-2n7qm" Mar 18 20:25:01 crc kubenswrapper[4950]: I0318 20:25:01.671259 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npcrb\" (UniqueName: \"kubernetes.io/projected/884d39cd-355d-491b-ac02-b47295da532f-kube-api-access-npcrb\") pod \"root-account-create-update-2n7qm\" (UID: \"884d39cd-355d-491b-ac02-b47295da532f\") " pod="openstack/root-account-create-update-2n7qm" Mar 18 20:25:01 crc kubenswrapper[4950]: I0318 20:25:01.802892 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-2n7qm" Mar 18 20:25:03 crc kubenswrapper[4950]: I0318 20:25:03.622439 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Mar 18 20:25:03 crc kubenswrapper[4950]: I0318 20:25:03.835869 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:25:03 crc kubenswrapper[4950]: I0318 20:25:03.836175 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:25:08 crc kubenswrapper[4950]: I0318 20:25:08.853222 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-zntd5" podUID="b72c4ff1-1234-41b7-9faa-b5fd15749084" containerName="ovn-controller" probeResult="failure" output=< Mar 18 20:25:08 crc kubenswrapper[4950]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 18 20:25:08 crc kubenswrapper[4950]: > Mar 18 20:25:08 crc kubenswrapper[4950]: I0318 20:25:08.976961 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:25:08 crc kubenswrapper[4950]: I0318 20:25:08.977341 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-lvf7g" Mar 18 20:25:08 crc kubenswrapper[4950]: I0318 20:25:08.985706 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-2n7qm"] Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.194915 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zntd5-config-nqq44"] Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.195863 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zntd5-config-nqq44" Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.199596 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.216684 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zntd5-config-nqq44"] Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.255042 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-var-run\") pod \"ovn-controller-zntd5-config-nqq44\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " pod="openstack/ovn-controller-zntd5-config-nqq44" Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.255469 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-var-log-ovn\") pod \"ovn-controller-zntd5-config-nqq44\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " pod="openstack/ovn-controller-zntd5-config-nqq44" Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.255509 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-additional-scripts\") pod \"ovn-controller-zntd5-config-nqq44\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " pod="openstack/ovn-controller-zntd5-config-nqq44" Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.255569 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-scripts\") pod \"ovn-controller-zntd5-config-nqq44\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " pod="openstack/ovn-controller-zntd5-config-nqq44" Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.255592 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcz7f\" (UniqueName: \"kubernetes.io/projected/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-kube-api-access-gcz7f\") pod \"ovn-controller-zntd5-config-nqq44\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " pod="openstack/ovn-controller-zntd5-config-nqq44" Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.255615 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-var-run-ovn\") pod \"ovn-controller-zntd5-config-nqq44\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " pod="openstack/ovn-controller-zntd5-config-nqq44" Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.357522 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcz7f\" (UniqueName: \"kubernetes.io/projected/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-kube-api-access-gcz7f\") pod \"ovn-controller-zntd5-config-nqq44\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " pod="openstack/ovn-controller-zntd5-config-nqq44" Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.357563 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-var-run-ovn\") pod \"ovn-controller-zntd5-config-nqq44\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " pod="openstack/ovn-controller-zntd5-config-nqq44" Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.357628 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-var-run\") pod \"ovn-controller-zntd5-config-nqq44\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " pod="openstack/ovn-controller-zntd5-config-nqq44" Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.357682 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-var-log-ovn\") pod \"ovn-controller-zntd5-config-nqq44\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " pod="openstack/ovn-controller-zntd5-config-nqq44" Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.357708 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-additional-scripts\") pod \"ovn-controller-zntd5-config-nqq44\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " pod="openstack/ovn-controller-zntd5-config-nqq44" Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.357747 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-scripts\") pod \"ovn-controller-zntd5-config-nqq44\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " pod="openstack/ovn-controller-zntd5-config-nqq44" Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.358099 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-var-run\") pod \"ovn-controller-zntd5-config-nqq44\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " pod="openstack/ovn-controller-zntd5-config-nqq44" Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.358542 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-var-run-ovn\") pod \"ovn-controller-zntd5-config-nqq44\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " pod="openstack/ovn-controller-zntd5-config-nqq44" Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.358600 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-var-log-ovn\") pod \"ovn-controller-zntd5-config-nqq44\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " pod="openstack/ovn-controller-zntd5-config-nqq44" Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.360602 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-additional-scripts\") pod \"ovn-controller-zntd5-config-nqq44\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " pod="openstack/ovn-controller-zntd5-config-nqq44" Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.361359 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-scripts\") pod \"ovn-controller-zntd5-config-nqq44\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " pod="openstack/ovn-controller-zntd5-config-nqq44" Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.383030 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcz7f\" (UniqueName: \"kubernetes.io/projected/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-kube-api-access-gcz7f\") pod \"ovn-controller-zntd5-config-nqq44\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " pod="openstack/ovn-controller-zntd5-config-nqq44" Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.511615 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zntd5-config-nqq44" Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.884935 4950 generic.go:334] "Generic (PLEG): container finished" podID="884d39cd-355d-491b-ac02-b47295da532f" containerID="a520c3de603944698f7246c1400a6eb9e6213248cb08ebdf7d5e4507728ba307" exitCode=0 Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.885234 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-2n7qm" event={"ID":"884d39cd-355d-491b-ac02-b47295da532f","Type":"ContainerDied","Data":"a520c3de603944698f7246c1400a6eb9e6213248cb08ebdf7d5e4507728ba307"} Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.885258 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-2n7qm" event={"ID":"884d39cd-355d-491b-ac02-b47295da532f","Type":"ContainerStarted","Data":"030ad0ece76ade3009dcab8b7468856840bd08b1b02d43b15adad5903249d5e5"} Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.897571 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sr482" event={"ID":"32f194c3-aab7-4bf3-ae91-2a7051efc958","Type":"ContainerStarted","Data":"735d0240236758f8873531be60822231e19073829893db212a7243f9d5cf82f2"} Mar 18 20:25:09 crc kubenswrapper[4950]: I0318 20:25:09.956372 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-sr482" podStartSLOduration=3.063263548 podStartE2EDuration="16.956354593s" podCreationTimestamp="2026-03-18 20:24:53 +0000 UTC" firstStartedPulling="2026-03-18 20:24:54.667950672 +0000 UTC m=+1107.908792550" lastFinishedPulling="2026-03-18 20:25:08.561041727 +0000 UTC m=+1121.801883595" observedRunningTime="2026-03-18 20:25:09.950448261 +0000 UTC m=+1123.191290129" watchObservedRunningTime="2026-03-18 20:25:09.956354593 +0000 UTC m=+1123.197196451" Mar 18 20:25:10 crc kubenswrapper[4950]: I0318 20:25:10.008610 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zntd5-config-nqq44"] Mar 18 20:25:10 crc kubenswrapper[4950]: W0318 20:25:10.016666 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a7bcdf6_0604_41cb_8c9d_ac76e955a57a.slice/crio-c4dbf311d7d6ee37dd071b27fbe95285398b0cd236ab9c143349616708431f7e WatchSource:0}: Error finding container c4dbf311d7d6ee37dd071b27fbe95285398b0cd236ab9c143349616708431f7e: Status 404 returned error can't find the container with id c4dbf311d7d6ee37dd071b27fbe95285398b0cd236ab9c143349616708431f7e Mar 18 20:25:10 crc kubenswrapper[4950]: I0318 20:25:10.905208 4950 generic.go:334] "Generic (PLEG): container finished" podID="7a7bcdf6-0604-41cb-8c9d-ac76e955a57a" containerID="e9069b3bc780d7e4876c5f477247669b5903a0aaf4a440d0d609149fdf78e0e1" exitCode=0 Mar 18 20:25:10 crc kubenswrapper[4950]: I0318 20:25:10.905274 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zntd5-config-nqq44" event={"ID":"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a","Type":"ContainerDied","Data":"e9069b3bc780d7e4876c5f477247669b5903a0aaf4a440d0d609149fdf78e0e1"} Mar 18 20:25:10 crc kubenswrapper[4950]: I0318 20:25:10.905721 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zntd5-config-nqq44" event={"ID":"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a","Type":"ContainerStarted","Data":"c4dbf311d7d6ee37dd071b27fbe95285398b0cd236ab9c143349616708431f7e"} Mar 18 20:25:11 crc kubenswrapper[4950]: I0318 20:25:11.246036 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-2n7qm" Mar 18 20:25:11 crc kubenswrapper[4950]: I0318 20:25:11.291956 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/884d39cd-355d-491b-ac02-b47295da532f-operator-scripts\") pod \"884d39cd-355d-491b-ac02-b47295da532f\" (UID: \"884d39cd-355d-491b-ac02-b47295da532f\") " Mar 18 20:25:11 crc kubenswrapper[4950]: I0318 20:25:11.292054 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npcrb\" (UniqueName: \"kubernetes.io/projected/884d39cd-355d-491b-ac02-b47295da532f-kube-api-access-npcrb\") pod \"884d39cd-355d-491b-ac02-b47295da532f\" (UID: \"884d39cd-355d-491b-ac02-b47295da532f\") " Mar 18 20:25:11 crc kubenswrapper[4950]: I0318 20:25:11.292713 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/884d39cd-355d-491b-ac02-b47295da532f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "884d39cd-355d-491b-ac02-b47295da532f" (UID: "884d39cd-355d-491b-ac02-b47295da532f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:25:11 crc kubenswrapper[4950]: I0318 20:25:11.297013 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/884d39cd-355d-491b-ac02-b47295da532f-kube-api-access-npcrb" (OuterVolumeSpecName: "kube-api-access-npcrb") pod "884d39cd-355d-491b-ac02-b47295da532f" (UID: "884d39cd-355d-491b-ac02-b47295da532f"). InnerVolumeSpecName "kube-api-access-npcrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:25:11 crc kubenswrapper[4950]: I0318 20:25:11.393566 4950 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/884d39cd-355d-491b-ac02-b47295da532f-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:11 crc kubenswrapper[4950]: I0318 20:25:11.393593 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npcrb\" (UniqueName: \"kubernetes.io/projected/884d39cd-355d-491b-ac02-b47295da532f-kube-api-access-npcrb\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:11 crc kubenswrapper[4950]: I0318 20:25:11.914635 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-2n7qm" event={"ID":"884d39cd-355d-491b-ac02-b47295da532f","Type":"ContainerDied","Data":"030ad0ece76ade3009dcab8b7468856840bd08b1b02d43b15adad5903249d5e5"} Mar 18 20:25:11 crc kubenswrapper[4950]: I0318 20:25:11.914711 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="030ad0ece76ade3009dcab8b7468856840bd08b1b02d43b15adad5903249d5e5" Mar 18 20:25:11 crc kubenswrapper[4950]: I0318 20:25:11.914659 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-2n7qm" Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.199810 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zntd5-config-nqq44" Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.314550 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-scripts\") pod \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.314816 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-var-run\") pod \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.314915 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-var-run" (OuterVolumeSpecName: "var-run") pod "7a7bcdf6-0604-41cb-8c9d-ac76e955a57a" (UID: "7a7bcdf6-0604-41cb-8c9d-ac76e955a57a"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.315174 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcz7f\" (UniqueName: \"kubernetes.io/projected/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-kube-api-access-gcz7f\") pod \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.315290 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-var-run-ovn\") pod \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.316210 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-var-log-ovn\") pod \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.316333 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-additional-scripts\") pod \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\" (UID: \"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a\") " Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.315500 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "7a7bcdf6-0604-41cb-8c9d-ac76e955a57a" (UID: "7a7bcdf6-0604-41cb-8c9d-ac76e955a57a"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.316340 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "7a7bcdf6-0604-41cb-8c9d-ac76e955a57a" (UID: "7a7bcdf6-0604-41cb-8c9d-ac76e955a57a"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.316954 4950 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-var-run\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.317030 4950 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.317090 4950 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.316971 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "7a7bcdf6-0604-41cb-8c9d-ac76e955a57a" (UID: "7a7bcdf6-0604-41cb-8c9d-ac76e955a57a"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.318242 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-scripts" (OuterVolumeSpecName: "scripts") pod "7a7bcdf6-0604-41cb-8c9d-ac76e955a57a" (UID: "7a7bcdf6-0604-41cb-8c9d-ac76e955a57a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.335122 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-kube-api-access-gcz7f" (OuterVolumeSpecName: "kube-api-access-gcz7f") pod "7a7bcdf6-0604-41cb-8c9d-ac76e955a57a" (UID: "7a7bcdf6-0604-41cb-8c9d-ac76e955a57a"). InnerVolumeSpecName "kube-api-access-gcz7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.418792 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcz7f\" (UniqueName: \"kubernetes.io/projected/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-kube-api-access-gcz7f\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.418826 4950 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.418836 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.932855 4950 generic.go:334] "Generic (PLEG): container finished" podID="b82f7668-ae1a-4bfe-be27-435cca5df467" containerID="a9215184422c6bfb281f115311fa646f8d17ed53337255e639a33ab2c4583605" exitCode=0 Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.933358 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b82f7668-ae1a-4bfe-be27-435cca5df467","Type":"ContainerDied","Data":"a9215184422c6bfb281f115311fa646f8d17ed53337255e639a33ab2c4583605"} Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.937387 4950 generic.go:334] "Generic (PLEG): container finished" podID="7502f5ea-31ca-4349-bd79-2219bf3903c0" containerID="c87e384b986ec489013da0da075d69803e2eecf7ea2534fbb8b55233e4c8990e" exitCode=0 Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.937520 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7502f5ea-31ca-4349-bd79-2219bf3903c0","Type":"ContainerDied","Data":"c87e384b986ec489013da0da075d69803e2eecf7ea2534fbb8b55233e4c8990e"} Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.941349 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zntd5-config-nqq44" event={"ID":"7a7bcdf6-0604-41cb-8c9d-ac76e955a57a","Type":"ContainerDied","Data":"c4dbf311d7d6ee37dd071b27fbe95285398b0cd236ab9c143349616708431f7e"} Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.941449 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4dbf311d7d6ee37dd071b27fbe95285398b0cd236ab9c143349616708431f7e" Mar 18 20:25:12 crc kubenswrapper[4950]: I0318 20:25:12.941568 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zntd5-config-nqq44" Mar 18 20:25:13 crc kubenswrapper[4950]: I0318 20:25:13.291154 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-zntd5-config-nqq44"] Mar 18 20:25:13 crc kubenswrapper[4950]: I0318 20:25:13.299870 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-zntd5-config-nqq44"] Mar 18 20:25:13 crc kubenswrapper[4950]: I0318 20:25:13.490308 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a7bcdf6-0604-41cb-8c9d-ac76e955a57a" path="/var/lib/kubelet/pods/7a7bcdf6-0604-41cb-8c9d-ac76e955a57a/volumes" Mar 18 20:25:13 crc kubenswrapper[4950]: I0318 20:25:13.894753 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-zntd5" Mar 18 20:25:13 crc kubenswrapper[4950]: I0318 20:25:13.971308 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b82f7668-ae1a-4bfe-be27-435cca5df467","Type":"ContainerStarted","Data":"ffa976652e557de2dbb7eaffaa7cda3491e64db27a9601f202aa674559da0c3b"} Mar 18 20:25:13 crc kubenswrapper[4950]: I0318 20:25:13.972212 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 18 20:25:13 crc kubenswrapper[4950]: I0318 20:25:13.976641 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7502f5ea-31ca-4349-bd79-2219bf3903c0","Type":"ContainerStarted","Data":"b069af21eaa4256e43fdff47e8842455ff9b0335a0b5b04aec42cea30f1c13fb"} Mar 18 20:25:13 crc kubenswrapper[4950]: I0318 20:25:13.976886 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:25:14 crc kubenswrapper[4950]: I0318 20:25:14.006814 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.381643316 podStartE2EDuration="1m16.006795473s" podCreationTimestamp="2026-03-18 20:23:58 +0000 UTC" firstStartedPulling="2026-03-18 20:24:01.185279928 +0000 UTC m=+1054.426121786" lastFinishedPulling="2026-03-18 20:24:38.810432075 +0000 UTC m=+1092.051273943" observedRunningTime="2026-03-18 20:25:13.999583627 +0000 UTC m=+1127.240425505" watchObservedRunningTime="2026-03-18 20:25:14.006795473 +0000 UTC m=+1127.247637341" Mar 18 20:25:14 crc kubenswrapper[4950]: I0318 20:25:14.032811 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.104611661 podStartE2EDuration="1m16.032784732s" podCreationTimestamp="2026-03-18 20:23:58 +0000 UTC" firstStartedPulling="2026-03-18 20:24:00.882876341 +0000 UTC m=+1054.123718209" lastFinishedPulling="2026-03-18 20:24:38.811049382 +0000 UTC m=+1092.051891280" observedRunningTime="2026-03-18 20:25:14.026948362 +0000 UTC m=+1127.267790220" watchObservedRunningTime="2026-03-18 20:25:14.032784732 +0000 UTC m=+1127.273626640" Mar 18 20:25:18 crc kubenswrapper[4950]: I0318 20:25:18.005593 4950 generic.go:334] "Generic (PLEG): container finished" podID="32f194c3-aab7-4bf3-ae91-2a7051efc958" containerID="735d0240236758f8873531be60822231e19073829893db212a7243f9d5cf82f2" exitCode=0 Mar 18 20:25:18 crc kubenswrapper[4950]: I0318 20:25:18.005733 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sr482" event={"ID":"32f194c3-aab7-4bf3-ae91-2a7051efc958","Type":"ContainerDied","Data":"735d0240236758f8873531be60822231e19073829893db212a7243f9d5cf82f2"} Mar 18 20:25:19 crc kubenswrapper[4950]: I0318 20:25:19.409590 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sr482" Mar 18 20:25:19 crc kubenswrapper[4950]: I0318 20:25:19.527822 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvhz4\" (UniqueName: \"kubernetes.io/projected/32f194c3-aab7-4bf3-ae91-2a7051efc958-kube-api-access-rvhz4\") pod \"32f194c3-aab7-4bf3-ae91-2a7051efc958\" (UID: \"32f194c3-aab7-4bf3-ae91-2a7051efc958\") " Mar 18 20:25:19 crc kubenswrapper[4950]: I0318 20:25:19.527899 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32f194c3-aab7-4bf3-ae91-2a7051efc958-combined-ca-bundle\") pod \"32f194c3-aab7-4bf3-ae91-2a7051efc958\" (UID: \"32f194c3-aab7-4bf3-ae91-2a7051efc958\") " Mar 18 20:25:19 crc kubenswrapper[4950]: I0318 20:25:19.527937 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32f194c3-aab7-4bf3-ae91-2a7051efc958-db-sync-config-data\") pod \"32f194c3-aab7-4bf3-ae91-2a7051efc958\" (UID: \"32f194c3-aab7-4bf3-ae91-2a7051efc958\") " Mar 18 20:25:19 crc kubenswrapper[4950]: I0318 20:25:19.527976 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32f194c3-aab7-4bf3-ae91-2a7051efc958-config-data\") pod \"32f194c3-aab7-4bf3-ae91-2a7051efc958\" (UID: \"32f194c3-aab7-4bf3-ae91-2a7051efc958\") " Mar 18 20:25:19 crc kubenswrapper[4950]: I0318 20:25:19.533278 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32f194c3-aab7-4bf3-ae91-2a7051efc958-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "32f194c3-aab7-4bf3-ae91-2a7051efc958" (UID: "32f194c3-aab7-4bf3-ae91-2a7051efc958"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:25:19 crc kubenswrapper[4950]: I0318 20:25:19.533739 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32f194c3-aab7-4bf3-ae91-2a7051efc958-kube-api-access-rvhz4" (OuterVolumeSpecName: "kube-api-access-rvhz4") pod "32f194c3-aab7-4bf3-ae91-2a7051efc958" (UID: "32f194c3-aab7-4bf3-ae91-2a7051efc958"). InnerVolumeSpecName "kube-api-access-rvhz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:25:19 crc kubenswrapper[4950]: I0318 20:25:19.556832 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32f194c3-aab7-4bf3-ae91-2a7051efc958-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32f194c3-aab7-4bf3-ae91-2a7051efc958" (UID: "32f194c3-aab7-4bf3-ae91-2a7051efc958"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:25:19 crc kubenswrapper[4950]: I0318 20:25:19.567275 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32f194c3-aab7-4bf3-ae91-2a7051efc958-config-data" (OuterVolumeSpecName: "config-data") pod "32f194c3-aab7-4bf3-ae91-2a7051efc958" (UID: "32f194c3-aab7-4bf3-ae91-2a7051efc958"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:25:19 crc kubenswrapper[4950]: I0318 20:25:19.629502 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32f194c3-aab7-4bf3-ae91-2a7051efc958-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:19 crc kubenswrapper[4950]: I0318 20:25:19.629538 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvhz4\" (UniqueName: \"kubernetes.io/projected/32f194c3-aab7-4bf3-ae91-2a7051efc958-kube-api-access-rvhz4\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:19 crc kubenswrapper[4950]: I0318 20:25:19.629548 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32f194c3-aab7-4bf3-ae91-2a7051efc958-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:19 crc kubenswrapper[4950]: I0318 20:25:19.629557 4950 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32f194c3-aab7-4bf3-ae91-2a7051efc958-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.023491 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sr482" event={"ID":"32f194c3-aab7-4bf3-ae91-2a7051efc958","Type":"ContainerDied","Data":"14151fe48c8de8deb97b4b631eb2419c1bd31a5d56ba2ccd67257d46a68c0d7a"} Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.023546 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14151fe48c8de8deb97b4b631eb2419c1bd31a5d56ba2ccd67257d46a68c0d7a" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.023571 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sr482" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.668499 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-547b76fb45-s9htl"] Mar 18 20:25:20 crc kubenswrapper[4950]: E0318 20:25:20.668795 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32f194c3-aab7-4bf3-ae91-2a7051efc958" containerName="glance-db-sync" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.668807 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="32f194c3-aab7-4bf3-ae91-2a7051efc958" containerName="glance-db-sync" Mar 18 20:25:20 crc kubenswrapper[4950]: E0318 20:25:20.668823 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="884d39cd-355d-491b-ac02-b47295da532f" containerName="mariadb-account-create-update" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.668829 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="884d39cd-355d-491b-ac02-b47295da532f" containerName="mariadb-account-create-update" Mar 18 20:25:20 crc kubenswrapper[4950]: E0318 20:25:20.668847 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a7bcdf6-0604-41cb-8c9d-ac76e955a57a" containerName="ovn-config" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.668853 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a7bcdf6-0604-41cb-8c9d-ac76e955a57a" containerName="ovn-config" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.668989 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="884d39cd-355d-491b-ac02-b47295da532f" containerName="mariadb-account-create-update" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.669005 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="32f194c3-aab7-4bf3-ae91-2a7051efc958" containerName="glance-db-sync" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.669018 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a7bcdf6-0604-41cb-8c9d-ac76e955a57a" containerName="ovn-config" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.669751 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-547b76fb45-s9htl" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.723471 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-547b76fb45-s9htl"] Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.845514 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-ovsdbserver-sb\") pod \"dnsmasq-dns-547b76fb45-s9htl\" (UID: \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\") " pod="openstack/dnsmasq-dns-547b76fb45-s9htl" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.845605 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-config\") pod \"dnsmasq-dns-547b76fb45-s9htl\" (UID: \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\") " pod="openstack/dnsmasq-dns-547b76fb45-s9htl" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.845677 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-ovsdbserver-nb\") pod \"dnsmasq-dns-547b76fb45-s9htl\" (UID: \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\") " pod="openstack/dnsmasq-dns-547b76fb45-s9htl" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.845699 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-dns-svc\") pod \"dnsmasq-dns-547b76fb45-s9htl\" (UID: \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\") " pod="openstack/dnsmasq-dns-547b76fb45-s9htl" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.845760 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmxv8\" (UniqueName: \"kubernetes.io/projected/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-kube-api-access-kmxv8\") pod \"dnsmasq-dns-547b76fb45-s9htl\" (UID: \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\") " pod="openstack/dnsmasq-dns-547b76fb45-s9htl" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.947611 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-ovsdbserver-sb\") pod \"dnsmasq-dns-547b76fb45-s9htl\" (UID: \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\") " pod="openstack/dnsmasq-dns-547b76fb45-s9htl" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.947675 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-config\") pod \"dnsmasq-dns-547b76fb45-s9htl\" (UID: \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\") " pod="openstack/dnsmasq-dns-547b76fb45-s9htl" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.947739 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-ovsdbserver-nb\") pod \"dnsmasq-dns-547b76fb45-s9htl\" (UID: \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\") " pod="openstack/dnsmasq-dns-547b76fb45-s9htl" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.947760 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-dns-svc\") pod \"dnsmasq-dns-547b76fb45-s9htl\" (UID: \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\") " pod="openstack/dnsmasq-dns-547b76fb45-s9htl" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.947798 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmxv8\" (UniqueName: \"kubernetes.io/projected/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-kube-api-access-kmxv8\") pod \"dnsmasq-dns-547b76fb45-s9htl\" (UID: \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\") " pod="openstack/dnsmasq-dns-547b76fb45-s9htl" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.948827 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-ovsdbserver-sb\") pod \"dnsmasq-dns-547b76fb45-s9htl\" (UID: \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\") " pod="openstack/dnsmasq-dns-547b76fb45-s9htl" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.948921 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-config\") pod \"dnsmasq-dns-547b76fb45-s9htl\" (UID: \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\") " pod="openstack/dnsmasq-dns-547b76fb45-s9htl" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.949391 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-ovsdbserver-nb\") pod \"dnsmasq-dns-547b76fb45-s9htl\" (UID: \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\") " pod="openstack/dnsmasq-dns-547b76fb45-s9htl" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.949499 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-dns-svc\") pod \"dnsmasq-dns-547b76fb45-s9htl\" (UID: \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\") " pod="openstack/dnsmasq-dns-547b76fb45-s9htl" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.980497 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmxv8\" (UniqueName: \"kubernetes.io/projected/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-kube-api-access-kmxv8\") pod \"dnsmasq-dns-547b76fb45-s9htl\" (UID: \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\") " pod="openstack/dnsmasq-dns-547b76fb45-s9htl" Mar 18 20:25:20 crc kubenswrapper[4950]: I0318 20:25:20.991818 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-547b76fb45-s9htl" Mar 18 20:25:21 crc kubenswrapper[4950]: I0318 20:25:21.591378 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-547b76fb45-s9htl"] Mar 18 20:25:22 crc kubenswrapper[4950]: I0318 20:25:22.037589 4950 generic.go:334] "Generic (PLEG): container finished" podID="96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac" containerID="ec73589967679f88800110dd92359f8262173fe346033fc696a416b664281a91" exitCode=0 Mar 18 20:25:22 crc kubenswrapper[4950]: I0318 20:25:22.037647 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-547b76fb45-s9htl" event={"ID":"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac","Type":"ContainerDied","Data":"ec73589967679f88800110dd92359f8262173fe346033fc696a416b664281a91"} Mar 18 20:25:22 crc kubenswrapper[4950]: I0318 20:25:22.038076 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-547b76fb45-s9htl" event={"ID":"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac","Type":"ContainerStarted","Data":"62916e59361722b7ce64cb4bd1b19fdc7ca2c4299ab9d9e48292b2d5f6c8d70b"} Mar 18 20:25:23 crc kubenswrapper[4950]: I0318 20:25:23.051927 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-547b76fb45-s9htl" event={"ID":"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac","Type":"ContainerStarted","Data":"b2c55f5f177417991f90810fa75e695c04b170f5a7d0bf92b93164c924bac67a"} Mar 18 20:25:23 crc kubenswrapper[4950]: I0318 20:25:23.052537 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-547b76fb45-s9htl" Mar 18 20:25:23 crc kubenswrapper[4950]: I0318 20:25:23.072539 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-547b76fb45-s9htl" podStartSLOduration=3.072523534 podStartE2EDuration="3.072523534s" podCreationTimestamp="2026-03-18 20:25:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:25:23.069985018 +0000 UTC m=+1136.310826886" watchObservedRunningTime="2026-03-18 20:25:23.072523534 +0000 UTC m=+1136.313365402" Mar 18 20:25:29 crc kubenswrapper[4950]: I0318 20:25:29.963752 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:25:30 crc kubenswrapper[4950]: I0318 20:25:30.614079 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 18 20:25:30 crc kubenswrapper[4950]: I0318 20:25:30.993684 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-547b76fb45-s9htl" Mar 18 20:25:31 crc kubenswrapper[4950]: I0318 20:25:31.051508 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f697c8bff-wvwhn"] Mar 18 20:25:31 crc kubenswrapper[4950]: I0318 20:25:31.051836 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" podUID="c4f9eed1-5dc4-4e60-b35a-812bc273acc6" containerName="dnsmasq-dns" containerID="cri-o://35ce87ad0e96b6e3da14255f5539d36606cf3e9e498844b9b1bea2ba280a43b6" gracePeriod=10 Mar 18 20:25:31 crc kubenswrapper[4950]: I0318 20:25:31.536697 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" Mar 18 20:25:31 crc kubenswrapper[4950]: I0318 20:25:31.631532 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-dns-svc\") pod \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\" (UID: \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\") " Mar 18 20:25:31 crc kubenswrapper[4950]: I0318 20:25:31.631876 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-ovsdbserver-sb\") pod \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\" (UID: \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\") " Mar 18 20:25:31 crc kubenswrapper[4950]: I0318 20:25:31.632054 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-ovsdbserver-nb\") pod \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\" (UID: \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\") " Mar 18 20:25:31 crc kubenswrapper[4950]: I0318 20:25:31.632150 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-config\") pod \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\" (UID: \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\") " Mar 18 20:25:31 crc kubenswrapper[4950]: I0318 20:25:31.632228 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djjt6\" (UniqueName: \"kubernetes.io/projected/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-kube-api-access-djjt6\") pod \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\" (UID: \"c4f9eed1-5dc4-4e60-b35a-812bc273acc6\") " Mar 18 20:25:31 crc kubenswrapper[4950]: I0318 20:25:31.654658 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-kube-api-access-djjt6" (OuterVolumeSpecName: "kube-api-access-djjt6") pod "c4f9eed1-5dc4-4e60-b35a-812bc273acc6" (UID: "c4f9eed1-5dc4-4e60-b35a-812bc273acc6"). InnerVolumeSpecName "kube-api-access-djjt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:25:31 crc kubenswrapper[4950]: I0318 20:25:31.689355 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-config" (OuterVolumeSpecName: "config") pod "c4f9eed1-5dc4-4e60-b35a-812bc273acc6" (UID: "c4f9eed1-5dc4-4e60-b35a-812bc273acc6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:25:31 crc kubenswrapper[4950]: I0318 20:25:31.694398 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c4f9eed1-5dc4-4e60-b35a-812bc273acc6" (UID: "c4f9eed1-5dc4-4e60-b35a-812bc273acc6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:25:31 crc kubenswrapper[4950]: I0318 20:25:31.699961 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c4f9eed1-5dc4-4e60-b35a-812bc273acc6" (UID: "c4f9eed1-5dc4-4e60-b35a-812bc273acc6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:25:31 crc kubenswrapper[4950]: I0318 20:25:31.709967 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c4f9eed1-5dc4-4e60-b35a-812bc273acc6" (UID: "c4f9eed1-5dc4-4e60-b35a-812bc273acc6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:25:31 crc kubenswrapper[4950]: I0318 20:25:31.734025 4950 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:31 crc kubenswrapper[4950]: I0318 20:25:31.734055 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:31 crc kubenswrapper[4950]: I0318 20:25:31.734066 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djjt6\" (UniqueName: \"kubernetes.io/projected/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-kube-api-access-djjt6\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:31 crc kubenswrapper[4950]: I0318 20:25:31.734075 4950 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:31 crc kubenswrapper[4950]: I0318 20:25:31.734083 4950 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4f9eed1-5dc4-4e60-b35a-812bc273acc6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.122674 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-z2ffj"] Mar 18 20:25:32 crc kubenswrapper[4950]: E0318 20:25:32.123232 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4f9eed1-5dc4-4e60-b35a-812bc273acc6" containerName="dnsmasq-dns" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.123243 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4f9eed1-5dc4-4e60-b35a-812bc273acc6" containerName="dnsmasq-dns" Mar 18 20:25:32 crc kubenswrapper[4950]: E0318 20:25:32.123271 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4f9eed1-5dc4-4e60-b35a-812bc273acc6" containerName="init" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.123278 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4f9eed1-5dc4-4e60-b35a-812bc273acc6" containerName="init" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.123508 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4f9eed1-5dc4-4e60-b35a-812bc273acc6" containerName="dnsmasq-dns" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.123954 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-z2ffj" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.136901 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-z2ffj"] Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.141564 4950 generic.go:334] "Generic (PLEG): container finished" podID="c4f9eed1-5dc4-4e60-b35a-812bc273acc6" containerID="35ce87ad0e96b6e3da14255f5539d36606cf3e9e498844b9b1bea2ba280a43b6" exitCode=0 Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.141609 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" event={"ID":"c4f9eed1-5dc4-4e60-b35a-812bc273acc6","Type":"ContainerDied","Data":"35ce87ad0e96b6e3da14255f5539d36606cf3e9e498844b9b1bea2ba280a43b6"} Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.141638 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" event={"ID":"c4f9eed1-5dc4-4e60-b35a-812bc273acc6","Type":"ContainerDied","Data":"9db2c9aed0821ee092af95f77e7005ba1c97396fe3a65ad17f620a7e1ee880ac"} Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.141655 4950 scope.go:117] "RemoveContainer" containerID="35ce87ad0e96b6e3da14255f5539d36606cf3e9e498844b9b1bea2ba280a43b6" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.141609 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f697c8bff-wvwhn" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.242387 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5f9t\" (UniqueName: \"kubernetes.io/projected/24db2c06-d91e-4fc1-bcb5-9f29546368e2-kube-api-access-q5f9t\") pod \"cinder-db-create-z2ffj\" (UID: \"24db2c06-d91e-4fc1-bcb5-9f29546368e2\") " pod="openstack/cinder-db-create-z2ffj" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.242495 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24db2c06-d91e-4fc1-bcb5-9f29546368e2-operator-scripts\") pod \"cinder-db-create-z2ffj\" (UID: \"24db2c06-d91e-4fc1-bcb5-9f29546368e2\") " pod="openstack/cinder-db-create-z2ffj" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.249182 4950 scope.go:117] "RemoveContainer" containerID="343707eb2fcc094bab69e01c5cf3a14e2e573896ed1c5f1fc16552c5d5afb6aa" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.274608 4950 scope.go:117] "RemoveContainer" containerID="35ce87ad0e96b6e3da14255f5539d36606cf3e9e498844b9b1bea2ba280a43b6" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.280301 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f697c8bff-wvwhn"] Mar 18 20:25:32 crc kubenswrapper[4950]: E0318 20:25:32.282508 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35ce87ad0e96b6e3da14255f5539d36606cf3e9e498844b9b1bea2ba280a43b6\": container with ID starting with 35ce87ad0e96b6e3da14255f5539d36606cf3e9e498844b9b1bea2ba280a43b6 not found: ID does not exist" containerID="35ce87ad0e96b6e3da14255f5539d36606cf3e9e498844b9b1bea2ba280a43b6" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.282551 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35ce87ad0e96b6e3da14255f5539d36606cf3e9e498844b9b1bea2ba280a43b6"} err="failed to get container status \"35ce87ad0e96b6e3da14255f5539d36606cf3e9e498844b9b1bea2ba280a43b6\": rpc error: code = NotFound desc = could not find container \"35ce87ad0e96b6e3da14255f5539d36606cf3e9e498844b9b1bea2ba280a43b6\": container with ID starting with 35ce87ad0e96b6e3da14255f5539d36606cf3e9e498844b9b1bea2ba280a43b6 not found: ID does not exist" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.282575 4950 scope.go:117] "RemoveContainer" containerID="343707eb2fcc094bab69e01c5cf3a14e2e573896ed1c5f1fc16552c5d5afb6aa" Mar 18 20:25:32 crc kubenswrapper[4950]: E0318 20:25:32.282875 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"343707eb2fcc094bab69e01c5cf3a14e2e573896ed1c5f1fc16552c5d5afb6aa\": container with ID starting with 343707eb2fcc094bab69e01c5cf3a14e2e573896ed1c5f1fc16552c5d5afb6aa not found: ID does not exist" containerID="343707eb2fcc094bab69e01c5cf3a14e2e573896ed1c5f1fc16552c5d5afb6aa" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.282924 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"343707eb2fcc094bab69e01c5cf3a14e2e573896ed1c5f1fc16552c5d5afb6aa"} err="failed to get container status \"343707eb2fcc094bab69e01c5cf3a14e2e573896ed1c5f1fc16552c5d5afb6aa\": rpc error: code = NotFound desc = could not find container \"343707eb2fcc094bab69e01c5cf3a14e2e573896ed1c5f1fc16552c5d5afb6aa\": container with ID starting with 343707eb2fcc094bab69e01c5cf3a14e2e573896ed1c5f1fc16552c5d5afb6aa not found: ID does not exist" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.296520 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f697c8bff-wvwhn"] Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.343620 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5f9t\" (UniqueName: \"kubernetes.io/projected/24db2c06-d91e-4fc1-bcb5-9f29546368e2-kube-api-access-q5f9t\") pod \"cinder-db-create-z2ffj\" (UID: \"24db2c06-d91e-4fc1-bcb5-9f29546368e2\") " pod="openstack/cinder-db-create-z2ffj" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.343713 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24db2c06-d91e-4fc1-bcb5-9f29546368e2-operator-scripts\") pod \"cinder-db-create-z2ffj\" (UID: \"24db2c06-d91e-4fc1-bcb5-9f29546368e2\") " pod="openstack/cinder-db-create-z2ffj" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.344452 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24db2c06-d91e-4fc1-bcb5-9f29546368e2-operator-scripts\") pod \"cinder-db-create-z2ffj\" (UID: \"24db2c06-d91e-4fc1-bcb5-9f29546368e2\") " pod="openstack/cinder-db-create-z2ffj" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.360405 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-32f0-account-create-update-7fwpt"] Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.361293 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-32f0-account-create-update-7fwpt" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.363036 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.368057 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5f9t\" (UniqueName: \"kubernetes.io/projected/24db2c06-d91e-4fc1-bcb5-9f29546368e2-kube-api-access-q5f9t\") pod \"cinder-db-create-z2ffj\" (UID: \"24db2c06-d91e-4fc1-bcb5-9f29546368e2\") " pod="openstack/cinder-db-create-z2ffj" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.417447 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-32f0-account-create-update-7fwpt"] Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.448281 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5a72833-ca01-4da1-84e9-2b144ac64814-operator-scripts\") pod \"cinder-32f0-account-create-update-7fwpt\" (UID: \"e5a72833-ca01-4da1-84e9-2b144ac64814\") " pod="openstack/cinder-32f0-account-create-update-7fwpt" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.448342 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82mq2\" (UniqueName: \"kubernetes.io/projected/e5a72833-ca01-4da1-84e9-2b144ac64814-kube-api-access-82mq2\") pod \"cinder-32f0-account-create-update-7fwpt\" (UID: \"e5a72833-ca01-4da1-84e9-2b144ac64814\") " pod="openstack/cinder-32f0-account-create-update-7fwpt" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.459388 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-vnl9g"] Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.460382 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-vnl9g" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.476241 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-vnl9g"] Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.539624 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-z2ffj" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.541266 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-t4zg5"] Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.542183 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-t4zg5" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.549809 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5a72833-ca01-4da1-84e9-2b144ac64814-operator-scripts\") pod \"cinder-32f0-account-create-update-7fwpt\" (UID: \"e5a72833-ca01-4da1-84e9-2b144ac64814\") " pod="openstack/cinder-32f0-account-create-update-7fwpt" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.550025 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3abd28b2-fbc4-4b22-8f6f-a34917f8cba3-operator-scripts\") pod \"barbican-db-create-vnl9g\" (UID: \"3abd28b2-fbc4-4b22-8f6f-a34917f8cba3\") " pod="openstack/barbican-db-create-vnl9g" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.550107 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82mq2\" (UniqueName: \"kubernetes.io/projected/e5a72833-ca01-4da1-84e9-2b144ac64814-kube-api-access-82mq2\") pod \"cinder-32f0-account-create-update-7fwpt\" (UID: \"e5a72833-ca01-4da1-84e9-2b144ac64814\") " pod="openstack/cinder-32f0-account-create-update-7fwpt" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.550182 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n65g\" (UniqueName: \"kubernetes.io/projected/3abd28b2-fbc4-4b22-8f6f-a34917f8cba3-kube-api-access-8n65g\") pod \"barbican-db-create-vnl9g\" (UID: \"3abd28b2-fbc4-4b22-8f6f-a34917f8cba3\") " pod="openstack/barbican-db-create-vnl9g" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.550946 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5a72833-ca01-4da1-84e9-2b144ac64814-operator-scripts\") pod \"cinder-32f0-account-create-update-7fwpt\" (UID: \"e5a72833-ca01-4da1-84e9-2b144ac64814\") " pod="openstack/cinder-32f0-account-create-update-7fwpt" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.561405 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-t4zg5"] Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.595033 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82mq2\" (UniqueName: \"kubernetes.io/projected/e5a72833-ca01-4da1-84e9-2b144ac64814-kube-api-access-82mq2\") pod \"cinder-32f0-account-create-update-7fwpt\" (UID: \"e5a72833-ca01-4da1-84e9-2b144ac64814\") " pod="openstack/cinder-32f0-account-create-update-7fwpt" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.651208 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzwd4\" (UniqueName: \"kubernetes.io/projected/0af1dd1d-4630-49a6-9f2d-a056fefecf78-kube-api-access-lzwd4\") pod \"neutron-db-create-t4zg5\" (UID: \"0af1dd1d-4630-49a6-9f2d-a056fefecf78\") " pod="openstack/neutron-db-create-t4zg5" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.651283 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0af1dd1d-4630-49a6-9f2d-a056fefecf78-operator-scripts\") pod \"neutron-db-create-t4zg5\" (UID: \"0af1dd1d-4630-49a6-9f2d-a056fefecf78\") " pod="openstack/neutron-db-create-t4zg5" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.651324 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3abd28b2-fbc4-4b22-8f6f-a34917f8cba3-operator-scripts\") pod \"barbican-db-create-vnl9g\" (UID: \"3abd28b2-fbc4-4b22-8f6f-a34917f8cba3\") " pod="openstack/barbican-db-create-vnl9g" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.651353 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n65g\" (UniqueName: \"kubernetes.io/projected/3abd28b2-fbc4-4b22-8f6f-a34917f8cba3-kube-api-access-8n65g\") pod \"barbican-db-create-vnl9g\" (UID: \"3abd28b2-fbc4-4b22-8f6f-a34917f8cba3\") " pod="openstack/barbican-db-create-vnl9g" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.652126 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3abd28b2-fbc4-4b22-8f6f-a34917f8cba3-operator-scripts\") pod \"barbican-db-create-vnl9g\" (UID: \"3abd28b2-fbc4-4b22-8f6f-a34917f8cba3\") " pod="openstack/barbican-db-create-vnl9g" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.652531 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-jz448"] Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.653641 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jz448" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.655961 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-bsdbw" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.656367 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.659227 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.659455 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.679538 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-jz448"] Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.683111 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8n65g\" (UniqueName: \"kubernetes.io/projected/3abd28b2-fbc4-4b22-8f6f-a34917f8cba3-kube-api-access-8n65g\") pod \"barbican-db-create-vnl9g\" (UID: \"3abd28b2-fbc4-4b22-8f6f-a34917f8cba3\") " pod="openstack/barbican-db-create-vnl9g" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.709916 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-32f0-account-create-update-7fwpt" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.772279 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90aa20e7-e9c9-46c5-ad83-8004caa09547-combined-ca-bundle\") pod \"keystone-db-sync-jz448\" (UID: \"90aa20e7-e9c9-46c5-ad83-8004caa09547\") " pod="openstack/keystone-db-sync-jz448" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.772375 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzwd4\" (UniqueName: \"kubernetes.io/projected/0af1dd1d-4630-49a6-9f2d-a056fefecf78-kube-api-access-lzwd4\") pod \"neutron-db-create-t4zg5\" (UID: \"0af1dd1d-4630-49a6-9f2d-a056fefecf78\") " pod="openstack/neutron-db-create-t4zg5" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.772400 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfgtc\" (UniqueName: \"kubernetes.io/projected/90aa20e7-e9c9-46c5-ad83-8004caa09547-kube-api-access-cfgtc\") pod \"keystone-db-sync-jz448\" (UID: \"90aa20e7-e9c9-46c5-ad83-8004caa09547\") " pod="openstack/keystone-db-sync-jz448" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.772466 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0af1dd1d-4630-49a6-9f2d-a056fefecf78-operator-scripts\") pod \"neutron-db-create-t4zg5\" (UID: \"0af1dd1d-4630-49a6-9f2d-a056fefecf78\") " pod="openstack/neutron-db-create-t4zg5" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.772502 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90aa20e7-e9c9-46c5-ad83-8004caa09547-config-data\") pod \"keystone-db-sync-jz448\" (UID: \"90aa20e7-e9c9-46c5-ad83-8004caa09547\") " pod="openstack/keystone-db-sync-jz448" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.773569 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0af1dd1d-4630-49a6-9f2d-a056fefecf78-operator-scripts\") pod \"neutron-db-create-t4zg5\" (UID: \"0af1dd1d-4630-49a6-9f2d-a056fefecf78\") " pod="openstack/neutron-db-create-t4zg5" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.779552 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-vnl9g" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.796256 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-7fe1-account-create-update-9rdjc"] Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.816135 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7fe1-account-create-update-9rdjc" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.819017 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzwd4\" (UniqueName: \"kubernetes.io/projected/0af1dd1d-4630-49a6-9f2d-a056fefecf78-kube-api-access-lzwd4\") pod \"neutron-db-create-t4zg5\" (UID: \"0af1dd1d-4630-49a6-9f2d-a056fefecf78\") " pod="openstack/neutron-db-create-t4zg5" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.830728 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.873830 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90aa20e7-e9c9-46c5-ad83-8004caa09547-config-data\") pod \"keystone-db-sync-jz448\" (UID: \"90aa20e7-e9c9-46c5-ad83-8004caa09547\") " pod="openstack/keystone-db-sync-jz448" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.874098 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90aa20e7-e9c9-46c5-ad83-8004caa09547-combined-ca-bundle\") pod \"keystone-db-sync-jz448\" (UID: \"90aa20e7-e9c9-46c5-ad83-8004caa09547\") " pod="openstack/keystone-db-sync-jz448" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.874253 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfgtc\" (UniqueName: \"kubernetes.io/projected/90aa20e7-e9c9-46c5-ad83-8004caa09547-kube-api-access-cfgtc\") pod \"keystone-db-sync-jz448\" (UID: \"90aa20e7-e9c9-46c5-ad83-8004caa09547\") " pod="openstack/keystone-db-sync-jz448" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.878044 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90aa20e7-e9c9-46c5-ad83-8004caa09547-config-data\") pod \"keystone-db-sync-jz448\" (UID: \"90aa20e7-e9c9-46c5-ad83-8004caa09547\") " pod="openstack/keystone-db-sync-jz448" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.886317 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90aa20e7-e9c9-46c5-ad83-8004caa09547-combined-ca-bundle\") pod \"keystone-db-sync-jz448\" (UID: \"90aa20e7-e9c9-46c5-ad83-8004caa09547\") " pod="openstack/keystone-db-sync-jz448" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.895635 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-7fe1-account-create-update-9rdjc"] Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.923657 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfgtc\" (UniqueName: \"kubernetes.io/projected/90aa20e7-e9c9-46c5-ad83-8004caa09547-kube-api-access-cfgtc\") pod \"keystone-db-sync-jz448\" (UID: \"90aa20e7-e9c9-46c5-ad83-8004caa09547\") " pod="openstack/keystone-db-sync-jz448" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.933355 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-t4zg5" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.975442 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kdz9\" (UniqueName: \"kubernetes.io/projected/875a302d-40dd-4c0f-a729-cd6f3dcb4956-kube-api-access-7kdz9\") pod \"barbican-7fe1-account-create-update-9rdjc\" (UID: \"875a302d-40dd-4c0f-a729-cd6f3dcb4956\") " pod="openstack/barbican-7fe1-account-create-update-9rdjc" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.975519 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/875a302d-40dd-4c0f-a729-cd6f3dcb4956-operator-scripts\") pod \"barbican-7fe1-account-create-update-9rdjc\" (UID: \"875a302d-40dd-4c0f-a729-cd6f3dcb4956\") " pod="openstack/barbican-7fe1-account-create-update-9rdjc" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.978634 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-9524-account-create-update-fznw7"] Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.980925 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jz448" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.983295 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9524-account-create-update-fznw7" Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.985540 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9524-account-create-update-fznw7"] Mar 18 20:25:32 crc kubenswrapper[4950]: I0318 20:25:32.986267 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.044872 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-z2ffj"] Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.079026 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kdz9\" (UniqueName: \"kubernetes.io/projected/875a302d-40dd-4c0f-a729-cd6f3dcb4956-kube-api-access-7kdz9\") pod \"barbican-7fe1-account-create-update-9rdjc\" (UID: \"875a302d-40dd-4c0f-a729-cd6f3dcb4956\") " pod="openstack/barbican-7fe1-account-create-update-9rdjc" Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.079110 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q84t7\" (UniqueName: \"kubernetes.io/projected/9fffb916-a3fd-400a-afd9-997515bf6850-kube-api-access-q84t7\") pod \"neutron-9524-account-create-update-fznw7\" (UID: \"9fffb916-a3fd-400a-afd9-997515bf6850\") " pod="openstack/neutron-9524-account-create-update-fznw7" Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.079145 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/875a302d-40dd-4c0f-a729-cd6f3dcb4956-operator-scripts\") pod \"barbican-7fe1-account-create-update-9rdjc\" (UID: \"875a302d-40dd-4c0f-a729-cd6f3dcb4956\") " pod="openstack/barbican-7fe1-account-create-update-9rdjc" Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.079167 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9fffb916-a3fd-400a-afd9-997515bf6850-operator-scripts\") pod \"neutron-9524-account-create-update-fznw7\" (UID: \"9fffb916-a3fd-400a-afd9-997515bf6850\") " pod="openstack/neutron-9524-account-create-update-fznw7" Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.080391 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/875a302d-40dd-4c0f-a729-cd6f3dcb4956-operator-scripts\") pod \"barbican-7fe1-account-create-update-9rdjc\" (UID: \"875a302d-40dd-4c0f-a729-cd6f3dcb4956\") " pod="openstack/barbican-7fe1-account-create-update-9rdjc" Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.098587 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kdz9\" (UniqueName: \"kubernetes.io/projected/875a302d-40dd-4c0f-a729-cd6f3dcb4956-kube-api-access-7kdz9\") pod \"barbican-7fe1-account-create-update-9rdjc\" (UID: \"875a302d-40dd-4c0f-a729-cd6f3dcb4956\") " pod="openstack/barbican-7fe1-account-create-update-9rdjc" Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.153676 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7fe1-account-create-update-9rdjc" Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.165575 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-z2ffj" event={"ID":"24db2c06-d91e-4fc1-bcb5-9f29546368e2","Type":"ContainerStarted","Data":"85ed251041c207da1a1063d7a59de0c2ed1305bc0882d19df68e7a0d62647bbf"} Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.180134 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q84t7\" (UniqueName: \"kubernetes.io/projected/9fffb916-a3fd-400a-afd9-997515bf6850-kube-api-access-q84t7\") pod \"neutron-9524-account-create-update-fznw7\" (UID: \"9fffb916-a3fd-400a-afd9-997515bf6850\") " pod="openstack/neutron-9524-account-create-update-fznw7" Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.180169 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9fffb916-a3fd-400a-afd9-997515bf6850-operator-scripts\") pod \"neutron-9524-account-create-update-fznw7\" (UID: \"9fffb916-a3fd-400a-afd9-997515bf6850\") " pod="openstack/neutron-9524-account-create-update-fznw7" Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.181394 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9fffb916-a3fd-400a-afd9-997515bf6850-operator-scripts\") pod \"neutron-9524-account-create-update-fznw7\" (UID: \"9fffb916-a3fd-400a-afd9-997515bf6850\") " pod="openstack/neutron-9524-account-create-update-fznw7" Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.218976 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q84t7\" (UniqueName: \"kubernetes.io/projected/9fffb916-a3fd-400a-afd9-997515bf6850-kube-api-access-q84t7\") pod \"neutron-9524-account-create-update-fznw7\" (UID: \"9fffb916-a3fd-400a-afd9-997515bf6850\") " pod="openstack/neutron-9524-account-create-update-fznw7" Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.423270 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9524-account-create-update-fznw7" Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.495641 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4f9eed1-5dc4-4e60-b35a-812bc273acc6" path="/var/lib/kubelet/pods/c4f9eed1-5dc4-4e60-b35a-812bc273acc6/volumes" Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.556866 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-32f0-account-create-update-7fwpt"] Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.688686 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-vnl9g"] Mar 18 20:25:33 crc kubenswrapper[4950]: W0318 20:25:33.701510 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3abd28b2_fbc4_4b22_8f6f_a34917f8cba3.slice/crio-0e003b9b05a373053cb80305eb2d413c764510885194bf88b0363e08a6db93aa WatchSource:0}: Error finding container 0e003b9b05a373053cb80305eb2d413c764510885194bf88b0363e08a6db93aa: Status 404 returned error can't find the container with id 0e003b9b05a373053cb80305eb2d413c764510885194bf88b0363e08a6db93aa Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.836985 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.837030 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.837069 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.837857 4950 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"80bb5b9fd4b42a17689aca26b1eb183f4363b2ddacc1fdf73b829648cfa1e873"} pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.837902 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" containerID="cri-o://80bb5b9fd4b42a17689aca26b1eb183f4363b2ddacc1fdf73b829648cfa1e873" gracePeriod=600 Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.879117 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-jz448"] Mar 18 20:25:33 crc kubenswrapper[4950]: W0318 20:25:33.896721 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90aa20e7_e9c9_46c5_ad83_8004caa09547.slice/crio-c44f6f548be434e0ef465991ec6ec1b9c0da1760fb61fc605eaf2da86cb2efaa WatchSource:0}: Error finding container c44f6f548be434e0ef465991ec6ec1b9c0da1760fb61fc605eaf2da86cb2efaa: Status 404 returned error can't find the container with id c44f6f548be434e0ef465991ec6ec1b9c0da1760fb61fc605eaf2da86cb2efaa Mar 18 20:25:33 crc kubenswrapper[4950]: I0318 20:25:33.897726 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-t4zg5"] Mar 18 20:25:33 crc kubenswrapper[4950]: W0318 20:25:33.900191 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0af1dd1d_4630_49a6_9f2d_a056fefecf78.slice/crio-0c81ffe6f0e0ef98d9cfd0e525714ac1d341a1b518a79eddb52d1971e2d24740 WatchSource:0}: Error finding container 0c81ffe6f0e0ef98d9cfd0e525714ac1d341a1b518a79eddb52d1971e2d24740: Status 404 returned error can't find the container with id 0c81ffe6f0e0ef98d9cfd0e525714ac1d341a1b518a79eddb52d1971e2d24740 Mar 18 20:25:34 crc kubenswrapper[4950]: I0318 20:25:34.019062 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9524-account-create-update-fznw7"] Mar 18 20:25:34 crc kubenswrapper[4950]: I0318 20:25:34.035390 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-7fe1-account-create-update-9rdjc"] Mar 18 20:25:34 crc kubenswrapper[4950]: W0318 20:25:34.038157 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fffb916_a3fd_400a_afd9_997515bf6850.slice/crio-dae6b3a58858a4a1b24ec47017247449f75fd23568d82d99134cc73887d191c5 WatchSource:0}: Error finding container dae6b3a58858a4a1b24ec47017247449f75fd23568d82d99134cc73887d191c5: Status 404 returned error can't find the container with id dae6b3a58858a4a1b24ec47017247449f75fd23568d82d99134cc73887d191c5 Mar 18 20:25:34 crc kubenswrapper[4950]: W0318 20:25:34.041461 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod875a302d_40dd_4c0f_a729_cd6f3dcb4956.slice/crio-cda1640e2c3ee2d9c30a1b99be0635a43fb0a2a3edc40bc510e81d4bb3b681f8 WatchSource:0}: Error finding container cda1640e2c3ee2d9c30a1b99be0635a43fb0a2a3edc40bc510e81d4bb3b681f8: Status 404 returned error can't find the container with id cda1640e2c3ee2d9c30a1b99be0635a43fb0a2a3edc40bc510e81d4bb3b681f8 Mar 18 20:25:34 crc kubenswrapper[4950]: I0318 20:25:34.181183 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jz448" event={"ID":"90aa20e7-e9c9-46c5-ad83-8004caa09547","Type":"ContainerStarted","Data":"c44f6f548be434e0ef465991ec6ec1b9c0da1760fb61fc605eaf2da86cb2efaa"} Mar 18 20:25:34 crc kubenswrapper[4950]: I0318 20:25:34.197648 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-vnl9g" event={"ID":"3abd28b2-fbc4-4b22-8f6f-a34917f8cba3","Type":"ContainerStarted","Data":"f75e0f84cb3b8ea848df445d53cb2dffeb6e6d3f41d31f462db1e9d5dc29c9e2"} Mar 18 20:25:34 crc kubenswrapper[4950]: I0318 20:25:34.197691 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-vnl9g" event={"ID":"3abd28b2-fbc4-4b22-8f6f-a34917f8cba3","Type":"ContainerStarted","Data":"0e003b9b05a373053cb80305eb2d413c764510885194bf88b0363e08a6db93aa"} Mar 18 20:25:34 crc kubenswrapper[4950]: I0318 20:25:34.206837 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-32f0-account-create-update-7fwpt" event={"ID":"e5a72833-ca01-4da1-84e9-2b144ac64814","Type":"ContainerStarted","Data":"2a537136389097a94c9a681b9048c1051b7469641cbea118bd175be37c57658e"} Mar 18 20:25:34 crc kubenswrapper[4950]: I0318 20:25:34.206878 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-32f0-account-create-update-7fwpt" event={"ID":"e5a72833-ca01-4da1-84e9-2b144ac64814","Type":"ContainerStarted","Data":"dde8c023c040a83669428cd870038858a44913ee848b849d7d65c9e70907658d"} Mar 18 20:25:34 crc kubenswrapper[4950]: I0318 20:25:34.211639 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9524-account-create-update-fznw7" event={"ID":"9fffb916-a3fd-400a-afd9-997515bf6850","Type":"ContainerStarted","Data":"dae6b3a58858a4a1b24ec47017247449f75fd23568d82d99134cc73887d191c5"} Mar 18 20:25:34 crc kubenswrapper[4950]: I0318 20:25:34.217387 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7fe1-account-create-update-9rdjc" event={"ID":"875a302d-40dd-4c0f-a729-cd6f3dcb4956","Type":"ContainerStarted","Data":"cda1640e2c3ee2d9c30a1b99be0635a43fb0a2a3edc40bc510e81d4bb3b681f8"} Mar 18 20:25:34 crc kubenswrapper[4950]: I0318 20:25:34.225742 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-t4zg5" event={"ID":"0af1dd1d-4630-49a6-9f2d-a056fefecf78","Type":"ContainerStarted","Data":"0c81ffe6f0e0ef98d9cfd0e525714ac1d341a1b518a79eddb52d1971e2d24740"} Mar 18 20:25:34 crc kubenswrapper[4950]: I0318 20:25:34.232994 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-vnl9g" podStartSLOduration=2.23297632 podStartE2EDuration="2.23297632s" podCreationTimestamp="2026-03-18 20:25:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:25:34.223027924 +0000 UTC m=+1147.463869792" watchObservedRunningTime="2026-03-18 20:25:34.23297632 +0000 UTC m=+1147.473818188" Mar 18 20:25:34 crc kubenswrapper[4950]: I0318 20:25:34.237481 4950 generic.go:334] "Generic (PLEG): container finished" podID="4048b439-3266-46e7-9de0-22377efacc46" containerID="80bb5b9fd4b42a17689aca26b1eb183f4363b2ddacc1fdf73b829648cfa1e873" exitCode=0 Mar 18 20:25:34 crc kubenswrapper[4950]: I0318 20:25:34.237563 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerDied","Data":"80bb5b9fd4b42a17689aca26b1eb183f4363b2ddacc1fdf73b829648cfa1e873"} Mar 18 20:25:34 crc kubenswrapper[4950]: I0318 20:25:34.237647 4950 scope.go:117] "RemoveContainer" containerID="bef1deca2fab314ca277f2c06d049ce203ead40181a917f46a9164a0f3e058c3" Mar 18 20:25:34 crc kubenswrapper[4950]: I0318 20:25:34.248748 4950 generic.go:334] "Generic (PLEG): container finished" podID="24db2c06-d91e-4fc1-bcb5-9f29546368e2" containerID="7b94ccfce68784ddcfbf806e3c9c79f6543d2028916875e5833b74b17696f3a3" exitCode=0 Mar 18 20:25:34 crc kubenswrapper[4950]: I0318 20:25:34.248788 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-z2ffj" event={"ID":"24db2c06-d91e-4fc1-bcb5-9f29546368e2","Type":"ContainerDied","Data":"7b94ccfce68784ddcfbf806e3c9c79f6543d2028916875e5833b74b17696f3a3"} Mar 18 20:25:34 crc kubenswrapper[4950]: I0318 20:25:34.260426 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-32f0-account-create-update-7fwpt" podStartSLOduration=2.260391636 podStartE2EDuration="2.260391636s" podCreationTimestamp="2026-03-18 20:25:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:25:34.254679869 +0000 UTC m=+1147.495521737" watchObservedRunningTime="2026-03-18 20:25:34.260391636 +0000 UTC m=+1147.501233494" Mar 18 20:25:34 crc kubenswrapper[4950]: I0318 20:25:34.302048 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-t4zg5" podStartSLOduration=2.302029788 podStartE2EDuration="2.302029788s" podCreationTimestamp="2026-03-18 20:25:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:25:34.287801142 +0000 UTC m=+1147.528643020" watchObservedRunningTime="2026-03-18 20:25:34.302029788 +0000 UTC m=+1147.542871656" Mar 18 20:25:35 crc kubenswrapper[4950]: I0318 20:25:35.260731 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerStarted","Data":"4e352dfc9da032a31ef3fccebe090411fdd509a1d2eb6df9588325dd52adecda"} Mar 18 20:25:35 crc kubenswrapper[4950]: I0318 20:25:35.267728 4950 generic.go:334] "Generic (PLEG): container finished" podID="3abd28b2-fbc4-4b22-8f6f-a34917f8cba3" containerID="f75e0f84cb3b8ea848df445d53cb2dffeb6e6d3f41d31f462db1e9d5dc29c9e2" exitCode=0 Mar 18 20:25:35 crc kubenswrapper[4950]: I0318 20:25:35.267954 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-vnl9g" event={"ID":"3abd28b2-fbc4-4b22-8f6f-a34917f8cba3","Type":"ContainerDied","Data":"f75e0f84cb3b8ea848df445d53cb2dffeb6e6d3f41d31f462db1e9d5dc29c9e2"} Mar 18 20:25:35 crc kubenswrapper[4950]: I0318 20:25:35.272190 4950 generic.go:334] "Generic (PLEG): container finished" podID="e5a72833-ca01-4da1-84e9-2b144ac64814" containerID="2a537136389097a94c9a681b9048c1051b7469641cbea118bd175be37c57658e" exitCode=0 Mar 18 20:25:35 crc kubenswrapper[4950]: I0318 20:25:35.272311 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-32f0-account-create-update-7fwpt" event={"ID":"e5a72833-ca01-4da1-84e9-2b144ac64814","Type":"ContainerDied","Data":"2a537136389097a94c9a681b9048c1051b7469641cbea118bd175be37c57658e"} Mar 18 20:25:35 crc kubenswrapper[4950]: I0318 20:25:35.282100 4950 generic.go:334] "Generic (PLEG): container finished" podID="875a302d-40dd-4c0f-a729-cd6f3dcb4956" containerID="5cd1f30806cdb193c9d3a294c6fe419e5156658820e1bb223ecc369c4dfc440e" exitCode=0 Mar 18 20:25:35 crc kubenswrapper[4950]: I0318 20:25:35.282274 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7fe1-account-create-update-9rdjc" event={"ID":"875a302d-40dd-4c0f-a729-cd6f3dcb4956","Type":"ContainerDied","Data":"5cd1f30806cdb193c9d3a294c6fe419e5156658820e1bb223ecc369c4dfc440e"} Mar 18 20:25:35 crc kubenswrapper[4950]: I0318 20:25:35.285571 4950 generic.go:334] "Generic (PLEG): container finished" podID="9fffb916-a3fd-400a-afd9-997515bf6850" containerID="5cf81e528e7dd2c3faddabd82b95a20fa458fcb198d8aa1978a9e8bbe30ec8cf" exitCode=0 Mar 18 20:25:35 crc kubenswrapper[4950]: I0318 20:25:35.285611 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9524-account-create-update-fznw7" event={"ID":"9fffb916-a3fd-400a-afd9-997515bf6850","Type":"ContainerDied","Data":"5cf81e528e7dd2c3faddabd82b95a20fa458fcb198d8aa1978a9e8bbe30ec8cf"} Mar 18 20:25:35 crc kubenswrapper[4950]: I0318 20:25:35.289808 4950 generic.go:334] "Generic (PLEG): container finished" podID="0af1dd1d-4630-49a6-9f2d-a056fefecf78" containerID="2cb382e16134e5b58b9d66cfbefac3335e78588db4a5110f353dc456e45a0327" exitCode=0 Mar 18 20:25:35 crc kubenswrapper[4950]: I0318 20:25:35.289843 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-t4zg5" event={"ID":"0af1dd1d-4630-49a6-9f2d-a056fefecf78","Type":"ContainerDied","Data":"2cb382e16134e5b58b9d66cfbefac3335e78588db4a5110f353dc456e45a0327"} Mar 18 20:25:35 crc kubenswrapper[4950]: I0318 20:25:35.809001 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-z2ffj" Mar 18 20:25:35 crc kubenswrapper[4950]: I0318 20:25:35.935161 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24db2c06-d91e-4fc1-bcb5-9f29546368e2-operator-scripts\") pod \"24db2c06-d91e-4fc1-bcb5-9f29546368e2\" (UID: \"24db2c06-d91e-4fc1-bcb5-9f29546368e2\") " Mar 18 20:25:35 crc kubenswrapper[4950]: I0318 20:25:35.935375 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5f9t\" (UniqueName: \"kubernetes.io/projected/24db2c06-d91e-4fc1-bcb5-9f29546368e2-kube-api-access-q5f9t\") pod \"24db2c06-d91e-4fc1-bcb5-9f29546368e2\" (UID: \"24db2c06-d91e-4fc1-bcb5-9f29546368e2\") " Mar 18 20:25:35 crc kubenswrapper[4950]: I0318 20:25:35.936268 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24db2c06-d91e-4fc1-bcb5-9f29546368e2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "24db2c06-d91e-4fc1-bcb5-9f29546368e2" (UID: "24db2c06-d91e-4fc1-bcb5-9f29546368e2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:25:35 crc kubenswrapper[4950]: I0318 20:25:35.937382 4950 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24db2c06-d91e-4fc1-bcb5-9f29546368e2-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:35 crc kubenswrapper[4950]: I0318 20:25:35.941942 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24db2c06-d91e-4fc1-bcb5-9f29546368e2-kube-api-access-q5f9t" (OuterVolumeSpecName: "kube-api-access-q5f9t") pod "24db2c06-d91e-4fc1-bcb5-9f29546368e2" (UID: "24db2c06-d91e-4fc1-bcb5-9f29546368e2"). InnerVolumeSpecName "kube-api-access-q5f9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:25:36 crc kubenswrapper[4950]: I0318 20:25:36.039108 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5f9t\" (UniqueName: \"kubernetes.io/projected/24db2c06-d91e-4fc1-bcb5-9f29546368e2-kube-api-access-q5f9t\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:36 crc kubenswrapper[4950]: I0318 20:25:36.303495 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-z2ffj" Mar 18 20:25:36 crc kubenswrapper[4950]: I0318 20:25:36.303397 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-z2ffj" event={"ID":"24db2c06-d91e-4fc1-bcb5-9f29546368e2","Type":"ContainerDied","Data":"85ed251041c207da1a1063d7a59de0c2ed1305bc0882d19df68e7a0d62647bbf"} Mar 18 20:25:36 crc kubenswrapper[4950]: I0318 20:25:36.303772 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85ed251041c207da1a1063d7a59de0c2ed1305bc0882d19df68e7a0d62647bbf" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.086941 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-t4zg5" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.143947 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-32f0-account-create-update-7fwpt" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.156216 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7fe1-account-create-update-9rdjc" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.171821 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-vnl9g" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.193435 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0af1dd1d-4630-49a6-9f2d-a056fefecf78-operator-scripts\") pod \"0af1dd1d-4630-49a6-9f2d-a056fefecf78\" (UID: \"0af1dd1d-4630-49a6-9f2d-a056fefecf78\") " Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.193638 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzwd4\" (UniqueName: \"kubernetes.io/projected/0af1dd1d-4630-49a6-9f2d-a056fefecf78-kube-api-access-lzwd4\") pod \"0af1dd1d-4630-49a6-9f2d-a056fefecf78\" (UID: \"0af1dd1d-4630-49a6-9f2d-a056fefecf78\") " Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.194307 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0af1dd1d-4630-49a6-9f2d-a056fefecf78-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0af1dd1d-4630-49a6-9f2d-a056fefecf78" (UID: "0af1dd1d-4630-49a6-9f2d-a056fefecf78"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.203818 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0af1dd1d-4630-49a6-9f2d-a056fefecf78-kube-api-access-lzwd4" (OuterVolumeSpecName: "kube-api-access-lzwd4") pod "0af1dd1d-4630-49a6-9f2d-a056fefecf78" (UID: "0af1dd1d-4630-49a6-9f2d-a056fefecf78"). InnerVolumeSpecName "kube-api-access-lzwd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.204978 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9524-account-create-update-fznw7" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.295911 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/875a302d-40dd-4c0f-a729-cd6f3dcb4956-operator-scripts\") pod \"875a302d-40dd-4c0f-a729-cd6f3dcb4956\" (UID: \"875a302d-40dd-4c0f-a729-cd6f3dcb4956\") " Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.295984 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8n65g\" (UniqueName: \"kubernetes.io/projected/3abd28b2-fbc4-4b22-8f6f-a34917f8cba3-kube-api-access-8n65g\") pod \"3abd28b2-fbc4-4b22-8f6f-a34917f8cba3\" (UID: \"3abd28b2-fbc4-4b22-8f6f-a34917f8cba3\") " Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.296032 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5a72833-ca01-4da1-84e9-2b144ac64814-operator-scripts\") pod \"e5a72833-ca01-4da1-84e9-2b144ac64814\" (UID: \"e5a72833-ca01-4da1-84e9-2b144ac64814\") " Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.296085 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82mq2\" (UniqueName: \"kubernetes.io/projected/e5a72833-ca01-4da1-84e9-2b144ac64814-kube-api-access-82mq2\") pod \"e5a72833-ca01-4da1-84e9-2b144ac64814\" (UID: \"e5a72833-ca01-4da1-84e9-2b144ac64814\") " Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.296129 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kdz9\" (UniqueName: \"kubernetes.io/projected/875a302d-40dd-4c0f-a729-cd6f3dcb4956-kube-api-access-7kdz9\") pod \"875a302d-40dd-4c0f-a729-cd6f3dcb4956\" (UID: \"875a302d-40dd-4c0f-a729-cd6f3dcb4956\") " Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.296146 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q84t7\" (UniqueName: \"kubernetes.io/projected/9fffb916-a3fd-400a-afd9-997515bf6850-kube-api-access-q84t7\") pod \"9fffb916-a3fd-400a-afd9-997515bf6850\" (UID: \"9fffb916-a3fd-400a-afd9-997515bf6850\") " Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.296224 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9fffb916-a3fd-400a-afd9-997515bf6850-operator-scripts\") pod \"9fffb916-a3fd-400a-afd9-997515bf6850\" (UID: \"9fffb916-a3fd-400a-afd9-997515bf6850\") " Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.296242 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3abd28b2-fbc4-4b22-8f6f-a34917f8cba3-operator-scripts\") pod \"3abd28b2-fbc4-4b22-8f6f-a34917f8cba3\" (UID: \"3abd28b2-fbc4-4b22-8f6f-a34917f8cba3\") " Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.296597 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzwd4\" (UniqueName: \"kubernetes.io/projected/0af1dd1d-4630-49a6-9f2d-a056fefecf78-kube-api-access-lzwd4\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.296621 4950 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0af1dd1d-4630-49a6-9f2d-a056fefecf78-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.297106 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/875a302d-40dd-4c0f-a729-cd6f3dcb4956-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "875a302d-40dd-4c0f-a729-cd6f3dcb4956" (UID: "875a302d-40dd-4c0f-a729-cd6f3dcb4956"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.297333 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fffb916-a3fd-400a-afd9-997515bf6850-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9fffb916-a3fd-400a-afd9-997515bf6850" (UID: "9fffb916-a3fd-400a-afd9-997515bf6850"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.297388 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3abd28b2-fbc4-4b22-8f6f-a34917f8cba3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3abd28b2-fbc4-4b22-8f6f-a34917f8cba3" (UID: "3abd28b2-fbc4-4b22-8f6f-a34917f8cba3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.297492 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5a72833-ca01-4da1-84e9-2b144ac64814-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e5a72833-ca01-4da1-84e9-2b144ac64814" (UID: "e5a72833-ca01-4da1-84e9-2b144ac64814"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.299877 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5a72833-ca01-4da1-84e9-2b144ac64814-kube-api-access-82mq2" (OuterVolumeSpecName: "kube-api-access-82mq2") pod "e5a72833-ca01-4da1-84e9-2b144ac64814" (UID: "e5a72833-ca01-4da1-84e9-2b144ac64814"). InnerVolumeSpecName "kube-api-access-82mq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.300535 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fffb916-a3fd-400a-afd9-997515bf6850-kube-api-access-q84t7" (OuterVolumeSpecName: "kube-api-access-q84t7") pod "9fffb916-a3fd-400a-afd9-997515bf6850" (UID: "9fffb916-a3fd-400a-afd9-997515bf6850"). InnerVolumeSpecName "kube-api-access-q84t7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.300798 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/875a302d-40dd-4c0f-a729-cd6f3dcb4956-kube-api-access-7kdz9" (OuterVolumeSpecName: "kube-api-access-7kdz9") pod "875a302d-40dd-4c0f-a729-cd6f3dcb4956" (UID: "875a302d-40dd-4c0f-a729-cd6f3dcb4956"). InnerVolumeSpecName "kube-api-access-7kdz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.300848 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3abd28b2-fbc4-4b22-8f6f-a34917f8cba3-kube-api-access-8n65g" (OuterVolumeSpecName: "kube-api-access-8n65g") pod "3abd28b2-fbc4-4b22-8f6f-a34917f8cba3" (UID: "3abd28b2-fbc4-4b22-8f6f-a34917f8cba3"). InnerVolumeSpecName "kube-api-access-8n65g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.328551 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jz448" event={"ID":"90aa20e7-e9c9-46c5-ad83-8004caa09547","Type":"ContainerStarted","Data":"2b08ed8575e0d50da93bca56a3da39611d55b13bb3028a048104c7cb5b26993d"} Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.329427 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-vnl9g" event={"ID":"3abd28b2-fbc4-4b22-8f6f-a34917f8cba3","Type":"ContainerDied","Data":"0e003b9b05a373053cb80305eb2d413c764510885194bf88b0363e08a6db93aa"} Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.329456 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e003b9b05a373053cb80305eb2d413c764510885194bf88b0363e08a6db93aa" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.329460 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-vnl9g" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.330179 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-32f0-account-create-update-7fwpt" event={"ID":"e5a72833-ca01-4da1-84e9-2b144ac64814","Type":"ContainerDied","Data":"dde8c023c040a83669428cd870038858a44913ee848b849d7d65c9e70907658d"} Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.330197 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dde8c023c040a83669428cd870038858a44913ee848b849d7d65c9e70907658d" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.330229 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-32f0-account-create-update-7fwpt" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.332308 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9524-account-create-update-fznw7" event={"ID":"9fffb916-a3fd-400a-afd9-997515bf6850","Type":"ContainerDied","Data":"dae6b3a58858a4a1b24ec47017247449f75fd23568d82d99134cc73887d191c5"} Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.332321 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9524-account-create-update-fznw7" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.332332 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dae6b3a58858a4a1b24ec47017247449f75fd23568d82d99134cc73887d191c5" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.338624 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7fe1-account-create-update-9rdjc" event={"ID":"875a302d-40dd-4c0f-a729-cd6f3dcb4956","Type":"ContainerDied","Data":"cda1640e2c3ee2d9c30a1b99be0635a43fb0a2a3edc40bc510e81d4bb3b681f8"} Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.338680 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cda1640e2c3ee2d9c30a1b99be0635a43fb0a2a3edc40bc510e81d4bb3b681f8" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.338857 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7fe1-account-create-update-9rdjc" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.341931 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-t4zg5" event={"ID":"0af1dd1d-4630-49a6-9f2d-a056fefecf78","Type":"ContainerDied","Data":"0c81ffe6f0e0ef98d9cfd0e525714ac1d341a1b518a79eddb52d1971e2d24740"} Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.341960 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c81ffe6f0e0ef98d9cfd0e525714ac1d341a1b518a79eddb52d1971e2d24740" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.341999 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-t4zg5" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.353427 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-jz448" podStartSLOduration=2.249501076 podStartE2EDuration="7.353378149s" podCreationTimestamp="2026-03-18 20:25:32 +0000 UTC" firstStartedPulling="2026-03-18 20:25:33.907161081 +0000 UTC m=+1147.148002949" lastFinishedPulling="2026-03-18 20:25:39.011038154 +0000 UTC m=+1152.251880022" observedRunningTime="2026-03-18 20:25:39.349171441 +0000 UTC m=+1152.590013309" watchObservedRunningTime="2026-03-18 20:25:39.353378149 +0000 UTC m=+1152.594220017" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.398037 4950 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9fffb916-a3fd-400a-afd9-997515bf6850-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.398069 4950 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3abd28b2-fbc4-4b22-8f6f-a34917f8cba3-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.398079 4950 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/875a302d-40dd-4c0f-a729-cd6f3dcb4956-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.398088 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8n65g\" (UniqueName: \"kubernetes.io/projected/3abd28b2-fbc4-4b22-8f6f-a34917f8cba3-kube-api-access-8n65g\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.398100 4950 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5a72833-ca01-4da1-84e9-2b144ac64814-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.398109 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82mq2\" (UniqueName: \"kubernetes.io/projected/e5a72833-ca01-4da1-84e9-2b144ac64814-kube-api-access-82mq2\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.398116 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kdz9\" (UniqueName: \"kubernetes.io/projected/875a302d-40dd-4c0f-a729-cd6f3dcb4956-kube-api-access-7kdz9\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:39 crc kubenswrapper[4950]: I0318 20:25:39.398124 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q84t7\" (UniqueName: \"kubernetes.io/projected/9fffb916-a3fd-400a-afd9-997515bf6850-kube-api-access-q84t7\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:43 crc kubenswrapper[4950]: I0318 20:25:43.378132 4950 generic.go:334] "Generic (PLEG): container finished" podID="90aa20e7-e9c9-46c5-ad83-8004caa09547" containerID="2b08ed8575e0d50da93bca56a3da39611d55b13bb3028a048104c7cb5b26993d" exitCode=0 Mar 18 20:25:43 crc kubenswrapper[4950]: I0318 20:25:43.378201 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jz448" event={"ID":"90aa20e7-e9c9-46c5-ad83-8004caa09547","Type":"ContainerDied","Data":"2b08ed8575e0d50da93bca56a3da39611d55b13bb3028a048104c7cb5b26993d"} Mar 18 20:25:44 crc kubenswrapper[4950]: I0318 20:25:44.709357 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jz448" Mar 18 20:25:44 crc kubenswrapper[4950]: I0318 20:25:44.799263 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90aa20e7-e9c9-46c5-ad83-8004caa09547-combined-ca-bundle\") pod \"90aa20e7-e9c9-46c5-ad83-8004caa09547\" (UID: \"90aa20e7-e9c9-46c5-ad83-8004caa09547\") " Mar 18 20:25:44 crc kubenswrapper[4950]: I0318 20:25:44.800118 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90aa20e7-e9c9-46c5-ad83-8004caa09547-config-data\") pod \"90aa20e7-e9c9-46c5-ad83-8004caa09547\" (UID: \"90aa20e7-e9c9-46c5-ad83-8004caa09547\") " Mar 18 20:25:44 crc kubenswrapper[4950]: I0318 20:25:44.800227 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfgtc\" (UniqueName: \"kubernetes.io/projected/90aa20e7-e9c9-46c5-ad83-8004caa09547-kube-api-access-cfgtc\") pod \"90aa20e7-e9c9-46c5-ad83-8004caa09547\" (UID: \"90aa20e7-e9c9-46c5-ad83-8004caa09547\") " Mar 18 20:25:44 crc kubenswrapper[4950]: I0318 20:25:44.804896 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90aa20e7-e9c9-46c5-ad83-8004caa09547-kube-api-access-cfgtc" (OuterVolumeSpecName: "kube-api-access-cfgtc") pod "90aa20e7-e9c9-46c5-ad83-8004caa09547" (UID: "90aa20e7-e9c9-46c5-ad83-8004caa09547"). InnerVolumeSpecName "kube-api-access-cfgtc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:25:44 crc kubenswrapper[4950]: I0318 20:25:44.825939 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90aa20e7-e9c9-46c5-ad83-8004caa09547-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "90aa20e7-e9c9-46c5-ad83-8004caa09547" (UID: "90aa20e7-e9c9-46c5-ad83-8004caa09547"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:25:44 crc kubenswrapper[4950]: I0318 20:25:44.846051 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90aa20e7-e9c9-46c5-ad83-8004caa09547-config-data" (OuterVolumeSpecName: "config-data") pod "90aa20e7-e9c9-46c5-ad83-8004caa09547" (UID: "90aa20e7-e9c9-46c5-ad83-8004caa09547"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:25:44 crc kubenswrapper[4950]: I0318 20:25:44.902095 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfgtc\" (UniqueName: \"kubernetes.io/projected/90aa20e7-e9c9-46c5-ad83-8004caa09547-kube-api-access-cfgtc\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:44 crc kubenswrapper[4950]: I0318 20:25:44.902381 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90aa20e7-e9c9-46c5-ad83-8004caa09547-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:44 crc kubenswrapper[4950]: I0318 20:25:44.902512 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90aa20e7-e9c9-46c5-ad83-8004caa09547-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.410306 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jz448" event={"ID":"90aa20e7-e9c9-46c5-ad83-8004caa09547","Type":"ContainerDied","Data":"c44f6f548be434e0ef465991ec6ec1b9c0da1760fb61fc605eaf2da86cb2efaa"} Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.410355 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c44f6f548be434e0ef465991ec6ec1b9c0da1760fb61fc605eaf2da86cb2efaa" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.410451 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jz448" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.606199 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-2bbb2"] Mar 18 20:25:45 crc kubenswrapper[4950]: E0318 20:25:45.606510 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0af1dd1d-4630-49a6-9f2d-a056fefecf78" containerName="mariadb-database-create" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.606525 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="0af1dd1d-4630-49a6-9f2d-a056fefecf78" containerName="mariadb-database-create" Mar 18 20:25:45 crc kubenswrapper[4950]: E0318 20:25:45.606539 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90aa20e7-e9c9-46c5-ad83-8004caa09547" containerName="keystone-db-sync" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.606545 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="90aa20e7-e9c9-46c5-ad83-8004caa09547" containerName="keystone-db-sync" Mar 18 20:25:45 crc kubenswrapper[4950]: E0318 20:25:45.606554 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5a72833-ca01-4da1-84e9-2b144ac64814" containerName="mariadb-account-create-update" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.606559 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5a72833-ca01-4da1-84e9-2b144ac64814" containerName="mariadb-account-create-update" Mar 18 20:25:45 crc kubenswrapper[4950]: E0318 20:25:45.606571 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3abd28b2-fbc4-4b22-8f6f-a34917f8cba3" containerName="mariadb-database-create" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.606577 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="3abd28b2-fbc4-4b22-8f6f-a34917f8cba3" containerName="mariadb-database-create" Mar 18 20:25:45 crc kubenswrapper[4950]: E0318 20:25:45.606589 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24db2c06-d91e-4fc1-bcb5-9f29546368e2" containerName="mariadb-database-create" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.606594 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="24db2c06-d91e-4fc1-bcb5-9f29546368e2" containerName="mariadb-database-create" Mar 18 20:25:45 crc kubenswrapper[4950]: E0318 20:25:45.606610 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fffb916-a3fd-400a-afd9-997515bf6850" containerName="mariadb-account-create-update" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.606616 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fffb916-a3fd-400a-afd9-997515bf6850" containerName="mariadb-account-create-update" Mar 18 20:25:45 crc kubenswrapper[4950]: E0318 20:25:45.606625 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="875a302d-40dd-4c0f-a729-cd6f3dcb4956" containerName="mariadb-account-create-update" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.606630 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="875a302d-40dd-4c0f-a729-cd6f3dcb4956" containerName="mariadb-account-create-update" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.606774 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="24db2c06-d91e-4fc1-bcb5-9f29546368e2" containerName="mariadb-database-create" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.606785 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="0af1dd1d-4630-49a6-9f2d-a056fefecf78" containerName="mariadb-database-create" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.606792 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="875a302d-40dd-4c0f-a729-cd6f3dcb4956" containerName="mariadb-account-create-update" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.606804 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5a72833-ca01-4da1-84e9-2b144ac64814" containerName="mariadb-account-create-update" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.606815 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="3abd28b2-fbc4-4b22-8f6f-a34917f8cba3" containerName="mariadb-database-create" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.606826 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fffb916-a3fd-400a-afd9-997515bf6850" containerName="mariadb-account-create-update" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.606835 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="90aa20e7-e9c9-46c5-ad83-8004caa09547" containerName="keystone-db-sync" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.607313 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2bbb2" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.608879 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-bsdbw" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.610362 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.611136 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.611877 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.613845 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.629450 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2bbb2"] Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.646355 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d6c8d744f-9mtrg"] Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.647933 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d6c8d744f-9mtrg" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.697131 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d6c8d744f-9mtrg"] Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.717140 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zskfs\" (UniqueName: \"kubernetes.io/projected/b2edc270-582d-4efe-84bc-f67ba8bbe372-kube-api-access-zskfs\") pod \"keystone-bootstrap-2bbb2\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " pod="openstack/keystone-bootstrap-2bbb2" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.717232 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-config-data\") pod \"keystone-bootstrap-2bbb2\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " pod="openstack/keystone-bootstrap-2bbb2" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.717259 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-scripts\") pod \"keystone-bootstrap-2bbb2\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " pod="openstack/keystone-bootstrap-2bbb2" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.717274 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-fernet-keys\") pod \"keystone-bootstrap-2bbb2\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " pod="openstack/keystone-bootstrap-2bbb2" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.717289 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-combined-ca-bundle\") pod \"keystone-bootstrap-2bbb2\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " pod="openstack/keystone-bootstrap-2bbb2" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.717336 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-credential-keys\") pod \"keystone-bootstrap-2bbb2\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " pod="openstack/keystone-bootstrap-2bbb2" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.818589 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-combined-ca-bundle\") pod \"keystone-bootstrap-2bbb2\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " pod="openstack/keystone-bootstrap-2bbb2" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.818830 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-ovsdbserver-nb\") pod \"dnsmasq-dns-5d6c8d744f-9mtrg\" (UID: \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\") " pod="openstack/dnsmasq-dns-5d6c8d744f-9mtrg" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.818920 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5529\" (UniqueName: \"kubernetes.io/projected/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-kube-api-access-k5529\") pod \"dnsmasq-dns-5d6c8d744f-9mtrg\" (UID: \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\") " pod="openstack/dnsmasq-dns-5d6c8d744f-9mtrg" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.818992 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-credential-keys\") pod \"keystone-bootstrap-2bbb2\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " pod="openstack/keystone-bootstrap-2bbb2" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.819058 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-dns-svc\") pod \"dnsmasq-dns-5d6c8d744f-9mtrg\" (UID: \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\") " pod="openstack/dnsmasq-dns-5d6c8d744f-9mtrg" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.819143 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zskfs\" (UniqueName: \"kubernetes.io/projected/b2edc270-582d-4efe-84bc-f67ba8bbe372-kube-api-access-zskfs\") pod \"keystone-bootstrap-2bbb2\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " pod="openstack/keystone-bootstrap-2bbb2" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.819215 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-config\") pod \"dnsmasq-dns-5d6c8d744f-9mtrg\" (UID: \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\") " pod="openstack/dnsmasq-dns-5d6c8d744f-9mtrg" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.819437 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-ovsdbserver-sb\") pod \"dnsmasq-dns-5d6c8d744f-9mtrg\" (UID: \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\") " pod="openstack/dnsmasq-dns-5d6c8d744f-9mtrg" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.819527 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-config-data\") pod \"keystone-bootstrap-2bbb2\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " pod="openstack/keystone-bootstrap-2bbb2" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.819580 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-scripts\") pod \"keystone-bootstrap-2bbb2\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " pod="openstack/keystone-bootstrap-2bbb2" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.819602 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-fernet-keys\") pod \"keystone-bootstrap-2bbb2\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " pod="openstack/keystone-bootstrap-2bbb2" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.829655 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-credential-keys\") pod \"keystone-bootstrap-2bbb2\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " pod="openstack/keystone-bootstrap-2bbb2" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.831355 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.833058 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-config-data\") pod \"keystone-bootstrap-2bbb2\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " pod="openstack/keystone-bootstrap-2bbb2" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.833570 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.835852 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-fernet-keys\") pod \"keystone-bootstrap-2bbb2\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " pod="openstack/keystone-bootstrap-2bbb2" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.841084 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-combined-ca-bundle\") pod \"keystone-bootstrap-2bbb2\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " pod="openstack/keystone-bootstrap-2bbb2" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.851914 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-scripts\") pod \"keystone-bootstrap-2bbb2\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " pod="openstack/keystone-bootstrap-2bbb2" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.852161 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.852852 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.856130 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zskfs\" (UniqueName: \"kubernetes.io/projected/b2edc270-582d-4efe-84bc-f67ba8bbe372-kube-api-access-zskfs\") pod \"keystone-bootstrap-2bbb2\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " pod="openstack/keystone-bootstrap-2bbb2" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.867829 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.921978 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvmgk\" (UniqueName: \"kubernetes.io/projected/7563774c-0a5c-4af9-802a-662f4e721f3c-kube-api-access-pvmgk\") pod \"ceilometer-0\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " pod="openstack/ceilometer-0" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.922285 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-config\") pod \"dnsmasq-dns-5d6c8d744f-9mtrg\" (UID: \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\") " pod="openstack/dnsmasq-dns-5d6c8d744f-9mtrg" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.922322 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7563774c-0a5c-4af9-802a-662f4e721f3c-log-httpd\") pod \"ceilometer-0\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " pod="openstack/ceilometer-0" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.922344 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-config-data\") pod \"ceilometer-0\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " pod="openstack/ceilometer-0" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.922364 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7563774c-0a5c-4af9-802a-662f4e721f3c-run-httpd\") pod \"ceilometer-0\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " pod="openstack/ceilometer-0" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.922402 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-ovsdbserver-sb\") pod \"dnsmasq-dns-5d6c8d744f-9mtrg\" (UID: \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\") " pod="openstack/dnsmasq-dns-5d6c8d744f-9mtrg" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.922457 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " pod="openstack/ceilometer-0" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.922487 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-scripts\") pod \"ceilometer-0\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " pod="openstack/ceilometer-0" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.922518 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-ovsdbserver-nb\") pod \"dnsmasq-dns-5d6c8d744f-9mtrg\" (UID: \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\") " pod="openstack/dnsmasq-dns-5d6c8d744f-9mtrg" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.922540 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " pod="openstack/ceilometer-0" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.922567 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5529\" (UniqueName: \"kubernetes.io/projected/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-kube-api-access-k5529\") pod \"dnsmasq-dns-5d6c8d744f-9mtrg\" (UID: \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\") " pod="openstack/dnsmasq-dns-5d6c8d744f-9mtrg" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.922600 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-dns-svc\") pod \"dnsmasq-dns-5d6c8d744f-9mtrg\" (UID: \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\") " pod="openstack/dnsmasq-dns-5d6c8d744f-9mtrg" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.923562 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-ovsdbserver-sb\") pod \"dnsmasq-dns-5d6c8d744f-9mtrg\" (UID: \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\") " pod="openstack/dnsmasq-dns-5d6c8d744f-9mtrg" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.931116 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2bbb2" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.956067 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-ovsdbserver-nb\") pod \"dnsmasq-dns-5d6c8d744f-9mtrg\" (UID: \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\") " pod="openstack/dnsmasq-dns-5d6c8d744f-9mtrg" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.956621 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-config\") pod \"dnsmasq-dns-5d6c8d744f-9mtrg\" (UID: \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\") " pod="openstack/dnsmasq-dns-5d6c8d744f-9mtrg" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.959366 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-dns-svc\") pod \"dnsmasq-dns-5d6c8d744f-9mtrg\" (UID: \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\") " pod="openstack/dnsmasq-dns-5d6c8d744f-9mtrg" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.968625 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5529\" (UniqueName: \"kubernetes.io/projected/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-kube-api-access-k5529\") pod \"dnsmasq-dns-5d6c8d744f-9mtrg\" (UID: \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\") " pod="openstack/dnsmasq-dns-5d6c8d744f-9mtrg" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.973479 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-j5m65"] Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.974483 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-j5m65" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.975740 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d6c8d744f-9mtrg" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.976998 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.977437 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 18 20:25:45 crc kubenswrapper[4950]: I0318 20:25:45.977655 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-dz6tt" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.034344 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-scripts\") pod \"ceilometer-0\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " pod="openstack/ceilometer-0" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.034424 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " pod="openstack/ceilometer-0" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.034484 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvmgk\" (UniqueName: \"kubernetes.io/projected/7563774c-0a5c-4af9-802a-662f4e721f3c-kube-api-access-pvmgk\") pod \"ceilometer-0\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " pod="openstack/ceilometer-0" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.034517 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7563774c-0a5c-4af9-802a-662f4e721f3c-log-httpd\") pod \"ceilometer-0\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " pod="openstack/ceilometer-0" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.034540 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-config-data\") pod \"ceilometer-0\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " pod="openstack/ceilometer-0" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.034558 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7563774c-0a5c-4af9-802a-662f4e721f3c-run-httpd\") pod \"ceilometer-0\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " pod="openstack/ceilometer-0" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.034592 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " pod="openstack/ceilometer-0" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.033558 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-j5m65"] Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.044981 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-scripts\") pod \"ceilometer-0\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " pod="openstack/ceilometer-0" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.047818 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7563774c-0a5c-4af9-802a-662f4e721f3c-run-httpd\") pod \"ceilometer-0\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " pod="openstack/ceilometer-0" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.048199 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7563774c-0a5c-4af9-802a-662f4e721f3c-log-httpd\") pod \"ceilometer-0\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " pod="openstack/ceilometer-0" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.048360 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " pod="openstack/ceilometer-0" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.048959 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " pod="openstack/ceilometer-0" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.061113 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-config-data\") pod \"ceilometer-0\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " pod="openstack/ceilometer-0" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.070257 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvmgk\" (UniqueName: \"kubernetes.io/projected/7563774c-0a5c-4af9-802a-662f4e721f3c-kube-api-access-pvmgk\") pod \"ceilometer-0\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " pod="openstack/ceilometer-0" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.079146 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-brgnq"] Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.080188 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-brgnq" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.085667 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.085855 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.086164 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-dv945" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.128895 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-brgnq"] Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.137422 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/dbf245a4-e400-4fc2-9d2e-3788c6baf25d-config\") pod \"neutron-db-sync-j5m65\" (UID: \"dbf245a4-e400-4fc2-9d2e-3788c6baf25d\") " pod="openstack/neutron-db-sync-j5m65" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.139172 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8zwx\" (UniqueName: \"kubernetes.io/projected/dbf245a4-e400-4fc2-9d2e-3788c6baf25d-kube-api-access-g8zwx\") pod \"neutron-db-sync-j5m65\" (UID: \"dbf245a4-e400-4fc2-9d2e-3788c6baf25d\") " pod="openstack/neutron-db-sync-j5m65" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.139382 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbf245a4-e400-4fc2-9d2e-3788c6baf25d-combined-ca-bundle\") pod \"neutron-db-sync-j5m65\" (UID: \"dbf245a4-e400-4fc2-9d2e-3788c6baf25d\") " pod="openstack/neutron-db-sync-j5m65" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.148557 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-tpdmj"] Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.157461 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-tpdmj" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.166018 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-qsjx9" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.166267 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.183159 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d6c8d744f-9mtrg"] Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.199294 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-42zwb"] Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.200313 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-42zwb" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.206698 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.206722 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-twmnm" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.206922 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.212017 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-tpdmj"] Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.217661 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.234280 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-42zwb"] Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.240519 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbf245a4-e400-4fc2-9d2e-3788c6baf25d-combined-ca-bundle\") pod \"neutron-db-sync-j5m65\" (UID: \"dbf245a4-e400-4fc2-9d2e-3788c6baf25d\") " pod="openstack/neutron-db-sync-j5m65" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.240576 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-combined-ca-bundle\") pod \"cinder-db-sync-brgnq\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " pod="openstack/cinder-db-sync-brgnq" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.240828 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wd8n\" (UniqueName: \"kubernetes.io/projected/82a7c2a4-9f57-460e-8bac-419f14222502-kube-api-access-9wd8n\") pod \"cinder-db-sync-brgnq\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " pod="openstack/cinder-db-sync-brgnq" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.240964 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82a7c2a4-9f57-460e-8bac-419f14222502-etc-machine-id\") pod \"cinder-db-sync-brgnq\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " pod="openstack/cinder-db-sync-brgnq" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.240997 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-scripts\") pod \"cinder-db-sync-brgnq\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " pod="openstack/cinder-db-sync-brgnq" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.241081 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-db-sync-config-data\") pod \"cinder-db-sync-brgnq\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " pod="openstack/cinder-db-sync-brgnq" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.241161 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-config-data\") pod \"cinder-db-sync-brgnq\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " pod="openstack/cinder-db-sync-brgnq" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.241231 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/dbf245a4-e400-4fc2-9d2e-3788c6baf25d-config\") pod \"neutron-db-sync-j5m65\" (UID: \"dbf245a4-e400-4fc2-9d2e-3788c6baf25d\") " pod="openstack/neutron-db-sync-j5m65" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.248222 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/dbf245a4-e400-4fc2-9d2e-3788c6baf25d-config\") pod \"neutron-db-sync-j5m65\" (UID: \"dbf245a4-e400-4fc2-9d2e-3788c6baf25d\") " pod="openstack/neutron-db-sync-j5m65" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.249027 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8zwx\" (UniqueName: \"kubernetes.io/projected/dbf245a4-e400-4fc2-9d2e-3788c6baf25d-kube-api-access-g8zwx\") pod \"neutron-db-sync-j5m65\" (UID: \"dbf245a4-e400-4fc2-9d2e-3788c6baf25d\") " pod="openstack/neutron-db-sync-j5m65" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.249491 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74b5448f7f-zdrjg"] Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.254183 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbf245a4-e400-4fc2-9d2e-3788c6baf25d-combined-ca-bundle\") pod \"neutron-db-sync-j5m65\" (UID: \"dbf245a4-e400-4fc2-9d2e-3788c6baf25d\") " pod="openstack/neutron-db-sync-j5m65" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.260642 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.269540 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74b5448f7f-zdrjg"] Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.271729 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8zwx\" (UniqueName: \"kubernetes.io/projected/dbf245a4-e400-4fc2-9d2e-3788c6baf25d-kube-api-access-g8zwx\") pod \"neutron-db-sync-j5m65\" (UID: \"dbf245a4-e400-4fc2-9d2e-3788c6baf25d\") " pod="openstack/neutron-db-sync-j5m65" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.352720 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-combined-ca-bundle\") pod \"cinder-db-sync-brgnq\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " pod="openstack/cinder-db-sync-brgnq" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.352776 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wd8n\" (UniqueName: \"kubernetes.io/projected/82a7c2a4-9f57-460e-8bac-419f14222502-kube-api-access-9wd8n\") pod \"cinder-db-sync-brgnq\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " pod="openstack/cinder-db-sync-brgnq" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.352810 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a4bb9e1a-0694-4e00-8ba6-f2765f82b983-db-sync-config-data\") pod \"barbican-db-sync-tpdmj\" (UID: \"a4bb9e1a-0694-4e00-8ba6-f2765f82b983\") " pod="openstack/barbican-db-sync-tpdmj" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.352839 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5989cc6f-e4c7-4511-a148-293f683ede13-combined-ca-bundle\") pod \"placement-db-sync-42zwb\" (UID: \"5989cc6f-e4c7-4511-a148-293f683ede13\") " pod="openstack/placement-db-sync-42zwb" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.352863 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82a7c2a4-9f57-460e-8bac-419f14222502-etc-machine-id\") pod \"cinder-db-sync-brgnq\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " pod="openstack/cinder-db-sync-brgnq" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.352887 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-scripts\") pod \"cinder-db-sync-brgnq\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " pod="openstack/cinder-db-sync-brgnq" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.352912 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tjdf\" (UniqueName: \"kubernetes.io/projected/a4bb9e1a-0694-4e00-8ba6-f2765f82b983-kube-api-access-5tjdf\") pod \"barbican-db-sync-tpdmj\" (UID: \"a4bb9e1a-0694-4e00-8ba6-f2765f82b983\") " pod="openstack/barbican-db-sync-tpdmj" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.352935 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5989cc6f-e4c7-4511-a148-293f683ede13-scripts\") pod \"placement-db-sync-42zwb\" (UID: \"5989cc6f-e4c7-4511-a148-293f683ede13\") " pod="openstack/placement-db-sync-42zwb" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.352955 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-db-sync-config-data\") pod \"cinder-db-sync-brgnq\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " pod="openstack/cinder-db-sync-brgnq" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.352989 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8rdc\" (UniqueName: \"kubernetes.io/projected/5989cc6f-e4c7-4511-a148-293f683ede13-kube-api-access-d8rdc\") pod \"placement-db-sync-42zwb\" (UID: \"5989cc6f-e4c7-4511-a148-293f683ede13\") " pod="openstack/placement-db-sync-42zwb" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.353014 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-config-data\") pod \"cinder-db-sync-brgnq\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " pod="openstack/cinder-db-sync-brgnq" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.353035 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5989cc6f-e4c7-4511-a148-293f683ede13-logs\") pod \"placement-db-sync-42zwb\" (UID: \"5989cc6f-e4c7-4511-a148-293f683ede13\") " pod="openstack/placement-db-sync-42zwb" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.353080 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5989cc6f-e4c7-4511-a148-293f683ede13-config-data\") pod \"placement-db-sync-42zwb\" (UID: \"5989cc6f-e4c7-4511-a148-293f683ede13\") " pod="openstack/placement-db-sync-42zwb" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.353115 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4bb9e1a-0694-4e00-8ba6-f2765f82b983-combined-ca-bundle\") pod \"barbican-db-sync-tpdmj\" (UID: \"a4bb9e1a-0694-4e00-8ba6-f2765f82b983\") " pod="openstack/barbican-db-sync-tpdmj" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.353696 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82a7c2a4-9f57-460e-8bac-419f14222502-etc-machine-id\") pod \"cinder-db-sync-brgnq\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " pod="openstack/cinder-db-sync-brgnq" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.356775 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-scripts\") pod \"cinder-db-sync-brgnq\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " pod="openstack/cinder-db-sync-brgnq" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.358440 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-combined-ca-bundle\") pod \"cinder-db-sync-brgnq\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " pod="openstack/cinder-db-sync-brgnq" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.359726 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-db-sync-config-data\") pod \"cinder-db-sync-brgnq\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " pod="openstack/cinder-db-sync-brgnq" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.367383 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-config-data\") pod \"cinder-db-sync-brgnq\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " pod="openstack/cinder-db-sync-brgnq" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.416704 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wd8n\" (UniqueName: \"kubernetes.io/projected/82a7c2a4-9f57-460e-8bac-419f14222502-kube-api-access-9wd8n\") pod \"cinder-db-sync-brgnq\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " pod="openstack/cinder-db-sync-brgnq" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.419625 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-j5m65" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.427811 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-brgnq" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.458779 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5989cc6f-e4c7-4511-a148-293f683ede13-config-data\") pod \"placement-db-sync-42zwb\" (UID: \"5989cc6f-e4c7-4511-a148-293f683ede13\") " pod="openstack/placement-db-sync-42zwb" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.458829 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-ovsdbserver-sb\") pod \"dnsmasq-dns-74b5448f7f-zdrjg\" (UID: \"ad591d5a-4120-4285-9cca-417b19176795\") " pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.458886 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4bb9e1a-0694-4e00-8ba6-f2765f82b983-combined-ca-bundle\") pod \"barbican-db-sync-tpdmj\" (UID: \"a4bb9e1a-0694-4e00-8ba6-f2765f82b983\") " pod="openstack/barbican-db-sync-tpdmj" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.458912 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4qnj\" (UniqueName: \"kubernetes.io/projected/ad591d5a-4120-4285-9cca-417b19176795-kube-api-access-j4qnj\") pod \"dnsmasq-dns-74b5448f7f-zdrjg\" (UID: \"ad591d5a-4120-4285-9cca-417b19176795\") " pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.458937 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-config\") pod \"dnsmasq-dns-74b5448f7f-zdrjg\" (UID: \"ad591d5a-4120-4285-9cca-417b19176795\") " pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.459017 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-ovsdbserver-nb\") pod \"dnsmasq-dns-74b5448f7f-zdrjg\" (UID: \"ad591d5a-4120-4285-9cca-417b19176795\") " pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.459043 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a4bb9e1a-0694-4e00-8ba6-f2765f82b983-db-sync-config-data\") pod \"barbican-db-sync-tpdmj\" (UID: \"a4bb9e1a-0694-4e00-8ba6-f2765f82b983\") " pod="openstack/barbican-db-sync-tpdmj" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.459071 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5989cc6f-e4c7-4511-a148-293f683ede13-combined-ca-bundle\") pod \"placement-db-sync-42zwb\" (UID: \"5989cc6f-e4c7-4511-a148-293f683ede13\") " pod="openstack/placement-db-sync-42zwb" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.459102 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tjdf\" (UniqueName: \"kubernetes.io/projected/a4bb9e1a-0694-4e00-8ba6-f2765f82b983-kube-api-access-5tjdf\") pod \"barbican-db-sync-tpdmj\" (UID: \"a4bb9e1a-0694-4e00-8ba6-f2765f82b983\") " pod="openstack/barbican-db-sync-tpdmj" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.459122 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-dns-svc\") pod \"dnsmasq-dns-74b5448f7f-zdrjg\" (UID: \"ad591d5a-4120-4285-9cca-417b19176795\") " pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.459144 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5989cc6f-e4c7-4511-a148-293f683ede13-scripts\") pod \"placement-db-sync-42zwb\" (UID: \"5989cc6f-e4c7-4511-a148-293f683ede13\") " pod="openstack/placement-db-sync-42zwb" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.459164 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8rdc\" (UniqueName: \"kubernetes.io/projected/5989cc6f-e4c7-4511-a148-293f683ede13-kube-api-access-d8rdc\") pod \"placement-db-sync-42zwb\" (UID: \"5989cc6f-e4c7-4511-a148-293f683ede13\") " pod="openstack/placement-db-sync-42zwb" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.459189 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5989cc6f-e4c7-4511-a148-293f683ede13-logs\") pod \"placement-db-sync-42zwb\" (UID: \"5989cc6f-e4c7-4511-a148-293f683ede13\") " pod="openstack/placement-db-sync-42zwb" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.463102 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4bb9e1a-0694-4e00-8ba6-f2765f82b983-combined-ca-bundle\") pod \"barbican-db-sync-tpdmj\" (UID: \"a4bb9e1a-0694-4e00-8ba6-f2765f82b983\") " pod="openstack/barbican-db-sync-tpdmj" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.464887 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5989cc6f-e4c7-4511-a148-293f683ede13-config-data\") pod \"placement-db-sync-42zwb\" (UID: \"5989cc6f-e4c7-4511-a148-293f683ede13\") " pod="openstack/placement-db-sync-42zwb" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.464973 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5989cc6f-e4c7-4511-a148-293f683ede13-logs\") pod \"placement-db-sync-42zwb\" (UID: \"5989cc6f-e4c7-4511-a148-293f683ede13\") " pod="openstack/placement-db-sync-42zwb" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.464987 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5989cc6f-e4c7-4511-a148-293f683ede13-combined-ca-bundle\") pod \"placement-db-sync-42zwb\" (UID: \"5989cc6f-e4c7-4511-a148-293f683ede13\") " pod="openstack/placement-db-sync-42zwb" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.465222 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a4bb9e1a-0694-4e00-8ba6-f2765f82b983-db-sync-config-data\") pod \"barbican-db-sync-tpdmj\" (UID: \"a4bb9e1a-0694-4e00-8ba6-f2765f82b983\") " pod="openstack/barbican-db-sync-tpdmj" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.466138 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5989cc6f-e4c7-4511-a148-293f683ede13-scripts\") pod \"placement-db-sync-42zwb\" (UID: \"5989cc6f-e4c7-4511-a148-293f683ede13\") " pod="openstack/placement-db-sync-42zwb" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.505935 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8rdc\" (UniqueName: \"kubernetes.io/projected/5989cc6f-e4c7-4511-a148-293f683ede13-kube-api-access-d8rdc\") pod \"placement-db-sync-42zwb\" (UID: \"5989cc6f-e4c7-4511-a148-293f683ede13\") " pod="openstack/placement-db-sync-42zwb" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.528309 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tjdf\" (UniqueName: \"kubernetes.io/projected/a4bb9e1a-0694-4e00-8ba6-f2765f82b983-kube-api-access-5tjdf\") pod \"barbican-db-sync-tpdmj\" (UID: \"a4bb9e1a-0694-4e00-8ba6-f2765f82b983\") " pod="openstack/barbican-db-sync-tpdmj" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.532963 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-42zwb" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.561628 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-dns-svc\") pod \"dnsmasq-dns-74b5448f7f-zdrjg\" (UID: \"ad591d5a-4120-4285-9cca-417b19176795\") " pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.561805 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-ovsdbserver-sb\") pod \"dnsmasq-dns-74b5448f7f-zdrjg\" (UID: \"ad591d5a-4120-4285-9cca-417b19176795\") " pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.561915 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4qnj\" (UniqueName: \"kubernetes.io/projected/ad591d5a-4120-4285-9cca-417b19176795-kube-api-access-j4qnj\") pod \"dnsmasq-dns-74b5448f7f-zdrjg\" (UID: \"ad591d5a-4120-4285-9cca-417b19176795\") " pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.561978 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-config\") pod \"dnsmasq-dns-74b5448f7f-zdrjg\" (UID: \"ad591d5a-4120-4285-9cca-417b19176795\") " pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.562070 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-ovsdbserver-nb\") pod \"dnsmasq-dns-74b5448f7f-zdrjg\" (UID: \"ad591d5a-4120-4285-9cca-417b19176795\") " pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.563106 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-ovsdbserver-nb\") pod \"dnsmasq-dns-74b5448f7f-zdrjg\" (UID: \"ad591d5a-4120-4285-9cca-417b19176795\") " pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.563488 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-config\") pod \"dnsmasq-dns-74b5448f7f-zdrjg\" (UID: \"ad591d5a-4120-4285-9cca-417b19176795\") " pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.564016 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-ovsdbserver-sb\") pod \"dnsmasq-dns-74b5448f7f-zdrjg\" (UID: \"ad591d5a-4120-4285-9cca-417b19176795\") " pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.564578 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-dns-svc\") pod \"dnsmasq-dns-74b5448f7f-zdrjg\" (UID: \"ad591d5a-4120-4285-9cca-417b19176795\") " pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.621088 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4qnj\" (UniqueName: \"kubernetes.io/projected/ad591d5a-4120-4285-9cca-417b19176795-kube-api-access-j4qnj\") pod \"dnsmasq-dns-74b5448f7f-zdrjg\" (UID: \"ad591d5a-4120-4285-9cca-417b19176795\") " pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.692346 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.796932 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-tpdmj" Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.865060 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d6c8d744f-9mtrg"] Mar 18 20:25:46 crc kubenswrapper[4950]: I0318 20:25:46.932645 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2bbb2"] Mar 18 20:25:47 crc kubenswrapper[4950]: W0318 20:25:47.021726 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb2edc270_582d_4efe_84bc_f67ba8bbe372.slice/crio-7d7cac9a2f4843cc91e1d7d0fbfd9ab177d54e3003f26bb31fb99ff98cf93bb4 WatchSource:0}: Error finding container 7d7cac9a2f4843cc91e1d7d0fbfd9ab177d54e3003f26bb31fb99ff98cf93bb4: Status 404 returned error can't find the container with id 7d7cac9a2f4843cc91e1d7d0fbfd9ab177d54e3003f26bb31fb99ff98cf93bb4 Mar 18 20:25:47 crc kubenswrapper[4950]: I0318 20:25:47.152518 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:25:47 crc kubenswrapper[4950]: W0318 20:25:47.169633 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7563774c_0a5c_4af9_802a_662f4e721f3c.slice/crio-7cbf38b3da364f13e1adf82ec0026977bf82c9432979f7f2f5fb3080546ffc16 WatchSource:0}: Error finding container 7cbf38b3da364f13e1adf82ec0026977bf82c9432979f7f2f5fb3080546ffc16: Status 404 returned error can't find the container with id 7cbf38b3da364f13e1adf82ec0026977bf82c9432979f7f2f5fb3080546ffc16 Mar 18 20:25:47 crc kubenswrapper[4950]: W0318 20:25:47.348304 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82a7c2a4_9f57_460e_8bac_419f14222502.slice/crio-97a26f08854fc0dfaabc62754c4b6d2acb95a35796f8ece839e9b7c0bd98db5a WatchSource:0}: Error finding container 97a26f08854fc0dfaabc62754c4b6d2acb95a35796f8ece839e9b7c0bd98db5a: Status 404 returned error can't find the container with id 97a26f08854fc0dfaabc62754c4b6d2acb95a35796f8ece839e9b7c0bd98db5a Mar 18 20:25:47 crc kubenswrapper[4950]: I0318 20:25:47.356137 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-brgnq"] Mar 18 20:25:47 crc kubenswrapper[4950]: I0318 20:25:47.449230 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d6c8d744f-9mtrg" event={"ID":"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb","Type":"ContainerStarted","Data":"c8ba7afa5ee101d497d8f03d8757366ee8228314f04e1a9ab8129e6d08f3c318"} Mar 18 20:25:47 crc kubenswrapper[4950]: I0318 20:25:47.449267 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d6c8d744f-9mtrg" event={"ID":"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb","Type":"ContainerStarted","Data":"c8648811a452cad83982b7d74934bc25b5d671a48e8cbee88ba72b4c37df0fc1"} Mar 18 20:25:47 crc kubenswrapper[4950]: I0318 20:25:47.449977 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7563774c-0a5c-4af9-802a-662f4e721f3c","Type":"ContainerStarted","Data":"7cbf38b3da364f13e1adf82ec0026977bf82c9432979f7f2f5fb3080546ffc16"} Mar 18 20:25:47 crc kubenswrapper[4950]: I0318 20:25:47.451065 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2bbb2" event={"ID":"b2edc270-582d-4efe-84bc-f67ba8bbe372","Type":"ContainerStarted","Data":"fafb42832d2dcda2f28bda37a748b550b176d5f3fdcdc2a73e44ebdaf4cb9e3c"} Mar 18 20:25:47 crc kubenswrapper[4950]: I0318 20:25:47.451083 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2bbb2" event={"ID":"b2edc270-582d-4efe-84bc-f67ba8bbe372","Type":"ContainerStarted","Data":"7d7cac9a2f4843cc91e1d7d0fbfd9ab177d54e3003f26bb31fb99ff98cf93bb4"} Mar 18 20:25:47 crc kubenswrapper[4950]: I0318 20:25:47.452079 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-brgnq" event={"ID":"82a7c2a4-9f57-460e-8bac-419f14222502","Type":"ContainerStarted","Data":"97a26f08854fc0dfaabc62754c4b6d2acb95a35796f8ece839e9b7c0bd98db5a"} Mar 18 20:25:47 crc kubenswrapper[4950]: I0318 20:25:47.541630 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-j5m65"] Mar 18 20:25:47 crc kubenswrapper[4950]: I0318 20:25:47.570227 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-42zwb"] Mar 18 20:25:47 crc kubenswrapper[4950]: W0318 20:25:47.570658 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddbf245a4_e400_4fc2_9d2e_3788c6baf25d.slice/crio-bea6819b326906cf836243b13284fed398d47e0480a5718e359e9cd69ff16f44 WatchSource:0}: Error finding container bea6819b326906cf836243b13284fed398d47e0480a5718e359e9cd69ff16f44: Status 404 returned error can't find the container with id bea6819b326906cf836243b13284fed398d47e0480a5718e359e9cd69ff16f44 Mar 18 20:25:47 crc kubenswrapper[4950]: I0318 20:25:47.599949 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74b5448f7f-zdrjg"] Mar 18 20:25:47 crc kubenswrapper[4950]: I0318 20:25:47.808208 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-tpdmj"] Mar 18 20:25:48 crc kubenswrapper[4950]: I0318 20:25:48.370598 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:25:48 crc kubenswrapper[4950]: I0318 20:25:48.470320 4950 generic.go:334] "Generic (PLEG): container finished" podID="ad591d5a-4120-4285-9cca-417b19176795" containerID="1123658f52ad30114e365676ca550db7d2c0a2ef49ba5180e2fed0dc0370cc00" exitCode=0 Mar 18 20:25:48 crc kubenswrapper[4950]: I0318 20:25:48.470395 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" event={"ID":"ad591d5a-4120-4285-9cca-417b19176795","Type":"ContainerDied","Data":"1123658f52ad30114e365676ca550db7d2c0a2ef49ba5180e2fed0dc0370cc00"} Mar 18 20:25:48 crc kubenswrapper[4950]: I0318 20:25:48.470438 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" event={"ID":"ad591d5a-4120-4285-9cca-417b19176795","Type":"ContainerStarted","Data":"8c293bb8351af4475bcd2918ff2d3a397db802e08287df076b7ab111bbf95677"} Mar 18 20:25:48 crc kubenswrapper[4950]: I0318 20:25:48.476852 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-42zwb" event={"ID":"5989cc6f-e4c7-4511-a148-293f683ede13","Type":"ContainerStarted","Data":"957572ed504871d866513e222027abb3fc3002243e48c31bae435a2d4ebcf2d3"} Mar 18 20:25:48 crc kubenswrapper[4950]: I0318 20:25:48.480201 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-j5m65" event={"ID":"dbf245a4-e400-4fc2-9d2e-3788c6baf25d","Type":"ContainerStarted","Data":"161b8640c2219b243d32be7dcefce1bd3ee90d650d799042fa629a4ca59da713"} Mar 18 20:25:48 crc kubenswrapper[4950]: I0318 20:25:48.480232 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-j5m65" event={"ID":"dbf245a4-e400-4fc2-9d2e-3788c6baf25d","Type":"ContainerStarted","Data":"bea6819b326906cf836243b13284fed398d47e0480a5718e359e9cd69ff16f44"} Mar 18 20:25:48 crc kubenswrapper[4950]: I0318 20:25:48.513208 4950 generic.go:334] "Generic (PLEG): container finished" podID="8853452a-b71e-43d1-8d3a-f18ba1b9c3cb" containerID="c8ba7afa5ee101d497d8f03d8757366ee8228314f04e1a9ab8129e6d08f3c318" exitCode=0 Mar 18 20:25:48 crc kubenswrapper[4950]: I0318 20:25:48.513281 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d6c8d744f-9mtrg" event={"ID":"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb","Type":"ContainerDied","Data":"c8ba7afa5ee101d497d8f03d8757366ee8228314f04e1a9ab8129e6d08f3c318"} Mar 18 20:25:48 crc kubenswrapper[4950]: I0318 20:25:48.532238 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-tpdmj" event={"ID":"a4bb9e1a-0694-4e00-8ba6-f2765f82b983","Type":"ContainerStarted","Data":"5d841f1711b2d20eb9beeb857c1703a9d8c135fead221b2e15a23a8700b18a1a"} Mar 18 20:25:48 crc kubenswrapper[4950]: I0318 20:25:48.575270 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-j5m65" podStartSLOduration=3.57524736 podStartE2EDuration="3.57524736s" podCreationTimestamp="2026-03-18 20:25:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:25:48.526889305 +0000 UTC m=+1161.767731173" watchObservedRunningTime="2026-03-18 20:25:48.57524736 +0000 UTC m=+1161.816089228" Mar 18 20:25:48 crc kubenswrapper[4950]: I0318 20:25:48.617087 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-2bbb2" podStartSLOduration=3.617066967 podStartE2EDuration="3.617066967s" podCreationTimestamp="2026-03-18 20:25:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:25:48.582945069 +0000 UTC m=+1161.823786947" watchObservedRunningTime="2026-03-18 20:25:48.617066967 +0000 UTC m=+1161.857908835" Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.086902 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d6c8d744f-9mtrg" Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.233310 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-config\") pod \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\" (UID: \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\") " Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.233355 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-ovsdbserver-nb\") pod \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\" (UID: \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\") " Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.233473 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-ovsdbserver-sb\") pod \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\" (UID: \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\") " Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.233589 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5529\" (UniqueName: \"kubernetes.io/projected/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-kube-api-access-k5529\") pod \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\" (UID: \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\") " Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.234176 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-dns-svc\") pod \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\" (UID: \"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb\") " Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.242247 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-kube-api-access-k5529" (OuterVolumeSpecName: "kube-api-access-k5529") pod "8853452a-b71e-43d1-8d3a-f18ba1b9c3cb" (UID: "8853452a-b71e-43d1-8d3a-f18ba1b9c3cb"). InnerVolumeSpecName "kube-api-access-k5529". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.261834 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8853452a-b71e-43d1-8d3a-f18ba1b9c3cb" (UID: "8853452a-b71e-43d1-8d3a-f18ba1b9c3cb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.268442 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8853452a-b71e-43d1-8d3a-f18ba1b9c3cb" (UID: "8853452a-b71e-43d1-8d3a-f18ba1b9c3cb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.269836 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8853452a-b71e-43d1-8d3a-f18ba1b9c3cb" (UID: "8853452a-b71e-43d1-8d3a-f18ba1b9c3cb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.273360 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-config" (OuterVolumeSpecName: "config") pod "8853452a-b71e-43d1-8d3a-f18ba1b9c3cb" (UID: "8853452a-b71e-43d1-8d3a-f18ba1b9c3cb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.336475 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5529\" (UniqueName: \"kubernetes.io/projected/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-kube-api-access-k5529\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.336520 4950 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.336533 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.336542 4950 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.336552 4950 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.541137 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" event={"ID":"ad591d5a-4120-4285-9cca-417b19176795","Type":"ContainerStarted","Data":"9868c2c62d92308f1e605a851fa729dafcb7f9155140705a4f262c2522d2d1dc"} Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.541605 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.545496 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d6c8d744f-9mtrg" event={"ID":"8853452a-b71e-43d1-8d3a-f18ba1b9c3cb","Type":"ContainerDied","Data":"c8648811a452cad83982b7d74934bc25b5d671a48e8cbee88ba72b4c37df0fc1"} Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.545940 4950 scope.go:117] "RemoveContainer" containerID="c8ba7afa5ee101d497d8f03d8757366ee8228314f04e1a9ab8129e6d08f3c318" Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.545704 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d6c8d744f-9mtrg" Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.563449 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" podStartSLOduration=3.563433114 podStartE2EDuration="3.563433114s" podCreationTimestamp="2026-03-18 20:25:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:25:49.562951112 +0000 UTC m=+1162.803792980" watchObservedRunningTime="2026-03-18 20:25:49.563433114 +0000 UTC m=+1162.804274982" Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.625469 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d6c8d744f-9mtrg"] Mar 18 20:25:49 crc kubenswrapper[4950]: I0318 20:25:49.629028 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d6c8d744f-9mtrg"] Mar 18 20:25:51 crc kubenswrapper[4950]: I0318 20:25:51.500628 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8853452a-b71e-43d1-8d3a-f18ba1b9c3cb" path="/var/lib/kubelet/pods/8853452a-b71e-43d1-8d3a-f18ba1b9c3cb/volumes" Mar 18 20:25:52 crc kubenswrapper[4950]: I0318 20:25:52.580616 4950 generic.go:334] "Generic (PLEG): container finished" podID="b2edc270-582d-4efe-84bc-f67ba8bbe372" containerID="fafb42832d2dcda2f28bda37a748b550b176d5f3fdcdc2a73e44ebdaf4cb9e3c" exitCode=0 Mar 18 20:25:52 crc kubenswrapper[4950]: I0318 20:25:52.580678 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2bbb2" event={"ID":"b2edc270-582d-4efe-84bc-f67ba8bbe372","Type":"ContainerDied","Data":"fafb42832d2dcda2f28bda37a748b550b176d5f3fdcdc2a73e44ebdaf4cb9e3c"} Mar 18 20:25:56 crc kubenswrapper[4950]: I0318 20:25:56.694534 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" Mar 18 20:25:56 crc kubenswrapper[4950]: I0318 20:25:56.766553 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-547b76fb45-s9htl"] Mar 18 20:25:56 crc kubenswrapper[4950]: I0318 20:25:56.766923 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-547b76fb45-s9htl" podUID="96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac" containerName="dnsmasq-dns" containerID="cri-o://b2c55f5f177417991f90810fa75e695c04b170f5a7d0bf92b93164c924bac67a" gracePeriod=10 Mar 18 20:25:57 crc kubenswrapper[4950]: I0318 20:25:57.615207 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-547b76fb45-s9htl" event={"ID":"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac","Type":"ContainerDied","Data":"b2c55f5f177417991f90810fa75e695c04b170f5a7d0bf92b93164c924bac67a"} Mar 18 20:25:57 crc kubenswrapper[4950]: I0318 20:25:57.615457 4950 generic.go:334] "Generic (PLEG): container finished" podID="96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac" containerID="b2c55f5f177417991f90810fa75e695c04b170f5a7d0bf92b93164c924bac67a" exitCode=0 Mar 18 20:25:59 crc kubenswrapper[4950]: I0318 20:25:59.426288 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2bbb2" Mar 18 20:25:59 crc kubenswrapper[4950]: I0318 20:25:59.519856 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-credential-keys\") pod \"b2edc270-582d-4efe-84bc-f67ba8bbe372\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " Mar 18 20:25:59 crc kubenswrapper[4950]: I0318 20:25:59.519915 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-combined-ca-bundle\") pod \"b2edc270-582d-4efe-84bc-f67ba8bbe372\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " Mar 18 20:25:59 crc kubenswrapper[4950]: I0318 20:25:59.519960 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-fernet-keys\") pod \"b2edc270-582d-4efe-84bc-f67ba8bbe372\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " Mar 18 20:25:59 crc kubenswrapper[4950]: I0318 20:25:59.519999 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-scripts\") pod \"b2edc270-582d-4efe-84bc-f67ba8bbe372\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " Mar 18 20:25:59 crc kubenswrapper[4950]: I0318 20:25:59.520051 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zskfs\" (UniqueName: \"kubernetes.io/projected/b2edc270-582d-4efe-84bc-f67ba8bbe372-kube-api-access-zskfs\") pod \"b2edc270-582d-4efe-84bc-f67ba8bbe372\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " Mar 18 20:25:59 crc kubenswrapper[4950]: I0318 20:25:59.520140 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-config-data\") pod \"b2edc270-582d-4efe-84bc-f67ba8bbe372\" (UID: \"b2edc270-582d-4efe-84bc-f67ba8bbe372\") " Mar 18 20:25:59 crc kubenswrapper[4950]: I0318 20:25:59.525935 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b2edc270-582d-4efe-84bc-f67ba8bbe372" (UID: "b2edc270-582d-4efe-84bc-f67ba8bbe372"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:25:59 crc kubenswrapper[4950]: I0318 20:25:59.525960 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "b2edc270-582d-4efe-84bc-f67ba8bbe372" (UID: "b2edc270-582d-4efe-84bc-f67ba8bbe372"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:25:59 crc kubenswrapper[4950]: I0318 20:25:59.526663 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-scripts" (OuterVolumeSpecName: "scripts") pod "b2edc270-582d-4efe-84bc-f67ba8bbe372" (UID: "b2edc270-582d-4efe-84bc-f67ba8bbe372"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:25:59 crc kubenswrapper[4950]: I0318 20:25:59.548913 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2edc270-582d-4efe-84bc-f67ba8bbe372-kube-api-access-zskfs" (OuterVolumeSpecName: "kube-api-access-zskfs") pod "b2edc270-582d-4efe-84bc-f67ba8bbe372" (UID: "b2edc270-582d-4efe-84bc-f67ba8bbe372"). InnerVolumeSpecName "kube-api-access-zskfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:25:59 crc kubenswrapper[4950]: I0318 20:25:59.556363 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b2edc270-582d-4efe-84bc-f67ba8bbe372" (UID: "b2edc270-582d-4efe-84bc-f67ba8bbe372"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:25:59 crc kubenswrapper[4950]: I0318 20:25:59.559972 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-config-data" (OuterVolumeSpecName: "config-data") pod "b2edc270-582d-4efe-84bc-f67ba8bbe372" (UID: "b2edc270-582d-4efe-84bc-f67ba8bbe372"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:25:59 crc kubenswrapper[4950]: I0318 20:25:59.622451 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:59 crc kubenswrapper[4950]: I0318 20:25:59.623094 4950 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:59 crc kubenswrapper[4950]: I0318 20:25:59.623320 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:59 crc kubenswrapper[4950]: I0318 20:25:59.623331 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zskfs\" (UniqueName: \"kubernetes.io/projected/b2edc270-582d-4efe-84bc-f67ba8bbe372-kube-api-access-zskfs\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:59 crc kubenswrapper[4950]: I0318 20:25:59.623342 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:59 crc kubenswrapper[4950]: I0318 20:25:59.623350 4950 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b2edc270-582d-4efe-84bc-f67ba8bbe372-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 18 20:25:59 crc kubenswrapper[4950]: I0318 20:25:59.632222 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2bbb2" event={"ID":"b2edc270-582d-4efe-84bc-f67ba8bbe372","Type":"ContainerDied","Data":"7d7cac9a2f4843cc91e1d7d0fbfd9ab177d54e3003f26bb31fb99ff98cf93bb4"} Mar 18 20:25:59 crc kubenswrapper[4950]: I0318 20:25:59.632251 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d7cac9a2f4843cc91e1d7d0fbfd9ab177d54e3003f26bb31fb99ff98cf93bb4" Mar 18 20:25:59 crc kubenswrapper[4950]: I0318 20:25:59.632291 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2bbb2" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.147085 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564426-4lcpm"] Mar 18 20:26:00 crc kubenswrapper[4950]: E0318 20:26:00.147517 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8853452a-b71e-43d1-8d3a-f18ba1b9c3cb" containerName="init" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.147532 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="8853452a-b71e-43d1-8d3a-f18ba1b9c3cb" containerName="init" Mar 18 20:26:00 crc kubenswrapper[4950]: E0318 20:26:00.147564 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2edc270-582d-4efe-84bc-f67ba8bbe372" containerName="keystone-bootstrap" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.147572 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2edc270-582d-4efe-84bc-f67ba8bbe372" containerName="keystone-bootstrap" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.147768 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2edc270-582d-4efe-84bc-f67ba8bbe372" containerName="keystone-bootstrap" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.147779 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="8853452a-b71e-43d1-8d3a-f18ba1b9c3cb" containerName="init" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.148479 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564426-4lcpm" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.153361 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.153708 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.156491 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.157704 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564426-4lcpm"] Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.232262 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtfrf\" (UniqueName: \"kubernetes.io/projected/619d4317-c521-4032-a59f-067f9914bb44-kube-api-access-dtfrf\") pod \"auto-csr-approver-29564426-4lcpm\" (UID: \"619d4317-c521-4032-a59f-067f9914bb44\") " pod="openshift-infra/auto-csr-approver-29564426-4lcpm" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.335854 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtfrf\" (UniqueName: \"kubernetes.io/projected/619d4317-c521-4032-a59f-067f9914bb44-kube-api-access-dtfrf\") pod \"auto-csr-approver-29564426-4lcpm\" (UID: \"619d4317-c521-4032-a59f-067f9914bb44\") " pod="openshift-infra/auto-csr-approver-29564426-4lcpm" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.352862 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtfrf\" (UniqueName: \"kubernetes.io/projected/619d4317-c521-4032-a59f-067f9914bb44-kube-api-access-dtfrf\") pod \"auto-csr-approver-29564426-4lcpm\" (UID: \"619d4317-c521-4032-a59f-067f9914bb44\") " pod="openshift-infra/auto-csr-approver-29564426-4lcpm" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.480425 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564426-4lcpm" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.523669 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-2bbb2"] Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.536181 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-2bbb2"] Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.616623 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-dfbp4"] Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.617763 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dfbp4" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.622835 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.623226 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.623251 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.623357 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-bsdbw" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.623386 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.628213 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dfbp4"] Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.741954 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-config-data\") pod \"keystone-bootstrap-dfbp4\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " pod="openstack/keystone-bootstrap-dfbp4" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.742021 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-combined-ca-bundle\") pod \"keystone-bootstrap-dfbp4\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " pod="openstack/keystone-bootstrap-dfbp4" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.742059 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-scripts\") pod \"keystone-bootstrap-dfbp4\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " pod="openstack/keystone-bootstrap-dfbp4" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.742077 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-credential-keys\") pod \"keystone-bootstrap-dfbp4\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " pod="openstack/keystone-bootstrap-dfbp4" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.742161 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn2ws\" (UniqueName: \"kubernetes.io/projected/d8589447-6f54-4a2b-80b5-29fdbccb4971-kube-api-access-xn2ws\") pod \"keystone-bootstrap-dfbp4\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " pod="openstack/keystone-bootstrap-dfbp4" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.742210 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-fernet-keys\") pod \"keystone-bootstrap-dfbp4\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " pod="openstack/keystone-bootstrap-dfbp4" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.843451 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-combined-ca-bundle\") pod \"keystone-bootstrap-dfbp4\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " pod="openstack/keystone-bootstrap-dfbp4" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.843514 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-scripts\") pod \"keystone-bootstrap-dfbp4\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " pod="openstack/keystone-bootstrap-dfbp4" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.843543 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-credential-keys\") pod \"keystone-bootstrap-dfbp4\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " pod="openstack/keystone-bootstrap-dfbp4" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.843604 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn2ws\" (UniqueName: \"kubernetes.io/projected/d8589447-6f54-4a2b-80b5-29fdbccb4971-kube-api-access-xn2ws\") pod \"keystone-bootstrap-dfbp4\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " pod="openstack/keystone-bootstrap-dfbp4" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.843655 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-fernet-keys\") pod \"keystone-bootstrap-dfbp4\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " pod="openstack/keystone-bootstrap-dfbp4" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.843701 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-config-data\") pod \"keystone-bootstrap-dfbp4\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " pod="openstack/keystone-bootstrap-dfbp4" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.848904 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-scripts\") pod \"keystone-bootstrap-dfbp4\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " pod="openstack/keystone-bootstrap-dfbp4" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.849119 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-combined-ca-bundle\") pod \"keystone-bootstrap-dfbp4\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " pod="openstack/keystone-bootstrap-dfbp4" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.849458 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-fernet-keys\") pod \"keystone-bootstrap-dfbp4\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " pod="openstack/keystone-bootstrap-dfbp4" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.850472 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-config-data\") pod \"keystone-bootstrap-dfbp4\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " pod="openstack/keystone-bootstrap-dfbp4" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.858195 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-credential-keys\") pod \"keystone-bootstrap-dfbp4\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " pod="openstack/keystone-bootstrap-dfbp4" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.859477 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn2ws\" (UniqueName: \"kubernetes.io/projected/d8589447-6f54-4a2b-80b5-29fdbccb4971-kube-api-access-xn2ws\") pod \"keystone-bootstrap-dfbp4\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " pod="openstack/keystone-bootstrap-dfbp4" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.951493 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dfbp4" Mar 18 20:26:00 crc kubenswrapper[4950]: I0318 20:26:00.993749 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-547b76fb45-s9htl" podUID="96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: connect: connection refused" Mar 18 20:26:01 crc kubenswrapper[4950]: I0318 20:26:01.493195 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2edc270-582d-4efe-84bc-f67ba8bbe372" path="/var/lib/kubelet/pods/b2edc270-582d-4efe-84bc-f67ba8bbe372/volumes" Mar 18 20:26:05 crc kubenswrapper[4950]: I0318 20:26:05.993545 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-547b76fb45-s9htl" podUID="96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: connect: connection refused" Mar 18 20:26:09 crc kubenswrapper[4950]: E0318 20:26:09.142134 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:574a17f0877c175128a764f2b37fc02456649c8514689125718ce6ca974bfb6b" Mar 18 20:26:09 crc kubenswrapper[4950]: E0318 20:26:09.142744 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:574a17f0877c175128a764f2b37fc02456649c8514689125718ce6ca974bfb6b,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9wd8n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-brgnq_openstack(82a7c2a4-9f57-460e-8bac-419f14222502): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 18 20:26:09 crc kubenswrapper[4950]: E0318 20:26:09.143901 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-brgnq" podUID="82a7c2a4-9f57-460e-8bac-419f14222502" Mar 18 20:26:09 crc kubenswrapper[4950]: E0318 20:26:09.651906 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:1240a45aec9c3e1599be762c5565556560849b49fd39c7283b8e5519dcaa501a" Mar 18 20:26:09 crc kubenswrapper[4950]: E0318 20:26:09.652478 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:1240a45aec9c3e1599be762c5565556560849b49fd39c7283b8e5519dcaa501a,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5tjdf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-tpdmj_openstack(a4bb9e1a-0694-4e00-8ba6-f2765f82b983): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 18 20:26:09 crc kubenswrapper[4950]: E0318 20:26:09.653934 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-tpdmj" podUID="a4bb9e1a-0694-4e00-8ba6-f2765f82b983" Mar 18 20:26:09 crc kubenswrapper[4950]: E0318 20:26:09.727387 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:574a17f0877c175128a764f2b37fc02456649c8514689125718ce6ca974bfb6b\\\"\"" pod="openstack/cinder-db-sync-brgnq" podUID="82a7c2a4-9f57-460e-8bac-419f14222502" Mar 18 20:26:09 crc kubenswrapper[4950]: E0318 20:26:09.727996 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:1240a45aec9c3e1599be762c5565556560849b49fd39c7283b8e5519dcaa501a\\\"\"" pod="openstack/barbican-db-sync-tpdmj" podUID="a4bb9e1a-0694-4e00-8ba6-f2765f82b983" Mar 18 20:26:09 crc kubenswrapper[4950]: I0318 20:26:09.993617 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-547b76fb45-s9htl" Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.140067 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-dns-svc\") pod \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\" (UID: \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\") " Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.140475 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-config\") pod \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\" (UID: \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\") " Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.140522 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmxv8\" (UniqueName: \"kubernetes.io/projected/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-kube-api-access-kmxv8\") pod \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\" (UID: \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\") " Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.140599 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-ovsdbserver-sb\") pod \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\" (UID: \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\") " Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.140739 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-ovsdbserver-nb\") pod \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\" (UID: \"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac\") " Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.147586 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-kube-api-access-kmxv8" (OuterVolumeSpecName: "kube-api-access-kmxv8") pod "96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac" (UID: "96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac"). InnerVolumeSpecName "kube-api-access-kmxv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.182467 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac" (UID: "96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.187501 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-config" (OuterVolumeSpecName: "config") pod "96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac" (UID: "96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.188654 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac" (UID: "96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.201049 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac" (UID: "96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.217635 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dfbp4"] Mar 18 20:26:10 crc kubenswrapper[4950]: W0318 20:26:10.227976 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod619d4317_c521_4032_a59f_067f9914bb44.slice/crio-73b58b8c55f6bf2f6de089b4b8dcd609ffc0086d47b64be2013435efd05aa797 WatchSource:0}: Error finding container 73b58b8c55f6bf2f6de089b4b8dcd609ffc0086d47b64be2013435efd05aa797: Status 404 returned error can't find the container with id 73b58b8c55f6bf2f6de089b4b8dcd609ffc0086d47b64be2013435efd05aa797 Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.241910 4950 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.241933 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.241943 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmxv8\" (UniqueName: \"kubernetes.io/projected/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-kube-api-access-kmxv8\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.241952 4950 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.241961 4950 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.243527 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564426-4lcpm"] Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.735134 4950 generic.go:334] "Generic (PLEG): container finished" podID="dbf245a4-e400-4fc2-9d2e-3788c6baf25d" containerID="161b8640c2219b243d32be7dcefce1bd3ee90d650d799042fa629a4ca59da713" exitCode=0 Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.735218 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-j5m65" event={"ID":"dbf245a4-e400-4fc2-9d2e-3788c6baf25d","Type":"ContainerDied","Data":"161b8640c2219b243d32be7dcefce1bd3ee90d650d799042fa629a4ca59da713"} Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.737818 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dfbp4" event={"ID":"d8589447-6f54-4a2b-80b5-29fdbccb4971","Type":"ContainerStarted","Data":"ef83c6302a1ab85f29e24e83bef380a024d32fad1c7855a688e1adbe340d26bf"} Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.737862 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dfbp4" event={"ID":"d8589447-6f54-4a2b-80b5-29fdbccb4971","Type":"ContainerStarted","Data":"842b6dc1d490c53b741456c0e5c030a31484193029da75c0986e28072fd8101b"} Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.739706 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564426-4lcpm" event={"ID":"619d4317-c521-4032-a59f-067f9914bb44","Type":"ContainerStarted","Data":"73b58b8c55f6bf2f6de089b4b8dcd609ffc0086d47b64be2013435efd05aa797"} Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.741403 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-547b76fb45-s9htl" event={"ID":"96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac","Type":"ContainerDied","Data":"62916e59361722b7ce64cb4bd1b19fdc7ca2c4299ab9d9e48292b2d5f6c8d70b"} Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.741507 4950 scope.go:117] "RemoveContainer" containerID="b2c55f5f177417991f90810fa75e695c04b170f5a7d0bf92b93164c924bac67a" Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.741637 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-547b76fb45-s9htl" Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.745002 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7563774c-0a5c-4af9-802a-662f4e721f3c","Type":"ContainerStarted","Data":"de1b4d05b6433c8bacfa4f5808c26a249120f01056e842ff8ef61a2c85cdcfad"} Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.755695 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-42zwb" event={"ID":"5989cc6f-e4c7-4511-a148-293f683ede13","Type":"ContainerStarted","Data":"076697f27bc810727a97bc72cc9216b889409ae3d998e9af07d0b9f87a899622"} Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.778385 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-dfbp4" podStartSLOduration=10.77836325 podStartE2EDuration="10.77836325s" podCreationTimestamp="2026-03-18 20:26:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:26:10.772858578 +0000 UTC m=+1184.013700446" watchObservedRunningTime="2026-03-18 20:26:10.77836325 +0000 UTC m=+1184.019205118" Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.810136 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-547b76fb45-s9htl"] Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.820284 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-547b76fb45-s9htl"] Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.825386 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-42zwb" podStartSLOduration=2.732156111 podStartE2EDuration="24.82536704s" podCreationTimestamp="2026-03-18 20:25:46 +0000 UTC" firstStartedPulling="2026-03-18 20:25:47.588554886 +0000 UTC m=+1160.829396754" lastFinishedPulling="2026-03-18 20:26:09.681765775 +0000 UTC m=+1182.922607683" observedRunningTime="2026-03-18 20:26:10.809609304 +0000 UTC m=+1184.050451182" watchObservedRunningTime="2026-03-18 20:26:10.82536704 +0000 UTC m=+1184.066208908" Mar 18 20:26:10 crc kubenswrapper[4950]: I0318 20:26:10.982671 4950 scope.go:117] "RemoveContainer" containerID="ec73589967679f88800110dd92359f8262173fe346033fc696a416b664281a91" Mar 18 20:26:11 crc kubenswrapper[4950]: I0318 20:26:11.503025 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac" path="/var/lib/kubelet/pods/96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac/volumes" Mar 18 20:26:11 crc kubenswrapper[4950]: I0318 20:26:11.763572 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7563774c-0a5c-4af9-802a-662f4e721f3c","Type":"ContainerStarted","Data":"a94ea562bd7c3e520d36e8bc6b0f62fe4b690bcbac96af7089b92400bea93599"} Mar 18 20:26:11 crc kubenswrapper[4950]: I0318 20:26:11.764962 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564426-4lcpm" event={"ID":"619d4317-c521-4032-a59f-067f9914bb44","Type":"ContainerStarted","Data":"1a47134f941a07b6ccc64ee6978cdd2910a4f674c4f2e2038ee93b17c562ff9c"} Mar 18 20:26:11 crc kubenswrapper[4950]: I0318 20:26:11.790166 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29564426-4lcpm" podStartSLOduration=10.823407089 podStartE2EDuration="11.790148621s" podCreationTimestamp="2026-03-18 20:26:00 +0000 UTC" firstStartedPulling="2026-03-18 20:26:10.240777128 +0000 UTC m=+1183.481618996" lastFinishedPulling="2026-03-18 20:26:11.20751866 +0000 UTC m=+1184.448360528" observedRunningTime="2026-03-18 20:26:11.785970904 +0000 UTC m=+1185.026812792" watchObservedRunningTime="2026-03-18 20:26:11.790148621 +0000 UTC m=+1185.030990489" Mar 18 20:26:12 crc kubenswrapper[4950]: I0318 20:26:12.179301 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-j5m65" Mar 18 20:26:12 crc kubenswrapper[4950]: I0318 20:26:12.273090 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbf245a4-e400-4fc2-9d2e-3788c6baf25d-combined-ca-bundle\") pod \"dbf245a4-e400-4fc2-9d2e-3788c6baf25d\" (UID: \"dbf245a4-e400-4fc2-9d2e-3788c6baf25d\") " Mar 18 20:26:12 crc kubenswrapper[4950]: I0318 20:26:12.273250 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8zwx\" (UniqueName: \"kubernetes.io/projected/dbf245a4-e400-4fc2-9d2e-3788c6baf25d-kube-api-access-g8zwx\") pod \"dbf245a4-e400-4fc2-9d2e-3788c6baf25d\" (UID: \"dbf245a4-e400-4fc2-9d2e-3788c6baf25d\") " Mar 18 20:26:12 crc kubenswrapper[4950]: I0318 20:26:12.273297 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/dbf245a4-e400-4fc2-9d2e-3788c6baf25d-config\") pod \"dbf245a4-e400-4fc2-9d2e-3788c6baf25d\" (UID: \"dbf245a4-e400-4fc2-9d2e-3788c6baf25d\") " Mar 18 20:26:12 crc kubenswrapper[4950]: I0318 20:26:12.284385 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbf245a4-e400-4fc2-9d2e-3788c6baf25d-kube-api-access-g8zwx" (OuterVolumeSpecName: "kube-api-access-g8zwx") pod "dbf245a4-e400-4fc2-9d2e-3788c6baf25d" (UID: "dbf245a4-e400-4fc2-9d2e-3788c6baf25d"). InnerVolumeSpecName "kube-api-access-g8zwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:26:12 crc kubenswrapper[4950]: I0318 20:26:12.304154 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbf245a4-e400-4fc2-9d2e-3788c6baf25d-config" (OuterVolumeSpecName: "config") pod "dbf245a4-e400-4fc2-9d2e-3788c6baf25d" (UID: "dbf245a4-e400-4fc2-9d2e-3788c6baf25d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:12 crc kubenswrapper[4950]: I0318 20:26:12.334003 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbf245a4-e400-4fc2-9d2e-3788c6baf25d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dbf245a4-e400-4fc2-9d2e-3788c6baf25d" (UID: "dbf245a4-e400-4fc2-9d2e-3788c6baf25d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:12 crc kubenswrapper[4950]: I0318 20:26:12.375484 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbf245a4-e400-4fc2-9d2e-3788c6baf25d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:12 crc kubenswrapper[4950]: I0318 20:26:12.375521 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8zwx\" (UniqueName: \"kubernetes.io/projected/dbf245a4-e400-4fc2-9d2e-3788c6baf25d-kube-api-access-g8zwx\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:12 crc kubenswrapper[4950]: I0318 20:26:12.375537 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/dbf245a4-e400-4fc2-9d2e-3788c6baf25d-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:12 crc kubenswrapper[4950]: I0318 20:26:12.795182 4950 generic.go:334] "Generic (PLEG): container finished" podID="619d4317-c521-4032-a59f-067f9914bb44" containerID="1a47134f941a07b6ccc64ee6978cdd2910a4f674c4f2e2038ee93b17c562ff9c" exitCode=0 Mar 18 20:26:12 crc kubenswrapper[4950]: I0318 20:26:12.795502 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564426-4lcpm" event={"ID":"619d4317-c521-4032-a59f-067f9914bb44","Type":"ContainerDied","Data":"1a47134f941a07b6ccc64ee6978cdd2910a4f674c4f2e2038ee93b17c562ff9c"} Mar 18 20:26:12 crc kubenswrapper[4950]: I0318 20:26:12.798696 4950 generic.go:334] "Generic (PLEG): container finished" podID="5989cc6f-e4c7-4511-a148-293f683ede13" containerID="076697f27bc810727a97bc72cc9216b889409ae3d998e9af07d0b9f87a899622" exitCode=0 Mar 18 20:26:12 crc kubenswrapper[4950]: I0318 20:26:12.798767 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-42zwb" event={"ID":"5989cc6f-e4c7-4511-a148-293f683ede13","Type":"ContainerDied","Data":"076697f27bc810727a97bc72cc9216b889409ae3d998e9af07d0b9f87a899622"} Mar 18 20:26:12 crc kubenswrapper[4950]: I0318 20:26:12.802298 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-j5m65" event={"ID":"dbf245a4-e400-4fc2-9d2e-3788c6baf25d","Type":"ContainerDied","Data":"bea6819b326906cf836243b13284fed398d47e0480a5718e359e9cd69ff16f44"} Mar 18 20:26:12 crc kubenswrapper[4950]: I0318 20:26:12.802323 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bea6819b326906cf836243b13284fed398d47e0480a5718e359e9cd69ff16f44" Mar 18 20:26:12 crc kubenswrapper[4950]: I0318 20:26:12.802342 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-j5m65" Mar 18 20:26:12 crc kubenswrapper[4950]: I0318 20:26:12.971230 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75bf4db767-zbs9v"] Mar 18 20:26:12 crc kubenswrapper[4950]: E0318 20:26:12.993806 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbf245a4-e400-4fc2-9d2e-3788c6baf25d" containerName="neutron-db-sync" Mar 18 20:26:12 crc kubenswrapper[4950]: I0318 20:26:12.993946 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbf245a4-e400-4fc2-9d2e-3788c6baf25d" containerName="neutron-db-sync" Mar 18 20:26:12 crc kubenswrapper[4950]: E0318 20:26:12.994117 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac" containerName="dnsmasq-dns" Mar 18 20:26:12 crc kubenswrapper[4950]: I0318 20:26:12.994128 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac" containerName="dnsmasq-dns" Mar 18 20:26:12 crc kubenswrapper[4950]: E0318 20:26:12.994162 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac" containerName="init" Mar 18 20:26:12 crc kubenswrapper[4950]: I0318 20:26:12.994170 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac" containerName="init" Mar 18 20:26:12 crc kubenswrapper[4950]: I0318 20:26:12.995839 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbf245a4-e400-4fc2-9d2e-3788c6baf25d" containerName="neutron-db-sync" Mar 18 20:26:12 crc kubenswrapper[4950]: I0318 20:26:12.995864 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="96f4cfa6-4ce7-4eb1-b8f9-57283cce64ac" containerName="dnsmasq-dns" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.002074 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.022257 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75bf4db767-zbs9v"] Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.097964 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nh2qg\" (UniqueName: \"kubernetes.io/projected/a95776b7-4a12-4de5-b60f-02d008344485-kube-api-access-nh2qg\") pod \"dnsmasq-dns-75bf4db767-zbs9v\" (UID: \"a95776b7-4a12-4de5-b60f-02d008344485\") " pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.098064 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-ovsdbserver-sb\") pod \"dnsmasq-dns-75bf4db767-zbs9v\" (UID: \"a95776b7-4a12-4de5-b60f-02d008344485\") " pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.098097 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-ovsdbserver-nb\") pod \"dnsmasq-dns-75bf4db767-zbs9v\" (UID: \"a95776b7-4a12-4de5-b60f-02d008344485\") " pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.098155 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-dns-svc\") pod \"dnsmasq-dns-75bf4db767-zbs9v\" (UID: \"a95776b7-4a12-4de5-b60f-02d008344485\") " pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.098179 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-config\") pod \"dnsmasq-dns-75bf4db767-zbs9v\" (UID: \"a95776b7-4a12-4de5-b60f-02d008344485\") " pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.107031 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-c888c6fcb-jnt6f"] Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.112010 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c888c6fcb-jnt6f" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.121448 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-dz6tt" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.121660 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.122211 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.123177 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c888c6fcb-jnt6f"] Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.123711 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.199367 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ftgk\" (UniqueName: \"kubernetes.io/projected/ed97e535-9755-40ef-a82f-176f0eb7e31b-kube-api-access-6ftgk\") pod \"neutron-c888c6fcb-jnt6f\" (UID: \"ed97e535-9755-40ef-a82f-176f0eb7e31b\") " pod="openstack/neutron-c888c6fcb-jnt6f" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.199447 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-combined-ca-bundle\") pod \"neutron-c888c6fcb-jnt6f\" (UID: \"ed97e535-9755-40ef-a82f-176f0eb7e31b\") " pod="openstack/neutron-c888c6fcb-jnt6f" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.199472 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-ovndb-tls-certs\") pod \"neutron-c888c6fcb-jnt6f\" (UID: \"ed97e535-9755-40ef-a82f-176f0eb7e31b\") " pod="openstack/neutron-c888c6fcb-jnt6f" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.199542 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-config\") pod \"neutron-c888c6fcb-jnt6f\" (UID: \"ed97e535-9755-40ef-a82f-176f0eb7e31b\") " pod="openstack/neutron-c888c6fcb-jnt6f" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.199605 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-ovsdbserver-sb\") pod \"dnsmasq-dns-75bf4db767-zbs9v\" (UID: \"a95776b7-4a12-4de5-b60f-02d008344485\") " pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.199668 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-ovsdbserver-nb\") pod \"dnsmasq-dns-75bf4db767-zbs9v\" (UID: \"a95776b7-4a12-4de5-b60f-02d008344485\") " pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.200986 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-ovsdbserver-sb\") pod \"dnsmasq-dns-75bf4db767-zbs9v\" (UID: \"a95776b7-4a12-4de5-b60f-02d008344485\") " pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.201058 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-httpd-config\") pod \"neutron-c888c6fcb-jnt6f\" (UID: \"ed97e535-9755-40ef-a82f-176f0eb7e31b\") " pod="openstack/neutron-c888c6fcb-jnt6f" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.201088 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-dns-svc\") pod \"dnsmasq-dns-75bf4db767-zbs9v\" (UID: \"a95776b7-4a12-4de5-b60f-02d008344485\") " pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.201100 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-ovsdbserver-nb\") pod \"dnsmasq-dns-75bf4db767-zbs9v\" (UID: \"a95776b7-4a12-4de5-b60f-02d008344485\") " pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.201112 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-config\") pod \"dnsmasq-dns-75bf4db767-zbs9v\" (UID: \"a95776b7-4a12-4de5-b60f-02d008344485\") " pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.201173 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nh2qg\" (UniqueName: \"kubernetes.io/projected/a95776b7-4a12-4de5-b60f-02d008344485-kube-api-access-nh2qg\") pod \"dnsmasq-dns-75bf4db767-zbs9v\" (UID: \"a95776b7-4a12-4de5-b60f-02d008344485\") " pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.201870 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-dns-svc\") pod \"dnsmasq-dns-75bf4db767-zbs9v\" (UID: \"a95776b7-4a12-4de5-b60f-02d008344485\") " pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.202221 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-config\") pod \"dnsmasq-dns-75bf4db767-zbs9v\" (UID: \"a95776b7-4a12-4de5-b60f-02d008344485\") " pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.230672 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nh2qg\" (UniqueName: \"kubernetes.io/projected/a95776b7-4a12-4de5-b60f-02d008344485-kube-api-access-nh2qg\") pod \"dnsmasq-dns-75bf4db767-zbs9v\" (UID: \"a95776b7-4a12-4de5-b60f-02d008344485\") " pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.302866 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ftgk\" (UniqueName: \"kubernetes.io/projected/ed97e535-9755-40ef-a82f-176f0eb7e31b-kube-api-access-6ftgk\") pod \"neutron-c888c6fcb-jnt6f\" (UID: \"ed97e535-9755-40ef-a82f-176f0eb7e31b\") " pod="openstack/neutron-c888c6fcb-jnt6f" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.303229 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-combined-ca-bundle\") pod \"neutron-c888c6fcb-jnt6f\" (UID: \"ed97e535-9755-40ef-a82f-176f0eb7e31b\") " pod="openstack/neutron-c888c6fcb-jnt6f" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.303253 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-ovndb-tls-certs\") pod \"neutron-c888c6fcb-jnt6f\" (UID: \"ed97e535-9755-40ef-a82f-176f0eb7e31b\") " pod="openstack/neutron-c888c6fcb-jnt6f" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.303292 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-config\") pod \"neutron-c888c6fcb-jnt6f\" (UID: \"ed97e535-9755-40ef-a82f-176f0eb7e31b\") " pod="openstack/neutron-c888c6fcb-jnt6f" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.303378 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-httpd-config\") pod \"neutron-c888c6fcb-jnt6f\" (UID: \"ed97e535-9755-40ef-a82f-176f0eb7e31b\") " pod="openstack/neutron-c888c6fcb-jnt6f" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.313286 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-httpd-config\") pod \"neutron-c888c6fcb-jnt6f\" (UID: \"ed97e535-9755-40ef-a82f-176f0eb7e31b\") " pod="openstack/neutron-c888c6fcb-jnt6f" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.313999 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-combined-ca-bundle\") pod \"neutron-c888c6fcb-jnt6f\" (UID: \"ed97e535-9755-40ef-a82f-176f0eb7e31b\") " pod="openstack/neutron-c888c6fcb-jnt6f" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.317210 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-config\") pod \"neutron-c888c6fcb-jnt6f\" (UID: \"ed97e535-9755-40ef-a82f-176f0eb7e31b\") " pod="openstack/neutron-c888c6fcb-jnt6f" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.317809 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-ovndb-tls-certs\") pod \"neutron-c888c6fcb-jnt6f\" (UID: \"ed97e535-9755-40ef-a82f-176f0eb7e31b\") " pod="openstack/neutron-c888c6fcb-jnt6f" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.322130 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ftgk\" (UniqueName: \"kubernetes.io/projected/ed97e535-9755-40ef-a82f-176f0eb7e31b-kube-api-access-6ftgk\") pod \"neutron-c888c6fcb-jnt6f\" (UID: \"ed97e535-9755-40ef-a82f-176f0eb7e31b\") " pod="openstack/neutron-c888c6fcb-jnt6f" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.352255 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" Mar 18 20:26:13 crc kubenswrapper[4950]: I0318 20:26:13.456941 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c888c6fcb-jnt6f" Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.743671 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-42zwb" Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.827522 4950 generic.go:334] "Generic (PLEG): container finished" podID="d8589447-6f54-4a2b-80b5-29fdbccb4971" containerID="ef83c6302a1ab85f29e24e83bef380a024d32fad1c7855a688e1adbe340d26bf" exitCode=0 Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.827627 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dfbp4" event={"ID":"d8589447-6f54-4a2b-80b5-29fdbccb4971","Type":"ContainerDied","Data":"ef83c6302a1ab85f29e24e83bef380a024d32fad1c7855a688e1adbe340d26bf"} Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.829732 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-42zwb" event={"ID":"5989cc6f-e4c7-4511-a148-293f683ede13","Type":"ContainerDied","Data":"957572ed504871d866513e222027abb3fc3002243e48c31bae435a2d4ebcf2d3"} Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.829763 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="957572ed504871d866513e222027abb3fc3002243e48c31bae435a2d4ebcf2d3" Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.832251 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-42zwb" Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.846224 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5989cc6f-e4c7-4511-a148-293f683ede13-scripts\") pod \"5989cc6f-e4c7-4511-a148-293f683ede13\" (UID: \"5989cc6f-e4c7-4511-a148-293f683ede13\") " Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.846275 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5989cc6f-e4c7-4511-a148-293f683ede13-config-data\") pod \"5989cc6f-e4c7-4511-a148-293f683ede13\" (UID: \"5989cc6f-e4c7-4511-a148-293f683ede13\") " Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.846317 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8rdc\" (UniqueName: \"kubernetes.io/projected/5989cc6f-e4c7-4511-a148-293f683ede13-kube-api-access-d8rdc\") pod \"5989cc6f-e4c7-4511-a148-293f683ede13\" (UID: \"5989cc6f-e4c7-4511-a148-293f683ede13\") " Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.846431 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5989cc6f-e4c7-4511-a148-293f683ede13-combined-ca-bundle\") pod \"5989cc6f-e4c7-4511-a148-293f683ede13\" (UID: \"5989cc6f-e4c7-4511-a148-293f683ede13\") " Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.846553 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5989cc6f-e4c7-4511-a148-293f683ede13-logs\") pod \"5989cc6f-e4c7-4511-a148-293f683ede13\" (UID: \"5989cc6f-e4c7-4511-a148-293f683ede13\") " Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.847283 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5989cc6f-e4c7-4511-a148-293f683ede13-logs" (OuterVolumeSpecName: "logs") pod "5989cc6f-e4c7-4511-a148-293f683ede13" (UID: "5989cc6f-e4c7-4511-a148-293f683ede13"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.856922 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5989cc6f-e4c7-4511-a148-293f683ede13-scripts" (OuterVolumeSpecName: "scripts") pod "5989cc6f-e4c7-4511-a148-293f683ede13" (UID: "5989cc6f-e4c7-4511-a148-293f683ede13"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.857000 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5989cc6f-e4c7-4511-a148-293f683ede13-kube-api-access-d8rdc" (OuterVolumeSpecName: "kube-api-access-d8rdc") pod "5989cc6f-e4c7-4511-a148-293f683ede13" (UID: "5989cc6f-e4c7-4511-a148-293f683ede13"). InnerVolumeSpecName "kube-api-access-d8rdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.883330 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5989cc6f-e4c7-4511-a148-293f683ede13-config-data" (OuterVolumeSpecName: "config-data") pod "5989cc6f-e4c7-4511-a148-293f683ede13" (UID: "5989cc6f-e4c7-4511-a148-293f683ede13"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.902574 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5989cc6f-e4c7-4511-a148-293f683ede13-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5989cc6f-e4c7-4511-a148-293f683ede13" (UID: "5989cc6f-e4c7-4511-a148-293f683ede13"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.963031 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5989cc6f-e4c7-4511-a148-293f683ede13-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.963066 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5989cc6f-e4c7-4511-a148-293f683ede13-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.963077 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8rdc\" (UniqueName: \"kubernetes.io/projected/5989cc6f-e4c7-4511-a148-293f683ede13-kube-api-access-d8rdc\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.963090 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5989cc6f-e4c7-4511-a148-293f683ede13-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.963115 4950 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5989cc6f-e4c7-4511-a148-293f683ede13-logs\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.969008 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5d979b4d49-bzq8q"] Mar 18 20:26:15 crc kubenswrapper[4950]: E0318 20:26:15.971912 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5989cc6f-e4c7-4511-a148-293f683ede13" containerName="placement-db-sync" Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.971965 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="5989cc6f-e4c7-4511-a148-293f683ede13" containerName="placement-db-sync" Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.972427 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="5989cc6f-e4c7-4511-a148-293f683ede13" containerName="placement-db-sync" Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.994241 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.998067 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Mar 18 20:26:15 crc kubenswrapper[4950]: I0318 20:26:15.998105 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.024040 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d979b4d49-bzq8q"] Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.064466 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-internal-tls-certs\") pod \"neutron-5d979b4d49-bzq8q\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.064518 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcj7f\" (UniqueName: \"kubernetes.io/projected/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-kube-api-access-vcj7f\") pod \"neutron-5d979b4d49-bzq8q\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.064566 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-httpd-config\") pod \"neutron-5d979b4d49-bzq8q\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.064656 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-combined-ca-bundle\") pod \"neutron-5d979b4d49-bzq8q\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.064707 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-config\") pod \"neutron-5d979b4d49-bzq8q\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.064737 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-public-tls-certs\") pod \"neutron-5d979b4d49-bzq8q\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.064762 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-ovndb-tls-certs\") pod \"neutron-5d979b4d49-bzq8q\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.166200 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-combined-ca-bundle\") pod \"neutron-5d979b4d49-bzq8q\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.166262 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-config\") pod \"neutron-5d979b4d49-bzq8q\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.166291 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-public-tls-certs\") pod \"neutron-5d979b4d49-bzq8q\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.166309 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-ovndb-tls-certs\") pod \"neutron-5d979b4d49-bzq8q\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.166359 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-internal-tls-certs\") pod \"neutron-5d979b4d49-bzq8q\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.166374 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcj7f\" (UniqueName: \"kubernetes.io/projected/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-kube-api-access-vcj7f\") pod \"neutron-5d979b4d49-bzq8q\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.166403 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-httpd-config\") pod \"neutron-5d979b4d49-bzq8q\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.173020 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-public-tls-certs\") pod \"neutron-5d979b4d49-bzq8q\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.173299 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-internal-tls-certs\") pod \"neutron-5d979b4d49-bzq8q\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.173399 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-config\") pod \"neutron-5d979b4d49-bzq8q\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.174695 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-ovndb-tls-certs\") pod \"neutron-5d979b4d49-bzq8q\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.177201 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-httpd-config\") pod \"neutron-5d979b4d49-bzq8q\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.180565 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-combined-ca-bundle\") pod \"neutron-5d979b4d49-bzq8q\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.183933 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcj7f\" (UniqueName: \"kubernetes.io/projected/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-kube-api-access-vcj7f\") pod \"neutron-5d979b4d49-bzq8q\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.318836 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.955692 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-67f985b6b-xsntq"] Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.957177 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.964208 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-twmnm" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.964284 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.964398 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.964589 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.964620 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Mar 18 20:26:16 crc kubenswrapper[4950]: I0318 20:26:16.970184 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-67f985b6b-xsntq"] Mar 18 20:26:17 crc kubenswrapper[4950]: I0318 20:26:17.080841 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-scripts\") pod \"placement-67f985b6b-xsntq\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:17 crc kubenswrapper[4950]: I0318 20:26:17.080887 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-config-data\") pod \"placement-67f985b6b-xsntq\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:17 crc kubenswrapper[4950]: I0318 20:26:17.080920 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-public-tls-certs\") pod \"placement-67f985b6b-xsntq\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:17 crc kubenswrapper[4950]: I0318 20:26:17.081001 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-internal-tls-certs\") pod \"placement-67f985b6b-xsntq\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:17 crc kubenswrapper[4950]: I0318 20:26:17.081029 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-combined-ca-bundle\") pod \"placement-67f985b6b-xsntq\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:17 crc kubenswrapper[4950]: I0318 20:26:17.081046 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cebbda36-bfd8-4115-bdaf-616051b00eb0-logs\") pod \"placement-67f985b6b-xsntq\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:17 crc kubenswrapper[4950]: I0318 20:26:17.081233 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq8gh\" (UniqueName: \"kubernetes.io/projected/cebbda36-bfd8-4115-bdaf-616051b00eb0-kube-api-access-sq8gh\") pod \"placement-67f985b6b-xsntq\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:17 crc kubenswrapper[4950]: I0318 20:26:17.182659 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq8gh\" (UniqueName: \"kubernetes.io/projected/cebbda36-bfd8-4115-bdaf-616051b00eb0-kube-api-access-sq8gh\") pod \"placement-67f985b6b-xsntq\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:17 crc kubenswrapper[4950]: I0318 20:26:17.182744 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-scripts\") pod \"placement-67f985b6b-xsntq\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:17 crc kubenswrapper[4950]: I0318 20:26:17.182767 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-config-data\") pod \"placement-67f985b6b-xsntq\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:17 crc kubenswrapper[4950]: I0318 20:26:17.182800 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-public-tls-certs\") pod \"placement-67f985b6b-xsntq\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:17 crc kubenswrapper[4950]: I0318 20:26:17.182839 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-internal-tls-certs\") pod \"placement-67f985b6b-xsntq\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:17 crc kubenswrapper[4950]: I0318 20:26:17.182865 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-combined-ca-bundle\") pod \"placement-67f985b6b-xsntq\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:17 crc kubenswrapper[4950]: I0318 20:26:17.182881 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cebbda36-bfd8-4115-bdaf-616051b00eb0-logs\") pod \"placement-67f985b6b-xsntq\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:17 crc kubenswrapper[4950]: I0318 20:26:17.183312 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cebbda36-bfd8-4115-bdaf-616051b00eb0-logs\") pod \"placement-67f985b6b-xsntq\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:17 crc kubenswrapper[4950]: I0318 20:26:17.188422 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-scripts\") pod \"placement-67f985b6b-xsntq\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:17 crc kubenswrapper[4950]: I0318 20:26:17.189389 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-combined-ca-bundle\") pod \"placement-67f985b6b-xsntq\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:17 crc kubenswrapper[4950]: I0318 20:26:17.189733 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-internal-tls-certs\") pod \"placement-67f985b6b-xsntq\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:17 crc kubenswrapper[4950]: I0318 20:26:17.190209 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-public-tls-certs\") pod \"placement-67f985b6b-xsntq\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:17 crc kubenswrapper[4950]: I0318 20:26:17.190294 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-config-data\") pod \"placement-67f985b6b-xsntq\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:17 crc kubenswrapper[4950]: I0318 20:26:17.216937 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq8gh\" (UniqueName: \"kubernetes.io/projected/cebbda36-bfd8-4115-bdaf-616051b00eb0-kube-api-access-sq8gh\") pod \"placement-67f985b6b-xsntq\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:17 crc kubenswrapper[4950]: I0318 20:26:17.281497 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.142022 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dfbp4" Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.148130 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564426-4lcpm" Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.272500 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-config-data\") pod \"d8589447-6f54-4a2b-80b5-29fdbccb4971\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.272907 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-credential-keys\") pod \"d8589447-6f54-4a2b-80b5-29fdbccb4971\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.272942 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-combined-ca-bundle\") pod \"d8589447-6f54-4a2b-80b5-29fdbccb4971\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.273027 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-scripts\") pod \"d8589447-6f54-4a2b-80b5-29fdbccb4971\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.273126 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xn2ws\" (UniqueName: \"kubernetes.io/projected/d8589447-6f54-4a2b-80b5-29fdbccb4971-kube-api-access-xn2ws\") pod \"d8589447-6f54-4a2b-80b5-29fdbccb4971\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.273177 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-fernet-keys\") pod \"d8589447-6f54-4a2b-80b5-29fdbccb4971\" (UID: \"d8589447-6f54-4a2b-80b5-29fdbccb4971\") " Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.273235 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtfrf\" (UniqueName: \"kubernetes.io/projected/619d4317-c521-4032-a59f-067f9914bb44-kube-api-access-dtfrf\") pod \"619d4317-c521-4032-a59f-067f9914bb44\" (UID: \"619d4317-c521-4032-a59f-067f9914bb44\") " Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.281089 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-scripts" (OuterVolumeSpecName: "scripts") pod "d8589447-6f54-4a2b-80b5-29fdbccb4971" (UID: "d8589447-6f54-4a2b-80b5-29fdbccb4971"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.284236 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "d8589447-6f54-4a2b-80b5-29fdbccb4971" (UID: "d8589447-6f54-4a2b-80b5-29fdbccb4971"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.299879 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/619d4317-c521-4032-a59f-067f9914bb44-kube-api-access-dtfrf" (OuterVolumeSpecName: "kube-api-access-dtfrf") pod "619d4317-c521-4032-a59f-067f9914bb44" (UID: "619d4317-c521-4032-a59f-067f9914bb44"). InnerVolumeSpecName "kube-api-access-dtfrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.306960 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8589447-6f54-4a2b-80b5-29fdbccb4971-kube-api-access-xn2ws" (OuterVolumeSpecName: "kube-api-access-xn2ws") pod "d8589447-6f54-4a2b-80b5-29fdbccb4971" (UID: "d8589447-6f54-4a2b-80b5-29fdbccb4971"). InnerVolumeSpecName "kube-api-access-xn2ws". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.318784 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d8589447-6f54-4a2b-80b5-29fdbccb4971" (UID: "d8589447-6f54-4a2b-80b5-29fdbccb4971"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.342964 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d8589447-6f54-4a2b-80b5-29fdbccb4971" (UID: "d8589447-6f54-4a2b-80b5-29fdbccb4971"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.347304 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-config-data" (OuterVolumeSpecName: "config-data") pod "d8589447-6f54-4a2b-80b5-29fdbccb4971" (UID: "d8589447-6f54-4a2b-80b5-29fdbccb4971"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.375474 4950 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.375515 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.375525 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.375536 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xn2ws\" (UniqueName: \"kubernetes.io/projected/d8589447-6f54-4a2b-80b5-29fdbccb4971-kube-api-access-xn2ws\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.375546 4950 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.375554 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtfrf\" (UniqueName: \"kubernetes.io/projected/619d4317-c521-4032-a59f-067f9914bb44-kube-api-access-dtfrf\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.375562 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8589447-6f54-4a2b-80b5-29fdbccb4971-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.469861 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75bf4db767-zbs9v"] Mar 18 20:26:21 crc kubenswrapper[4950]: W0318 20:26:21.473648 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda95776b7_4a12_4de5_b60f_02d008344485.slice/crio-dc760a0ee02c4c6778cec4f216b5ae8418671bdb044ea749af5a31f372f92640 WatchSource:0}: Error finding container dc760a0ee02c4c6778cec4f216b5ae8418671bdb044ea749af5a31f372f92640: Status 404 returned error can't find the container with id dc760a0ee02c4c6778cec4f216b5ae8418671bdb044ea749af5a31f372f92640 Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.729084 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d979b4d49-bzq8q"] Mar 18 20:26:21 crc kubenswrapper[4950]: W0318 20:26:21.734884 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31bbaba1_cd09_469b_9b89_f7ff2213cd8e.slice/crio-4883a046fb74a10d04787d9d53f3adeb5aadbda6f5aea920fd78682a805ac980 WatchSource:0}: Error finding container 4883a046fb74a10d04787d9d53f3adeb5aadbda6f5aea920fd78682a805ac980: Status 404 returned error can't find the container with id 4883a046fb74a10d04787d9d53f3adeb5aadbda6f5aea920fd78682a805ac980 Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.782198 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-67f985b6b-xsntq"] Mar 18 20:26:21 crc kubenswrapper[4950]: W0318 20:26:21.786039 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcebbda36_bfd8_4115_bdaf_616051b00eb0.slice/crio-366be91b14609f8f748dc76b9a1376386ea107b218e76a95421561e8e5f46136 WatchSource:0}: Error finding container 366be91b14609f8f748dc76b9a1376386ea107b218e76a95421561e8e5f46136: Status 404 returned error can't find the container with id 366be91b14609f8f748dc76b9a1376386ea107b218e76a95421561e8e5f46136 Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.893526 4950 generic.go:334] "Generic (PLEG): container finished" podID="a95776b7-4a12-4de5-b60f-02d008344485" containerID="a52ada126793a6eb000243037a35dc25cd43e2dfca4ac4a2872b35cc11e88ae8" exitCode=0 Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.893582 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" event={"ID":"a95776b7-4a12-4de5-b60f-02d008344485","Type":"ContainerDied","Data":"a52ada126793a6eb000243037a35dc25cd43e2dfca4ac4a2872b35cc11e88ae8"} Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.893606 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" event={"ID":"a95776b7-4a12-4de5-b60f-02d008344485","Type":"ContainerStarted","Data":"dc760a0ee02c4c6778cec4f216b5ae8418671bdb044ea749af5a31f372f92640"} Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.897959 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7563774c-0a5c-4af9-802a-662f4e721f3c","Type":"ContainerStarted","Data":"e609f6cce6ca82ac4a1ba760e29aadc780269f67316528034724eda1d330732d"} Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.900978 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67f985b6b-xsntq" event={"ID":"cebbda36-bfd8-4115-bdaf-616051b00eb0","Type":"ContainerStarted","Data":"366be91b14609f8f748dc76b9a1376386ea107b218e76a95421561e8e5f46136"} Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.902774 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d979b4d49-bzq8q" event={"ID":"31bbaba1-cd09-469b-9b89-f7ff2213cd8e","Type":"ContainerStarted","Data":"0c993414a215e6519fba3c4c77718278e352a22c3d3bad0350254d5c89624b76"} Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.902802 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d979b4d49-bzq8q" event={"ID":"31bbaba1-cd09-469b-9b89-f7ff2213cd8e","Type":"ContainerStarted","Data":"4883a046fb74a10d04787d9d53f3adeb5aadbda6f5aea920fd78682a805ac980"} Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.904397 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dfbp4" event={"ID":"d8589447-6f54-4a2b-80b5-29fdbccb4971","Type":"ContainerDied","Data":"842b6dc1d490c53b741456c0e5c030a31484193029da75c0986e28072fd8101b"} Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.904437 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="842b6dc1d490c53b741456c0e5c030a31484193029da75c0986e28072fd8101b" Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.904494 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dfbp4" Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.906680 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564426-4lcpm" event={"ID":"619d4317-c521-4032-a59f-067f9914bb44","Type":"ContainerDied","Data":"73b58b8c55f6bf2f6de089b4b8dcd609ffc0086d47b64be2013435efd05aa797"} Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.906701 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73b58b8c55f6bf2f6de089b4b8dcd609ffc0086d47b64be2013435efd05aa797" Mar 18 20:26:21 crc kubenswrapper[4950]: I0318 20:26:21.906747 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564426-4lcpm" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.245220 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564420-54tbm"] Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.258065 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564420-54tbm"] Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.272401 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-564b994b87-rbnt8"] Mar 18 20:26:22 crc kubenswrapper[4950]: E0318 20:26:22.273049 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="619d4317-c521-4032-a59f-067f9914bb44" containerName="oc" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.273110 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="619d4317-c521-4032-a59f-067f9914bb44" containerName="oc" Mar 18 20:26:22 crc kubenswrapper[4950]: E0318 20:26:22.273167 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8589447-6f54-4a2b-80b5-29fdbccb4971" containerName="keystone-bootstrap" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.273214 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8589447-6f54-4a2b-80b5-29fdbccb4971" containerName="keystone-bootstrap" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.273635 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="619d4317-c521-4032-a59f-067f9914bb44" containerName="oc" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.273721 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8589447-6f54-4a2b-80b5-29fdbccb4971" containerName="keystone-bootstrap" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.274257 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.281762 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-bsdbw" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.281904 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.281951 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.282065 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.282139 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.284244 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.289719 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-564b994b87-rbnt8"] Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.390389 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb4d97b7-3492-4ccb-8489-98a9882ea782-public-tls-certs\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.390446 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fb4d97b7-3492-4ccb-8489-98a9882ea782-fernet-keys\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.390470 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnpzc\" (UniqueName: \"kubernetes.io/projected/fb4d97b7-3492-4ccb-8489-98a9882ea782-kube-api-access-mnpzc\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.390492 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb4d97b7-3492-4ccb-8489-98a9882ea782-combined-ca-bundle\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.390561 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb4d97b7-3492-4ccb-8489-98a9882ea782-internal-tls-certs\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.390582 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fb4d97b7-3492-4ccb-8489-98a9882ea782-credential-keys\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.390639 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb4d97b7-3492-4ccb-8489-98a9882ea782-scripts\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.390661 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb4d97b7-3492-4ccb-8489-98a9882ea782-config-data\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.492033 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnpzc\" (UniqueName: \"kubernetes.io/projected/fb4d97b7-3492-4ccb-8489-98a9882ea782-kube-api-access-mnpzc\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.492090 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb4d97b7-3492-4ccb-8489-98a9882ea782-combined-ca-bundle\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.492180 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb4d97b7-3492-4ccb-8489-98a9882ea782-internal-tls-certs\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.492202 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fb4d97b7-3492-4ccb-8489-98a9882ea782-credential-keys\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.492286 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb4d97b7-3492-4ccb-8489-98a9882ea782-scripts\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.492328 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb4d97b7-3492-4ccb-8489-98a9882ea782-config-data\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.492349 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb4d97b7-3492-4ccb-8489-98a9882ea782-public-tls-certs\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.492383 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fb4d97b7-3492-4ccb-8489-98a9882ea782-fernet-keys\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.497265 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb4d97b7-3492-4ccb-8489-98a9882ea782-public-tls-certs\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.498104 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fb4d97b7-3492-4ccb-8489-98a9882ea782-fernet-keys\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.500153 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb4d97b7-3492-4ccb-8489-98a9882ea782-config-data\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.505725 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb4d97b7-3492-4ccb-8489-98a9882ea782-combined-ca-bundle\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.506923 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fb4d97b7-3492-4ccb-8489-98a9882ea782-credential-keys\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.509004 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb4d97b7-3492-4ccb-8489-98a9882ea782-internal-tls-certs\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.509103 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb4d97b7-3492-4ccb-8489-98a9882ea782-scripts\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.511827 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnpzc\" (UniqueName: \"kubernetes.io/projected/fb4d97b7-3492-4ccb-8489-98a9882ea782-kube-api-access-mnpzc\") pod \"keystone-564b994b87-rbnt8\" (UID: \"fb4d97b7-3492-4ccb-8489-98a9882ea782\") " pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.670314 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.842156 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c888c6fcb-jnt6f"] Mar 18 20:26:22 crc kubenswrapper[4950]: W0318 20:26:22.895868 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded97e535_9755_40ef_a82f_176f0eb7e31b.slice/crio-901e81991ebfa5935dfb7f6c7620365fc8bb2119a1f1e2e42f586862cab0a475 WatchSource:0}: Error finding container 901e81991ebfa5935dfb7f6c7620365fc8bb2119a1f1e2e42f586862cab0a475: Status 404 returned error can't find the container with id 901e81991ebfa5935dfb7f6c7620365fc8bb2119a1f1e2e42f586862cab0a475 Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.949225 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67f985b6b-xsntq" event={"ID":"cebbda36-bfd8-4115-bdaf-616051b00eb0","Type":"ContainerStarted","Data":"071eed9b9bf77cd21c4fb507848fedfa3b7301e0abdc706543d2fea57787a0d5"} Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.949592 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67f985b6b-xsntq" event={"ID":"cebbda36-bfd8-4115-bdaf-616051b00eb0","Type":"ContainerStarted","Data":"37b7d0561b5fbb0d11bf9f3c9faf271dff5effbb99739498ac66dc22afa16116"} Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.950506 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.950531 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.956689 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d979b4d49-bzq8q" event={"ID":"31bbaba1-cd09-469b-9b89-f7ff2213cd8e","Type":"ContainerStarted","Data":"8c5841219d263c4c634cb638c391cc932d8c4bf30892f6dc7b784816023de26e"} Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.957342 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.962565 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c888c6fcb-jnt6f" event={"ID":"ed97e535-9755-40ef-a82f-176f0eb7e31b","Type":"ContainerStarted","Data":"901e81991ebfa5935dfb7f6c7620365fc8bb2119a1f1e2e42f586862cab0a475"} Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.982727 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-brgnq" event={"ID":"82a7c2a4-9f57-460e-8bac-419f14222502","Type":"ContainerStarted","Data":"08cf42bf1bb71302f9083c4aa0908a4cd8aed1c546a9aa614f9927b0b0ad5bf8"} Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.983309 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-67f985b6b-xsntq" podStartSLOduration=6.983286209 podStartE2EDuration="6.983286209s" podCreationTimestamp="2026-03-18 20:26:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:26:22.981127583 +0000 UTC m=+1196.221969451" watchObservedRunningTime="2026-03-18 20:26:22.983286209 +0000 UTC m=+1196.224128077" Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.990347 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" event={"ID":"a95776b7-4a12-4de5-b60f-02d008344485","Type":"ContainerStarted","Data":"8c8dfa2860fb1e62b894f7d9111f01263d57835534226ef39a1bca9a7368ddba"} Mar 18 20:26:22 crc kubenswrapper[4950]: I0318 20:26:22.991157 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" Mar 18 20:26:23 crc kubenswrapper[4950]: I0318 20:26:23.004275 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5d979b4d49-bzq8q" podStartSLOduration=8.004229788 podStartE2EDuration="8.004229788s" podCreationTimestamp="2026-03-18 20:26:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:26:22.998969213 +0000 UTC m=+1196.239811081" watchObservedRunningTime="2026-03-18 20:26:23.004229788 +0000 UTC m=+1196.245071656" Mar 18 20:26:23 crc kubenswrapper[4950]: I0318 20:26:23.050015 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" podStartSLOduration=11.049993366 podStartE2EDuration="11.049993366s" podCreationTimestamp="2026-03-18 20:26:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:26:23.031062489 +0000 UTC m=+1196.271904357" watchObservedRunningTime="2026-03-18 20:26:23.049993366 +0000 UTC m=+1196.290835234" Mar 18 20:26:23 crc kubenswrapper[4950]: I0318 20:26:23.063628 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-brgnq" podStartSLOduration=3.376439392 podStartE2EDuration="38.063609987s" podCreationTimestamp="2026-03-18 20:25:45 +0000 UTC" firstStartedPulling="2026-03-18 20:25:47.357081256 +0000 UTC m=+1160.597923114" lastFinishedPulling="2026-03-18 20:26:22.044251841 +0000 UTC m=+1195.285093709" observedRunningTime="2026-03-18 20:26:23.055666623 +0000 UTC m=+1196.296508491" watchObservedRunningTime="2026-03-18 20:26:23.063609987 +0000 UTC m=+1196.304451855" Mar 18 20:26:23 crc kubenswrapper[4950]: I0318 20:26:23.193117 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-564b994b87-rbnt8"] Mar 18 20:26:23 crc kubenswrapper[4950]: W0318 20:26:23.195379 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb4d97b7_3492_4ccb_8489_98a9882ea782.slice/crio-f212e0430eedd48518f3eb337befd0c250b89a7c388f4f42ef85cbd381d15bdf WatchSource:0}: Error finding container f212e0430eedd48518f3eb337befd0c250b89a7c388f4f42ef85cbd381d15bdf: Status 404 returned error can't find the container with id f212e0430eedd48518f3eb337befd0c250b89a7c388f4f42ef85cbd381d15bdf Mar 18 20:26:23 crc kubenswrapper[4950]: I0318 20:26:23.504037 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bad5973-c35e-4a25-9ba4-50958354a198" path="/var/lib/kubelet/pods/3bad5973-c35e-4a25-9ba4-50958354a198/volumes" Mar 18 20:26:24 crc kubenswrapper[4950]: I0318 20:26:24.001506 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c888c6fcb-jnt6f" event={"ID":"ed97e535-9755-40ef-a82f-176f0eb7e31b","Type":"ContainerStarted","Data":"e3df6aff85a9f7990f7a5c74c826e0e7d768d6648c7249bf7e4b60b739cb95f6"} Mar 18 20:26:24 crc kubenswrapper[4950]: I0318 20:26:24.001820 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c888c6fcb-jnt6f" event={"ID":"ed97e535-9755-40ef-a82f-176f0eb7e31b","Type":"ContainerStarted","Data":"74f0ac325be8ead0f0e1eaebbcab54a265d706045bb6cc501683c742e2f9b28f"} Mar 18 20:26:24 crc kubenswrapper[4950]: I0318 20:26:24.002846 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-c888c6fcb-jnt6f" Mar 18 20:26:24 crc kubenswrapper[4950]: I0318 20:26:24.006466 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-564b994b87-rbnt8" event={"ID":"fb4d97b7-3492-4ccb-8489-98a9882ea782","Type":"ContainerStarted","Data":"98d6f9599427df12da4c7b5ed592fb02001f42da29a687245735d40b74f31413"} Mar 18 20:26:24 crc kubenswrapper[4950]: I0318 20:26:24.006494 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-564b994b87-rbnt8" event={"ID":"fb4d97b7-3492-4ccb-8489-98a9882ea782","Type":"ContainerStarted","Data":"f212e0430eedd48518f3eb337befd0c250b89a7c388f4f42ef85cbd381d15bdf"} Mar 18 20:26:24 crc kubenswrapper[4950]: I0318 20:26:24.006985 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:24 crc kubenswrapper[4950]: I0318 20:26:24.032326 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-c888c6fcb-jnt6f" podStartSLOduration=11.032309108 podStartE2EDuration="11.032309108s" podCreationTimestamp="2026-03-18 20:26:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:26:24.028585352 +0000 UTC m=+1197.269427220" watchObservedRunningTime="2026-03-18 20:26:24.032309108 +0000 UTC m=+1197.273150976" Mar 18 20:26:25 crc kubenswrapper[4950]: I0318 20:26:25.531053 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-564b994b87-rbnt8" podStartSLOduration=3.531038807 podStartE2EDuration="3.531038807s" podCreationTimestamp="2026-03-18 20:26:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:26:24.072692908 +0000 UTC m=+1197.313534776" watchObservedRunningTime="2026-03-18 20:26:25.531038807 +0000 UTC m=+1198.771880675" Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.628376 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-c888c6fcb-jnt6f"] Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.656474 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-d5c6d4b4c-6cdzp"] Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.657960 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.673144 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d5c6d4b4c-6cdzp"] Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.773292 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2afba89a-006e-472f-8e49-0e8d69cd86c7-public-tls-certs\") pod \"neutron-d5c6d4b4c-6cdzp\" (UID: \"2afba89a-006e-472f-8e49-0e8d69cd86c7\") " pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.773368 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2afba89a-006e-472f-8e49-0e8d69cd86c7-config\") pod \"neutron-d5c6d4b4c-6cdzp\" (UID: \"2afba89a-006e-472f-8e49-0e8d69cd86c7\") " pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.773386 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2afba89a-006e-472f-8e49-0e8d69cd86c7-httpd-config\") pod \"neutron-d5c6d4b4c-6cdzp\" (UID: \"2afba89a-006e-472f-8e49-0e8d69cd86c7\") " pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.773444 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2afba89a-006e-472f-8e49-0e8d69cd86c7-internal-tls-certs\") pod \"neutron-d5c6d4b4c-6cdzp\" (UID: \"2afba89a-006e-472f-8e49-0e8d69cd86c7\") " pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.773587 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2afba89a-006e-472f-8e49-0e8d69cd86c7-ovndb-tls-certs\") pod \"neutron-d5c6d4b4c-6cdzp\" (UID: \"2afba89a-006e-472f-8e49-0e8d69cd86c7\") " pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.773612 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2afba89a-006e-472f-8e49-0e8d69cd86c7-combined-ca-bundle\") pod \"neutron-d5c6d4b4c-6cdzp\" (UID: \"2afba89a-006e-472f-8e49-0e8d69cd86c7\") " pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.773633 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcj6b\" (UniqueName: \"kubernetes.io/projected/2afba89a-006e-472f-8e49-0e8d69cd86c7-kube-api-access-zcj6b\") pod \"neutron-d5c6d4b4c-6cdzp\" (UID: \"2afba89a-006e-472f-8e49-0e8d69cd86c7\") " pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.874999 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2afba89a-006e-472f-8e49-0e8d69cd86c7-ovndb-tls-certs\") pod \"neutron-d5c6d4b4c-6cdzp\" (UID: \"2afba89a-006e-472f-8e49-0e8d69cd86c7\") " pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.875959 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2afba89a-006e-472f-8e49-0e8d69cd86c7-combined-ca-bundle\") pod \"neutron-d5c6d4b4c-6cdzp\" (UID: \"2afba89a-006e-472f-8e49-0e8d69cd86c7\") " pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.875984 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcj6b\" (UniqueName: \"kubernetes.io/projected/2afba89a-006e-472f-8e49-0e8d69cd86c7-kube-api-access-zcj6b\") pod \"neutron-d5c6d4b4c-6cdzp\" (UID: \"2afba89a-006e-472f-8e49-0e8d69cd86c7\") " pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.876044 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2afba89a-006e-472f-8e49-0e8d69cd86c7-public-tls-certs\") pod \"neutron-d5c6d4b4c-6cdzp\" (UID: \"2afba89a-006e-472f-8e49-0e8d69cd86c7\") " pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.876075 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2afba89a-006e-472f-8e49-0e8d69cd86c7-httpd-config\") pod \"neutron-d5c6d4b4c-6cdzp\" (UID: \"2afba89a-006e-472f-8e49-0e8d69cd86c7\") " pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.876091 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2afba89a-006e-472f-8e49-0e8d69cd86c7-config\") pod \"neutron-d5c6d4b4c-6cdzp\" (UID: \"2afba89a-006e-472f-8e49-0e8d69cd86c7\") " pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.876127 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2afba89a-006e-472f-8e49-0e8d69cd86c7-internal-tls-certs\") pod \"neutron-d5c6d4b4c-6cdzp\" (UID: \"2afba89a-006e-472f-8e49-0e8d69cd86c7\") " pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.880939 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2afba89a-006e-472f-8e49-0e8d69cd86c7-internal-tls-certs\") pod \"neutron-d5c6d4b4c-6cdzp\" (UID: \"2afba89a-006e-472f-8e49-0e8d69cd86c7\") " pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.881629 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2afba89a-006e-472f-8e49-0e8d69cd86c7-config\") pod \"neutron-d5c6d4b4c-6cdzp\" (UID: \"2afba89a-006e-472f-8e49-0e8d69cd86c7\") " pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.886144 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2afba89a-006e-472f-8e49-0e8d69cd86c7-ovndb-tls-certs\") pod \"neutron-d5c6d4b4c-6cdzp\" (UID: \"2afba89a-006e-472f-8e49-0e8d69cd86c7\") " pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.886978 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2afba89a-006e-472f-8e49-0e8d69cd86c7-combined-ca-bundle\") pod \"neutron-d5c6d4b4c-6cdzp\" (UID: \"2afba89a-006e-472f-8e49-0e8d69cd86c7\") " pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.889090 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2afba89a-006e-472f-8e49-0e8d69cd86c7-httpd-config\") pod \"neutron-d5c6d4b4c-6cdzp\" (UID: \"2afba89a-006e-472f-8e49-0e8d69cd86c7\") " pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.894790 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcj6b\" (UniqueName: \"kubernetes.io/projected/2afba89a-006e-472f-8e49-0e8d69cd86c7-kube-api-access-zcj6b\") pod \"neutron-d5c6d4b4c-6cdzp\" (UID: \"2afba89a-006e-472f-8e49-0e8d69cd86c7\") " pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.894974 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2afba89a-006e-472f-8e49-0e8d69cd86c7-public-tls-certs\") pod \"neutron-d5c6d4b4c-6cdzp\" (UID: \"2afba89a-006e-472f-8e49-0e8d69cd86c7\") " pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:26 crc kubenswrapper[4950]: I0318 20:26:26.976374 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:27 crc kubenswrapper[4950]: I0318 20:26:27.035746 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-c888c6fcb-jnt6f" podUID="ed97e535-9755-40ef-a82f-176f0eb7e31b" containerName="neutron-api" containerID="cri-o://74f0ac325be8ead0f0e1eaebbcab54a265d706045bb6cc501683c742e2f9b28f" gracePeriod=30 Mar 18 20:26:27 crc kubenswrapper[4950]: I0318 20:26:27.035797 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-c888c6fcb-jnt6f" podUID="ed97e535-9755-40ef-a82f-176f0eb7e31b" containerName="neutron-httpd" containerID="cri-o://e3df6aff85a9f7990f7a5c74c826e0e7d768d6648c7249bf7e4b60b739cb95f6" gracePeriod=30 Mar 18 20:26:28 crc kubenswrapper[4950]: I0318 20:26:28.046522 4950 generic.go:334] "Generic (PLEG): container finished" podID="ed97e535-9755-40ef-a82f-176f0eb7e31b" containerID="e3df6aff85a9f7990f7a5c74c826e0e7d768d6648c7249bf7e4b60b739cb95f6" exitCode=0 Mar 18 20:26:28 crc kubenswrapper[4950]: I0318 20:26:28.046604 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c888c6fcb-jnt6f" event={"ID":"ed97e535-9755-40ef-a82f-176f0eb7e31b","Type":"ContainerDied","Data":"e3df6aff85a9f7990f7a5c74c826e0e7d768d6648c7249bf7e4b60b739cb95f6"} Mar 18 20:26:28 crc kubenswrapper[4950]: I0318 20:26:28.354580 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" Mar 18 20:26:28 crc kubenswrapper[4950]: I0318 20:26:28.412932 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74b5448f7f-zdrjg"] Mar 18 20:26:28 crc kubenswrapper[4950]: I0318 20:26:28.413152 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" podUID="ad591d5a-4120-4285-9cca-417b19176795" containerName="dnsmasq-dns" containerID="cri-o://9868c2c62d92308f1e605a851fa729dafcb7f9155140705a4f262c2522d2d1dc" gracePeriod=10 Mar 18 20:26:29 crc kubenswrapper[4950]: I0318 20:26:29.060476 4950 generic.go:334] "Generic (PLEG): container finished" podID="ad591d5a-4120-4285-9cca-417b19176795" containerID="9868c2c62d92308f1e605a851fa729dafcb7f9155140705a4f262c2522d2d1dc" exitCode=0 Mar 18 20:26:29 crc kubenswrapper[4950]: I0318 20:26:29.060652 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" event={"ID":"ad591d5a-4120-4285-9cca-417b19176795","Type":"ContainerDied","Data":"9868c2c62d92308f1e605a851fa729dafcb7f9155140705a4f262c2522d2d1dc"} Mar 18 20:26:31 crc kubenswrapper[4950]: I0318 20:26:31.681016 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" Mar 18 20:26:31 crc kubenswrapper[4950]: I0318 20:26:31.765446 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4qnj\" (UniqueName: \"kubernetes.io/projected/ad591d5a-4120-4285-9cca-417b19176795-kube-api-access-j4qnj\") pod \"ad591d5a-4120-4285-9cca-417b19176795\" (UID: \"ad591d5a-4120-4285-9cca-417b19176795\") " Mar 18 20:26:31 crc kubenswrapper[4950]: I0318 20:26:31.765852 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-ovsdbserver-nb\") pod \"ad591d5a-4120-4285-9cca-417b19176795\" (UID: \"ad591d5a-4120-4285-9cca-417b19176795\") " Mar 18 20:26:31 crc kubenswrapper[4950]: I0318 20:26:31.765935 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-config\") pod \"ad591d5a-4120-4285-9cca-417b19176795\" (UID: \"ad591d5a-4120-4285-9cca-417b19176795\") " Mar 18 20:26:31 crc kubenswrapper[4950]: I0318 20:26:31.766009 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-ovsdbserver-sb\") pod \"ad591d5a-4120-4285-9cca-417b19176795\" (UID: \"ad591d5a-4120-4285-9cca-417b19176795\") " Mar 18 20:26:31 crc kubenswrapper[4950]: I0318 20:26:31.766090 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-dns-svc\") pod \"ad591d5a-4120-4285-9cca-417b19176795\" (UID: \"ad591d5a-4120-4285-9cca-417b19176795\") " Mar 18 20:26:31 crc kubenswrapper[4950]: I0318 20:26:31.779660 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad591d5a-4120-4285-9cca-417b19176795-kube-api-access-j4qnj" (OuterVolumeSpecName: "kube-api-access-j4qnj") pod "ad591d5a-4120-4285-9cca-417b19176795" (UID: "ad591d5a-4120-4285-9cca-417b19176795"). InnerVolumeSpecName "kube-api-access-j4qnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:26:31 crc kubenswrapper[4950]: I0318 20:26:31.869564 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4qnj\" (UniqueName: \"kubernetes.io/projected/ad591d5a-4120-4285-9cca-417b19176795-kube-api-access-j4qnj\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:31 crc kubenswrapper[4950]: I0318 20:26:31.885783 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ad591d5a-4120-4285-9cca-417b19176795" (UID: "ad591d5a-4120-4285-9cca-417b19176795"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:26:31 crc kubenswrapper[4950]: I0318 20:26:31.889114 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ad591d5a-4120-4285-9cca-417b19176795" (UID: "ad591d5a-4120-4285-9cca-417b19176795"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:26:31 crc kubenswrapper[4950]: I0318 20:26:31.891631 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-config" (OuterVolumeSpecName: "config") pod "ad591d5a-4120-4285-9cca-417b19176795" (UID: "ad591d5a-4120-4285-9cca-417b19176795"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:26:31 crc kubenswrapper[4950]: I0318 20:26:31.892167 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ad591d5a-4120-4285-9cca-417b19176795" (UID: "ad591d5a-4120-4285-9cca-417b19176795"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:26:31 crc kubenswrapper[4950]: I0318 20:26:31.971255 4950 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:31 crc kubenswrapper[4950]: I0318 20:26:31.971336 4950 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:31 crc kubenswrapper[4950]: I0318 20:26:31.971350 4950 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:31 crc kubenswrapper[4950]: I0318 20:26:31.971363 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad591d5a-4120-4285-9cca-417b19176795-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:32 crc kubenswrapper[4950]: I0318 20:26:32.092779 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" event={"ID":"ad591d5a-4120-4285-9cca-417b19176795","Type":"ContainerDied","Data":"8c293bb8351af4475bcd2918ff2d3a397db802e08287df076b7ab111bbf95677"} Mar 18 20:26:32 crc kubenswrapper[4950]: I0318 20:26:32.092858 4950 scope.go:117] "RemoveContainer" containerID="9868c2c62d92308f1e605a851fa729dafcb7f9155140705a4f262c2522d2d1dc" Mar 18 20:26:32 crc kubenswrapper[4950]: I0318 20:26:32.093058 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74b5448f7f-zdrjg" Mar 18 20:26:32 crc kubenswrapper[4950]: I0318 20:26:32.096091 4950 generic.go:334] "Generic (PLEG): container finished" podID="82a7c2a4-9f57-460e-8bac-419f14222502" containerID="08cf42bf1bb71302f9083c4aa0908a4cd8aed1c546a9aa614f9927b0b0ad5bf8" exitCode=0 Mar 18 20:26:32 crc kubenswrapper[4950]: I0318 20:26:32.096162 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-brgnq" event={"ID":"82a7c2a4-9f57-460e-8bac-419f14222502","Type":"ContainerDied","Data":"08cf42bf1bb71302f9083c4aa0908a4cd8aed1c546a9aa614f9927b0b0ad5bf8"} Mar 18 20:26:32 crc kubenswrapper[4950]: I0318 20:26:32.101964 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7563774c-0a5c-4af9-802a-662f4e721f3c","Type":"ContainerStarted","Data":"da6d7c641e3ff94fdefa562e2753b88297237917dc8d3b63b84a609a7f7d2450"} Mar 18 20:26:32 crc kubenswrapper[4950]: I0318 20:26:32.102227 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7563774c-0a5c-4af9-802a-662f4e721f3c" containerName="ceilometer-central-agent" containerID="cri-o://de1b4d05b6433c8bacfa4f5808c26a249120f01056e842ff8ef61a2c85cdcfad" gracePeriod=30 Mar 18 20:26:32 crc kubenswrapper[4950]: I0318 20:26:32.103338 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 18 20:26:32 crc kubenswrapper[4950]: I0318 20:26:32.103456 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7563774c-0a5c-4af9-802a-662f4e721f3c" containerName="proxy-httpd" containerID="cri-o://da6d7c641e3ff94fdefa562e2753b88297237917dc8d3b63b84a609a7f7d2450" gracePeriod=30 Mar 18 20:26:32 crc kubenswrapper[4950]: I0318 20:26:32.103578 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7563774c-0a5c-4af9-802a-662f4e721f3c" containerName="sg-core" containerID="cri-o://e609f6cce6ca82ac4a1ba760e29aadc780269f67316528034724eda1d330732d" gracePeriod=30 Mar 18 20:26:32 crc kubenswrapper[4950]: I0318 20:26:32.103663 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7563774c-0a5c-4af9-802a-662f4e721f3c" containerName="ceilometer-notification-agent" containerID="cri-o://a94ea562bd7c3e520d36e8bc6b0f62fe4b690bcbac96af7089b92400bea93599" gracePeriod=30 Mar 18 20:26:32 crc kubenswrapper[4950]: I0318 20:26:32.108448 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-tpdmj" event={"ID":"a4bb9e1a-0694-4e00-8ba6-f2765f82b983","Type":"ContainerStarted","Data":"de9ab05a4141d7f72e470b7a98a89d526812c15d9a4174be30b8a51cc0d0d4b3"} Mar 18 20:26:32 crc kubenswrapper[4950]: I0318 20:26:32.137022 4950 scope.go:117] "RemoveContainer" containerID="1123658f52ad30114e365676ca550db7d2c0a2ef49ba5180e2fed0dc0370cc00" Mar 18 20:26:32 crc kubenswrapper[4950]: I0318 20:26:32.154859 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d5c6d4b4c-6cdzp"] Mar 18 20:26:32 crc kubenswrapper[4950]: I0318 20:26:32.158397 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.736837294 podStartE2EDuration="47.158377826s" podCreationTimestamp="2026-03-18 20:25:45 +0000 UTC" firstStartedPulling="2026-03-18 20:25:47.176845565 +0000 UTC m=+1160.417687433" lastFinishedPulling="2026-03-18 20:26:31.598386107 +0000 UTC m=+1204.839227965" observedRunningTime="2026-03-18 20:26:32.152640398 +0000 UTC m=+1205.393482266" watchObservedRunningTime="2026-03-18 20:26:32.158377826 +0000 UTC m=+1205.399219694" Mar 18 20:26:32 crc kubenswrapper[4950]: I0318 20:26:32.174074 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-tpdmj" podStartSLOduration=2.497417486 podStartE2EDuration="46.1740586s" podCreationTimestamp="2026-03-18 20:25:46 +0000 UTC" firstStartedPulling="2026-03-18 20:25:47.84625096 +0000 UTC m=+1161.087092828" lastFinishedPulling="2026-03-18 20:26:31.522892074 +0000 UTC m=+1204.763733942" observedRunningTime="2026-03-18 20:26:32.166923156 +0000 UTC m=+1205.407765024" watchObservedRunningTime="2026-03-18 20:26:32.1740586 +0000 UTC m=+1205.414900468" Mar 18 20:26:32 crc kubenswrapper[4950]: I0318 20:26:32.188113 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74b5448f7f-zdrjg"] Mar 18 20:26:32 crc kubenswrapper[4950]: I0318 20:26:32.194039 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74b5448f7f-zdrjg"] Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.119154 4950 generic.go:334] "Generic (PLEG): container finished" podID="7563774c-0a5c-4af9-802a-662f4e721f3c" containerID="da6d7c641e3ff94fdefa562e2753b88297237917dc8d3b63b84a609a7f7d2450" exitCode=0 Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.119480 4950 generic.go:334] "Generic (PLEG): container finished" podID="7563774c-0a5c-4af9-802a-662f4e721f3c" containerID="e609f6cce6ca82ac4a1ba760e29aadc780269f67316528034724eda1d330732d" exitCode=2 Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.119497 4950 generic.go:334] "Generic (PLEG): container finished" podID="7563774c-0a5c-4af9-802a-662f4e721f3c" containerID="de1b4d05b6433c8bacfa4f5808c26a249120f01056e842ff8ef61a2c85cdcfad" exitCode=0 Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.119229 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7563774c-0a5c-4af9-802a-662f4e721f3c","Type":"ContainerDied","Data":"da6d7c641e3ff94fdefa562e2753b88297237917dc8d3b63b84a609a7f7d2450"} Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.119589 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7563774c-0a5c-4af9-802a-662f4e721f3c","Type":"ContainerDied","Data":"e609f6cce6ca82ac4a1ba760e29aadc780269f67316528034724eda1d330732d"} Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.119605 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7563774c-0a5c-4af9-802a-662f4e721f3c","Type":"ContainerDied","Data":"de1b4d05b6433c8bacfa4f5808c26a249120f01056e842ff8ef61a2c85cdcfad"} Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.121485 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d5c6d4b4c-6cdzp" event={"ID":"2afba89a-006e-472f-8e49-0e8d69cd86c7","Type":"ContainerStarted","Data":"710823210f1a20088abd62720e0217f90ee4f975e4d072376c4fe0b935e0d064"} Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.121519 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d5c6d4b4c-6cdzp" event={"ID":"2afba89a-006e-472f-8e49-0e8d69cd86c7","Type":"ContainerStarted","Data":"27680c5ed5cb90c7f896bba09dfc3469387c0c378ee7efe74332861dc56601f8"} Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.121532 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d5c6d4b4c-6cdzp" event={"ID":"2afba89a-006e-472f-8e49-0e8d69cd86c7","Type":"ContainerStarted","Data":"eafaaf98694d4f051f242102d91ff820b0c23c70255102f98b68b532bf105d77"} Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.121582 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.146148 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-d5c6d4b4c-6cdzp" podStartSLOduration=7.146131149 podStartE2EDuration="7.146131149s" podCreationTimestamp="2026-03-18 20:26:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:26:33.142152536 +0000 UTC m=+1206.382994404" watchObservedRunningTime="2026-03-18 20:26:33.146131149 +0000 UTC m=+1206.386973017" Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.489518 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad591d5a-4120-4285-9cca-417b19176795" path="/var/lib/kubelet/pods/ad591d5a-4120-4285-9cca-417b19176795/volumes" Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.524845 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-brgnq" Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.605624 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wd8n\" (UniqueName: \"kubernetes.io/projected/82a7c2a4-9f57-460e-8bac-419f14222502-kube-api-access-9wd8n\") pod \"82a7c2a4-9f57-460e-8bac-419f14222502\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.605690 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-config-data\") pod \"82a7c2a4-9f57-460e-8bac-419f14222502\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.605713 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-scripts\") pod \"82a7c2a4-9f57-460e-8bac-419f14222502\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.605733 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-combined-ca-bundle\") pod \"82a7c2a4-9f57-460e-8bac-419f14222502\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.605749 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82a7c2a4-9f57-460e-8bac-419f14222502-etc-machine-id\") pod \"82a7c2a4-9f57-460e-8bac-419f14222502\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.605768 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-db-sync-config-data\") pod \"82a7c2a4-9f57-460e-8bac-419f14222502\" (UID: \"82a7c2a4-9f57-460e-8bac-419f14222502\") " Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.606394 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/82a7c2a4-9f57-460e-8bac-419f14222502-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "82a7c2a4-9f57-460e-8bac-419f14222502" (UID: "82a7c2a4-9f57-460e-8bac-419f14222502"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.614484 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82a7c2a4-9f57-460e-8bac-419f14222502-kube-api-access-9wd8n" (OuterVolumeSpecName: "kube-api-access-9wd8n") pod "82a7c2a4-9f57-460e-8bac-419f14222502" (UID: "82a7c2a4-9f57-460e-8bac-419f14222502"). InnerVolumeSpecName "kube-api-access-9wd8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.614856 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-scripts" (OuterVolumeSpecName: "scripts") pod "82a7c2a4-9f57-460e-8bac-419f14222502" (UID: "82a7c2a4-9f57-460e-8bac-419f14222502"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.633470 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "82a7c2a4-9f57-460e-8bac-419f14222502" (UID: "82a7c2a4-9f57-460e-8bac-419f14222502"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.682386 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-config-data" (OuterVolumeSpecName: "config-data") pod "82a7c2a4-9f57-460e-8bac-419f14222502" (UID: "82a7c2a4-9f57-460e-8bac-419f14222502"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.687496 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82a7c2a4-9f57-460e-8bac-419f14222502" (UID: "82a7c2a4-9f57-460e-8bac-419f14222502"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.707385 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wd8n\" (UniqueName: \"kubernetes.io/projected/82a7c2a4-9f57-460e-8bac-419f14222502-kube-api-access-9wd8n\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.707429 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.707439 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.707448 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.707457 4950 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82a7c2a4-9f57-460e-8bac-419f14222502-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:33 crc kubenswrapper[4950]: I0318 20:26:33.707467 4950 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82a7c2a4-9f57-460e-8bac-419f14222502-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.131292 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-brgnq" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.132524 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-brgnq" event={"ID":"82a7c2a4-9f57-460e-8bac-419f14222502","Type":"ContainerDied","Data":"97a26f08854fc0dfaabc62754c4b6d2acb95a35796f8ece839e9b7c0bd98db5a"} Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.132572 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97a26f08854fc0dfaabc62754c4b6d2acb95a35796f8ece839e9b7c0bd98db5a" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.548208 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 18 20:26:34 crc kubenswrapper[4950]: E0318 20:26:34.548632 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad591d5a-4120-4285-9cca-417b19176795" containerName="init" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.548653 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad591d5a-4120-4285-9cca-417b19176795" containerName="init" Mar 18 20:26:34 crc kubenswrapper[4950]: E0318 20:26:34.548666 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad591d5a-4120-4285-9cca-417b19176795" containerName="dnsmasq-dns" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.548676 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad591d5a-4120-4285-9cca-417b19176795" containerName="dnsmasq-dns" Mar 18 20:26:34 crc kubenswrapper[4950]: E0318 20:26:34.548691 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82a7c2a4-9f57-460e-8bac-419f14222502" containerName="cinder-db-sync" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.548698 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="82a7c2a4-9f57-460e-8bac-419f14222502" containerName="cinder-db-sync" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.548888 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="82a7c2a4-9f57-460e-8bac-419f14222502" containerName="cinder-db-sync" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.548911 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad591d5a-4120-4285-9cca-417b19176795" containerName="dnsmasq-dns" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.549848 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.552460 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.552641 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.552770 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.552966 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-dv945" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.583221 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.636511 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-config-data\") pod \"cinder-scheduler-0\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.636745 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-scripts\") pod \"cinder-scheduler-0\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.636775 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.636822 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpbnc\" (UniqueName: \"kubernetes.io/projected/eac8a754-52b8-471c-bce1-85c449f801ba-kube-api-access-vpbnc\") pod \"cinder-scheduler-0\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.636883 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.636907 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eac8a754-52b8-471c-bce1-85c449f801ba-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.656824 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6648966d55-b42h7"] Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.670004 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6648966d55-b42h7" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.689243 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6648966d55-b42h7"] Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.738338 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9zks\" (UniqueName: \"kubernetes.io/projected/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-kube-api-access-h9zks\") pod \"dnsmasq-dns-6648966d55-b42h7\" (UID: \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\") " pod="openstack/dnsmasq-dns-6648966d55-b42h7" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.738394 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.738443 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-ovsdbserver-nb\") pod \"dnsmasq-dns-6648966d55-b42h7\" (UID: \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\") " pod="openstack/dnsmasq-dns-6648966d55-b42h7" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.738465 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eac8a754-52b8-471c-bce1-85c449f801ba-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.738489 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-config-data\") pod \"cinder-scheduler-0\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.738526 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-dns-svc\") pod \"dnsmasq-dns-6648966d55-b42h7\" (UID: \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\") " pod="openstack/dnsmasq-dns-6648966d55-b42h7" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.738546 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-ovsdbserver-sb\") pod \"dnsmasq-dns-6648966d55-b42h7\" (UID: \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\") " pod="openstack/dnsmasq-dns-6648966d55-b42h7" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.738601 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-scripts\") pod \"cinder-scheduler-0\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.738621 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.738639 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-config\") pod \"dnsmasq-dns-6648966d55-b42h7\" (UID: \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\") " pod="openstack/dnsmasq-dns-6648966d55-b42h7" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.738664 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpbnc\" (UniqueName: \"kubernetes.io/projected/eac8a754-52b8-471c-bce1-85c449f801ba-kube-api-access-vpbnc\") pod \"cinder-scheduler-0\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.739455 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eac8a754-52b8-471c-bce1-85c449f801ba-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.766061 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.766998 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.767311 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-scripts\") pod \"cinder-scheduler-0\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.776508 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpbnc\" (UniqueName: \"kubernetes.io/projected/eac8a754-52b8-471c-bce1-85c449f801ba-kube-api-access-vpbnc\") pod \"cinder-scheduler-0\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.782716 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-config-data\") pod \"cinder-scheduler-0\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.839933 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-ovsdbserver-sb\") pod \"dnsmasq-dns-6648966d55-b42h7\" (UID: \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\") " pod="openstack/dnsmasq-dns-6648966d55-b42h7" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.840048 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-config\") pod \"dnsmasq-dns-6648966d55-b42h7\" (UID: \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\") " pod="openstack/dnsmasq-dns-6648966d55-b42h7" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.840108 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9zks\" (UniqueName: \"kubernetes.io/projected/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-kube-api-access-h9zks\") pod \"dnsmasq-dns-6648966d55-b42h7\" (UID: \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\") " pod="openstack/dnsmasq-dns-6648966d55-b42h7" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.840129 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-ovsdbserver-nb\") pod \"dnsmasq-dns-6648966d55-b42h7\" (UID: \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\") " pod="openstack/dnsmasq-dns-6648966d55-b42h7" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.840188 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-dns-svc\") pod \"dnsmasq-dns-6648966d55-b42h7\" (UID: \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\") " pod="openstack/dnsmasq-dns-6648966d55-b42h7" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.841209 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-dns-svc\") pod \"dnsmasq-dns-6648966d55-b42h7\" (UID: \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\") " pod="openstack/dnsmasq-dns-6648966d55-b42h7" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.841272 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-config\") pod \"dnsmasq-dns-6648966d55-b42h7\" (UID: \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\") " pod="openstack/dnsmasq-dns-6648966d55-b42h7" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.841925 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-ovsdbserver-sb\") pod \"dnsmasq-dns-6648966d55-b42h7\" (UID: \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\") " pod="openstack/dnsmasq-dns-6648966d55-b42h7" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.842070 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-ovsdbserver-nb\") pod \"dnsmasq-dns-6648966d55-b42h7\" (UID: \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\") " pod="openstack/dnsmasq-dns-6648966d55-b42h7" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.844278 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.848327 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.850716 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.865972 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.866716 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.868254 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9zks\" (UniqueName: \"kubernetes.io/projected/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-kube-api-access-h9zks\") pod \"dnsmasq-dns-6648966d55-b42h7\" (UID: \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\") " pod="openstack/dnsmasq-dns-6648966d55-b42h7" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.943286 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6ae778b5-53f6-4fbf-b779-83a194d4da7d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " pod="openstack/cinder-api-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.943354 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " pod="openstack/cinder-api-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.943377 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-scripts\") pod \"cinder-api-0\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " pod="openstack/cinder-api-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.943473 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-config-data\") pod \"cinder-api-0\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " pod="openstack/cinder-api-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.943640 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-config-data-custom\") pod \"cinder-api-0\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " pod="openstack/cinder-api-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.943841 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6k6r\" (UniqueName: \"kubernetes.io/projected/6ae778b5-53f6-4fbf-b779-83a194d4da7d-kube-api-access-h6k6r\") pod \"cinder-api-0\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " pod="openstack/cinder-api-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.943910 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ae778b5-53f6-4fbf-b779-83a194d4da7d-logs\") pod \"cinder-api-0\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " pod="openstack/cinder-api-0" Mar 18 20:26:34 crc kubenswrapper[4950]: I0318 20:26:34.993268 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6648966d55-b42h7" Mar 18 20:26:35 crc kubenswrapper[4950]: I0318 20:26:35.047352 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-config-data\") pod \"cinder-api-0\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " pod="openstack/cinder-api-0" Mar 18 20:26:35 crc kubenswrapper[4950]: I0318 20:26:35.047443 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-config-data-custom\") pod \"cinder-api-0\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " pod="openstack/cinder-api-0" Mar 18 20:26:35 crc kubenswrapper[4950]: I0318 20:26:35.047503 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6k6r\" (UniqueName: \"kubernetes.io/projected/6ae778b5-53f6-4fbf-b779-83a194d4da7d-kube-api-access-h6k6r\") pod \"cinder-api-0\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " pod="openstack/cinder-api-0" Mar 18 20:26:35 crc kubenswrapper[4950]: I0318 20:26:35.047542 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ae778b5-53f6-4fbf-b779-83a194d4da7d-logs\") pod \"cinder-api-0\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " pod="openstack/cinder-api-0" Mar 18 20:26:35 crc kubenswrapper[4950]: I0318 20:26:35.047590 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6ae778b5-53f6-4fbf-b779-83a194d4da7d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " pod="openstack/cinder-api-0" Mar 18 20:26:35 crc kubenswrapper[4950]: I0318 20:26:35.047621 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " pod="openstack/cinder-api-0" Mar 18 20:26:35 crc kubenswrapper[4950]: I0318 20:26:35.047640 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-scripts\") pod \"cinder-api-0\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " pod="openstack/cinder-api-0" Mar 18 20:26:35 crc kubenswrapper[4950]: I0318 20:26:35.048297 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6ae778b5-53f6-4fbf-b779-83a194d4da7d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " pod="openstack/cinder-api-0" Mar 18 20:26:35 crc kubenswrapper[4950]: I0318 20:26:35.048698 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ae778b5-53f6-4fbf-b779-83a194d4da7d-logs\") pod \"cinder-api-0\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " pod="openstack/cinder-api-0" Mar 18 20:26:35 crc kubenswrapper[4950]: I0318 20:26:35.052991 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-config-data-custom\") pod \"cinder-api-0\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " pod="openstack/cinder-api-0" Mar 18 20:26:35 crc kubenswrapper[4950]: I0318 20:26:35.054178 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " pod="openstack/cinder-api-0" Mar 18 20:26:35 crc kubenswrapper[4950]: I0318 20:26:35.054458 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-scripts\") pod \"cinder-api-0\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " pod="openstack/cinder-api-0" Mar 18 20:26:35 crc kubenswrapper[4950]: I0318 20:26:35.054986 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-config-data\") pod \"cinder-api-0\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " pod="openstack/cinder-api-0" Mar 18 20:26:35 crc kubenswrapper[4950]: I0318 20:26:35.070964 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6k6r\" (UniqueName: \"kubernetes.io/projected/6ae778b5-53f6-4fbf-b779-83a194d4da7d-kube-api-access-h6k6r\") pod \"cinder-api-0\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " pod="openstack/cinder-api-0" Mar 18 20:26:35 crc kubenswrapper[4950]: I0318 20:26:35.275273 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 18 20:26:35 crc kubenswrapper[4950]: I0318 20:26:35.410790 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 18 20:26:35 crc kubenswrapper[4950]: I0318 20:26:35.442418 4950 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 20:26:35 crc kubenswrapper[4950]: I0318 20:26:35.582628 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6648966d55-b42h7"] Mar 18 20:26:35 crc kubenswrapper[4950]: I0318 20:26:35.826925 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.011037 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.065953 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-combined-ca-bundle\") pod \"7563774c-0a5c-4af9-802a-662f4e721f3c\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.065993 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-config-data\") pod \"7563774c-0a5c-4af9-802a-662f4e721f3c\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.066060 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7563774c-0a5c-4af9-802a-662f4e721f3c-run-httpd\") pod \"7563774c-0a5c-4af9-802a-662f4e721f3c\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.066097 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-scripts\") pod \"7563774c-0a5c-4af9-802a-662f4e721f3c\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.066162 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvmgk\" (UniqueName: \"kubernetes.io/projected/7563774c-0a5c-4af9-802a-662f4e721f3c-kube-api-access-pvmgk\") pod \"7563774c-0a5c-4af9-802a-662f4e721f3c\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.066216 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7563774c-0a5c-4af9-802a-662f4e721f3c-log-httpd\") pod \"7563774c-0a5c-4af9-802a-662f4e721f3c\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.066283 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-sg-core-conf-yaml\") pod \"7563774c-0a5c-4af9-802a-662f4e721f3c\" (UID: \"7563774c-0a5c-4af9-802a-662f4e721f3c\") " Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.069883 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7563774c-0a5c-4af9-802a-662f4e721f3c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7563774c-0a5c-4af9-802a-662f4e721f3c" (UID: "7563774c-0a5c-4af9-802a-662f4e721f3c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.070091 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7563774c-0a5c-4af9-802a-662f4e721f3c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7563774c-0a5c-4af9-802a-662f4e721f3c" (UID: "7563774c-0a5c-4af9-802a-662f4e721f3c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.087793 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-scripts" (OuterVolumeSpecName: "scripts") pod "7563774c-0a5c-4af9-802a-662f4e721f3c" (UID: "7563774c-0a5c-4af9-802a-662f4e721f3c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.094376 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7563774c-0a5c-4af9-802a-662f4e721f3c-kube-api-access-pvmgk" (OuterVolumeSpecName: "kube-api-access-pvmgk") pod "7563774c-0a5c-4af9-802a-662f4e721f3c" (UID: "7563774c-0a5c-4af9-802a-662f4e721f3c"). InnerVolumeSpecName "kube-api-access-pvmgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.124701 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7563774c-0a5c-4af9-802a-662f4e721f3c" (UID: "7563774c-0a5c-4af9-802a-662f4e721f3c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.167834 4950 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7563774c-0a5c-4af9-802a-662f4e721f3c-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.167861 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.167869 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvmgk\" (UniqueName: \"kubernetes.io/projected/7563774c-0a5c-4af9-802a-662f4e721f3c-kube-api-access-pvmgk\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.167880 4950 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7563774c-0a5c-4af9-802a-662f4e721f3c-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.167889 4950 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.180005 4950 generic.go:334] "Generic (PLEG): container finished" podID="292b3f8d-075a-4d1d-8aa9-56dd2c00943c" containerID="92cb7ef93cf371a845543fb38f5e1870998e9149cf186cce812098557089a974" exitCode=0 Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.180064 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6648966d55-b42h7" event={"ID":"292b3f8d-075a-4d1d-8aa9-56dd2c00943c","Type":"ContainerDied","Data":"92cb7ef93cf371a845543fb38f5e1870998e9149cf186cce812098557089a974"} Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.180088 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6648966d55-b42h7" event={"ID":"292b3f8d-075a-4d1d-8aa9-56dd2c00943c","Type":"ContainerStarted","Data":"927554ad0952af8811296c017ee1311035e9da9c03bfd58a0c23e295a7fb5f38"} Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.181298 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7563774c-0a5c-4af9-802a-662f4e721f3c" (UID: "7563774c-0a5c-4af9-802a-662f4e721f3c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.221259 4950 generic.go:334] "Generic (PLEG): container finished" podID="7563774c-0a5c-4af9-802a-662f4e721f3c" containerID="a94ea562bd7c3e520d36e8bc6b0f62fe4b690bcbac96af7089b92400bea93599" exitCode=0 Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.221345 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7563774c-0a5c-4af9-802a-662f4e721f3c","Type":"ContainerDied","Data":"a94ea562bd7c3e520d36e8bc6b0f62fe4b690bcbac96af7089b92400bea93599"} Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.221377 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7563774c-0a5c-4af9-802a-662f4e721f3c","Type":"ContainerDied","Data":"7cbf38b3da364f13e1adf82ec0026977bf82c9432979f7f2f5fb3080546ffc16"} Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.221394 4950 scope.go:117] "RemoveContainer" containerID="da6d7c641e3ff94fdefa562e2753b88297237917dc8d3b63b84a609a7f7d2450" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.221550 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.263623 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6ae778b5-53f6-4fbf-b779-83a194d4da7d","Type":"ContainerStarted","Data":"1b834b5c5abb7f4f3e4bb8f034fe0ececf866881271852ca8eb09119b789ab2d"} Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.269495 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.284582 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-config-data" (OuterVolumeSpecName: "config-data") pod "7563774c-0a5c-4af9-802a-662f4e721f3c" (UID: "7563774c-0a5c-4af9-802a-662f4e721f3c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.285797 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"eac8a754-52b8-471c-bce1-85c449f801ba","Type":"ContainerStarted","Data":"d3ccce4ca8ca5f25f58a935906e075ef2e4680759dd54bb310fc16bd73ae07ca"} Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.374345 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7563774c-0a5c-4af9-802a-662f4e721f3c-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.508593 4950 scope.go:117] "RemoveContainer" containerID="e609f6cce6ca82ac4a1ba760e29aadc780269f67316528034724eda1d330732d" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.578246 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.593065 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.618202 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:26:36 crc kubenswrapper[4950]: E0318 20:26:36.618536 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7563774c-0a5c-4af9-802a-662f4e721f3c" containerName="proxy-httpd" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.618551 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="7563774c-0a5c-4af9-802a-662f4e721f3c" containerName="proxy-httpd" Mar 18 20:26:36 crc kubenswrapper[4950]: E0318 20:26:36.618575 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7563774c-0a5c-4af9-802a-662f4e721f3c" containerName="ceilometer-central-agent" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.618581 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="7563774c-0a5c-4af9-802a-662f4e721f3c" containerName="ceilometer-central-agent" Mar 18 20:26:36 crc kubenswrapper[4950]: E0318 20:26:36.618595 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7563774c-0a5c-4af9-802a-662f4e721f3c" containerName="sg-core" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.618601 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="7563774c-0a5c-4af9-802a-662f4e721f3c" containerName="sg-core" Mar 18 20:26:36 crc kubenswrapper[4950]: E0318 20:26:36.618616 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7563774c-0a5c-4af9-802a-662f4e721f3c" containerName="ceilometer-notification-agent" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.618623 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="7563774c-0a5c-4af9-802a-662f4e721f3c" containerName="ceilometer-notification-agent" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.618746 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="7563774c-0a5c-4af9-802a-662f4e721f3c" containerName="sg-core" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.618759 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="7563774c-0a5c-4af9-802a-662f4e721f3c" containerName="proxy-httpd" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.618768 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="7563774c-0a5c-4af9-802a-662f4e721f3c" containerName="ceilometer-central-agent" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.618775 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="7563774c-0a5c-4af9-802a-662f4e721f3c" containerName="ceilometer-notification-agent" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.620095 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.622861 4950 scope.go:117] "RemoveContainer" containerID="a94ea562bd7c3e520d36e8bc6b0f62fe4b690bcbac96af7089b92400bea93599" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.628572 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.628622 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.665513 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.679573 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-config-data\") pod \"ceilometer-0\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.679610 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-scripts\") pod \"ceilometer-0\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.679680 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-run-httpd\") pod \"ceilometer-0\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.679708 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sv7h\" (UniqueName: \"kubernetes.io/projected/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-kube-api-access-6sv7h\") pod \"ceilometer-0\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.679787 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.679843 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.679872 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-log-httpd\") pod \"ceilometer-0\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.728592 4950 scope.go:117] "RemoveContainer" containerID="de1b4d05b6433c8bacfa4f5808c26a249120f01056e842ff8ef61a2c85cdcfad" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.781484 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-config-data\") pod \"ceilometer-0\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.781531 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-scripts\") pod \"ceilometer-0\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.781584 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-run-httpd\") pod \"ceilometer-0\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.781609 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sv7h\" (UniqueName: \"kubernetes.io/projected/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-kube-api-access-6sv7h\") pod \"ceilometer-0\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.781658 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.781689 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.781713 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-log-httpd\") pod \"ceilometer-0\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.782269 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-log-httpd\") pod \"ceilometer-0\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.782530 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-run-httpd\") pod \"ceilometer-0\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.801345 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-config-data\") pod \"ceilometer-0\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.801754 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.824388 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.825691 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-scripts\") pod \"ceilometer-0\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.846371 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sv7h\" (UniqueName: \"kubernetes.io/projected/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-kube-api-access-6sv7h\") pod \"ceilometer-0\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " pod="openstack/ceilometer-0" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.857393 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.952243 4950 scope.go:117] "RemoveContainer" containerID="da6d7c641e3ff94fdefa562e2753b88297237917dc8d3b63b84a609a7f7d2450" Mar 18 20:26:36 crc kubenswrapper[4950]: E0318 20:26:36.954438 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da6d7c641e3ff94fdefa562e2753b88297237917dc8d3b63b84a609a7f7d2450\": container with ID starting with da6d7c641e3ff94fdefa562e2753b88297237917dc8d3b63b84a609a7f7d2450 not found: ID does not exist" containerID="da6d7c641e3ff94fdefa562e2753b88297237917dc8d3b63b84a609a7f7d2450" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.954473 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da6d7c641e3ff94fdefa562e2753b88297237917dc8d3b63b84a609a7f7d2450"} err="failed to get container status \"da6d7c641e3ff94fdefa562e2753b88297237917dc8d3b63b84a609a7f7d2450\": rpc error: code = NotFound desc = could not find container \"da6d7c641e3ff94fdefa562e2753b88297237917dc8d3b63b84a609a7f7d2450\": container with ID starting with da6d7c641e3ff94fdefa562e2753b88297237917dc8d3b63b84a609a7f7d2450 not found: ID does not exist" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.954495 4950 scope.go:117] "RemoveContainer" containerID="e609f6cce6ca82ac4a1ba760e29aadc780269f67316528034724eda1d330732d" Mar 18 20:26:36 crc kubenswrapper[4950]: E0318 20:26:36.954723 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e609f6cce6ca82ac4a1ba760e29aadc780269f67316528034724eda1d330732d\": container with ID starting with e609f6cce6ca82ac4a1ba760e29aadc780269f67316528034724eda1d330732d not found: ID does not exist" containerID="e609f6cce6ca82ac4a1ba760e29aadc780269f67316528034724eda1d330732d" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.954743 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e609f6cce6ca82ac4a1ba760e29aadc780269f67316528034724eda1d330732d"} err="failed to get container status \"e609f6cce6ca82ac4a1ba760e29aadc780269f67316528034724eda1d330732d\": rpc error: code = NotFound desc = could not find container \"e609f6cce6ca82ac4a1ba760e29aadc780269f67316528034724eda1d330732d\": container with ID starting with e609f6cce6ca82ac4a1ba760e29aadc780269f67316528034724eda1d330732d not found: ID does not exist" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.954755 4950 scope.go:117] "RemoveContainer" containerID="a94ea562bd7c3e520d36e8bc6b0f62fe4b690bcbac96af7089b92400bea93599" Mar 18 20:26:36 crc kubenswrapper[4950]: E0318 20:26:36.954957 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a94ea562bd7c3e520d36e8bc6b0f62fe4b690bcbac96af7089b92400bea93599\": container with ID starting with a94ea562bd7c3e520d36e8bc6b0f62fe4b690bcbac96af7089b92400bea93599 not found: ID does not exist" containerID="a94ea562bd7c3e520d36e8bc6b0f62fe4b690bcbac96af7089b92400bea93599" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.954978 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a94ea562bd7c3e520d36e8bc6b0f62fe4b690bcbac96af7089b92400bea93599"} err="failed to get container status \"a94ea562bd7c3e520d36e8bc6b0f62fe4b690bcbac96af7089b92400bea93599\": rpc error: code = NotFound desc = could not find container \"a94ea562bd7c3e520d36e8bc6b0f62fe4b690bcbac96af7089b92400bea93599\": container with ID starting with a94ea562bd7c3e520d36e8bc6b0f62fe4b690bcbac96af7089b92400bea93599 not found: ID does not exist" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.954990 4950 scope.go:117] "RemoveContainer" containerID="de1b4d05b6433c8bacfa4f5808c26a249120f01056e842ff8ef61a2c85cdcfad" Mar 18 20:26:36 crc kubenswrapper[4950]: E0318 20:26:36.955204 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de1b4d05b6433c8bacfa4f5808c26a249120f01056e842ff8ef61a2c85cdcfad\": container with ID starting with de1b4d05b6433c8bacfa4f5808c26a249120f01056e842ff8ef61a2c85cdcfad not found: ID does not exist" containerID="de1b4d05b6433c8bacfa4f5808c26a249120f01056e842ff8ef61a2c85cdcfad" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.955222 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de1b4d05b6433c8bacfa4f5808c26a249120f01056e842ff8ef61a2c85cdcfad"} err="failed to get container status \"de1b4d05b6433c8bacfa4f5808c26a249120f01056e842ff8ef61a2c85cdcfad\": rpc error: code = NotFound desc = could not find container \"de1b4d05b6433c8bacfa4f5808c26a249120f01056e842ff8ef61a2c85cdcfad\": container with ID starting with de1b4d05b6433c8bacfa4f5808c26a249120f01056e842ff8ef61a2c85cdcfad not found: ID does not exist" Mar 18 20:26:36 crc kubenswrapper[4950]: I0318 20:26:36.958690 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:26:37 crc kubenswrapper[4950]: I0318 20:26:37.333751 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6ae778b5-53f6-4fbf-b779-83a194d4da7d","Type":"ContainerStarted","Data":"a1291c01d23d49a5770ca05d1127cad33c511df6e4bffd22c195d2178762aebb"} Mar 18 20:26:37 crc kubenswrapper[4950]: I0318 20:26:37.348789 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6648966d55-b42h7" event={"ID":"292b3f8d-075a-4d1d-8aa9-56dd2c00943c","Type":"ContainerStarted","Data":"9e39560939f2c6a63b846fe7631faa7dc91c92edf7fece72891f4dfc17d751a2"} Mar 18 20:26:37 crc kubenswrapper[4950]: I0318 20:26:37.349257 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6648966d55-b42h7" Mar 18 20:26:37 crc kubenswrapper[4950]: I0318 20:26:37.379916 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6648966d55-b42h7" podStartSLOduration=3.379899349 podStartE2EDuration="3.379899349s" podCreationTimestamp="2026-03-18 20:26:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:26:37.379432657 +0000 UTC m=+1210.620274525" watchObservedRunningTime="2026-03-18 20:26:37.379899349 +0000 UTC m=+1210.620741207" Mar 18 20:26:37 crc kubenswrapper[4950]: I0318 20:26:37.504619 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7563774c-0a5c-4af9-802a-662f4e721f3c" path="/var/lib/kubelet/pods/7563774c-0a5c-4af9-802a-662f4e721f3c/volumes" Mar 18 20:26:37 crc kubenswrapper[4950]: I0318 20:26:37.552709 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:26:38 crc kubenswrapper[4950]: I0318 20:26:38.359832 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a00a076-ca9d-46e9-b3e7-af13dcaca76e","Type":"ContainerStarted","Data":"15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca"} Mar 18 20:26:38 crc kubenswrapper[4950]: I0318 20:26:38.360145 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a00a076-ca9d-46e9-b3e7-af13dcaca76e","Type":"ContainerStarted","Data":"de71a314bd6be6cf438243681b817c44b5906f260c505a93a236ea484a733d7e"} Mar 18 20:26:38 crc kubenswrapper[4950]: I0318 20:26:38.361744 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"eac8a754-52b8-471c-bce1-85c449f801ba","Type":"ContainerStarted","Data":"42d560c87b7d449e8e9b9bb91455358f6f1c39aa6f8e2cb7378e5b8ac837ccfa"} Mar 18 20:26:38 crc kubenswrapper[4950]: I0318 20:26:38.363137 4950 generic.go:334] "Generic (PLEG): container finished" podID="a4bb9e1a-0694-4e00-8ba6-f2765f82b983" containerID="de9ab05a4141d7f72e470b7a98a89d526812c15d9a4174be30b8a51cc0d0d4b3" exitCode=0 Mar 18 20:26:38 crc kubenswrapper[4950]: I0318 20:26:38.363185 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-tpdmj" event={"ID":"a4bb9e1a-0694-4e00-8ba6-f2765f82b983","Type":"ContainerDied","Data":"de9ab05a4141d7f72e470b7a98a89d526812c15d9a4174be30b8a51cc0d0d4b3"} Mar 18 20:26:38 crc kubenswrapper[4950]: I0318 20:26:38.375460 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6ae778b5-53f6-4fbf-b779-83a194d4da7d","Type":"ContainerStarted","Data":"9b8971bb26768fa3b1732ed9bb879b77e747608497828e2f327f5cbfc179b4fc"} Mar 18 20:26:38 crc kubenswrapper[4950]: I0318 20:26:38.375494 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="6ae778b5-53f6-4fbf-b779-83a194d4da7d" containerName="cinder-api-log" containerID="cri-o://a1291c01d23d49a5770ca05d1127cad33c511df6e4bffd22c195d2178762aebb" gracePeriod=30 Mar 18 20:26:38 crc kubenswrapper[4950]: I0318 20:26:38.375531 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 18 20:26:38 crc kubenswrapper[4950]: I0318 20:26:38.375561 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="6ae778b5-53f6-4fbf-b779-83a194d4da7d" containerName="cinder-api" containerID="cri-o://9b8971bb26768fa3b1732ed9bb879b77e747608497828e2f327f5cbfc179b4fc" gracePeriod=30 Mar 18 20:26:38 crc kubenswrapper[4950]: I0318 20:26:38.425910 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.42589205 podStartE2EDuration="4.42589205s" podCreationTimestamp="2026-03-18 20:26:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:26:38.423466378 +0000 UTC m=+1211.664308246" watchObservedRunningTime="2026-03-18 20:26:38.42589205 +0000 UTC m=+1211.666733918" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.388934 4950 generic.go:334] "Generic (PLEG): container finished" podID="6ae778b5-53f6-4fbf-b779-83a194d4da7d" containerID="9b8971bb26768fa3b1732ed9bb879b77e747608497828e2f327f5cbfc179b4fc" exitCode=0 Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.389221 4950 generic.go:334] "Generic (PLEG): container finished" podID="6ae778b5-53f6-4fbf-b779-83a194d4da7d" containerID="a1291c01d23d49a5770ca05d1127cad33c511df6e4bffd22c195d2178762aebb" exitCode=143 Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.389317 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6ae778b5-53f6-4fbf-b779-83a194d4da7d","Type":"ContainerDied","Data":"9b8971bb26768fa3b1732ed9bb879b77e747608497828e2f327f5cbfc179b4fc"} Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.389343 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6ae778b5-53f6-4fbf-b779-83a194d4da7d","Type":"ContainerDied","Data":"a1291c01d23d49a5770ca05d1127cad33c511df6e4bffd22c195d2178762aebb"} Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.389353 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6ae778b5-53f6-4fbf-b779-83a194d4da7d","Type":"ContainerDied","Data":"1b834b5c5abb7f4f3e4bb8f034fe0ececf866881271852ca8eb09119b789ab2d"} Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.389362 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b834b5c5abb7f4f3e4bb8f034fe0ececf866881271852ca8eb09119b789ab2d" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.395821 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"eac8a754-52b8-471c-bce1-85c449f801ba","Type":"ContainerStarted","Data":"c956a40169bc02e7c0273be7eebc398074072389554d2646b06f1a2f18bd2b83"} Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.399975 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a00a076-ca9d-46e9-b3e7-af13dcaca76e","Type":"ContainerStarted","Data":"c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc"} Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.423604 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.433781 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.210149085 podStartE2EDuration="5.433767831s" podCreationTimestamp="2026-03-18 20:26:34 +0000 UTC" firstStartedPulling="2026-03-18 20:26:35.442175356 +0000 UTC m=+1208.683017224" lastFinishedPulling="2026-03-18 20:26:36.665794102 +0000 UTC m=+1209.906635970" observedRunningTime="2026-03-18 20:26:39.433437842 +0000 UTC m=+1212.674279710" watchObservedRunningTime="2026-03-18 20:26:39.433767831 +0000 UTC m=+1212.674609699" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.554763 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-combined-ca-bundle\") pod \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.554817 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6k6r\" (UniqueName: \"kubernetes.io/projected/6ae778b5-53f6-4fbf-b779-83a194d4da7d-kube-api-access-h6k6r\") pod \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.554984 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-config-data-custom\") pod \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.555091 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-config-data\") pod \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.555158 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6ae778b5-53f6-4fbf-b779-83a194d4da7d-etc-machine-id\") pod \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.556001 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ae778b5-53f6-4fbf-b779-83a194d4da7d-logs\") pod \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.556065 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-scripts\") pod \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\" (UID: \"6ae778b5-53f6-4fbf-b779-83a194d4da7d\") " Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.558497 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ae778b5-53f6-4fbf-b779-83a194d4da7d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6ae778b5-53f6-4fbf-b779-83a194d4da7d" (UID: "6ae778b5-53f6-4fbf-b779-83a194d4da7d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.560403 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6ae778b5-53f6-4fbf-b779-83a194d4da7d" (UID: "6ae778b5-53f6-4fbf-b779-83a194d4da7d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.561071 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ae778b5-53f6-4fbf-b779-83a194d4da7d-kube-api-access-h6k6r" (OuterVolumeSpecName: "kube-api-access-h6k6r") pod "6ae778b5-53f6-4fbf-b779-83a194d4da7d" (UID: "6ae778b5-53f6-4fbf-b779-83a194d4da7d"). InnerVolumeSpecName "kube-api-access-h6k6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.561100 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ae778b5-53f6-4fbf-b779-83a194d4da7d-logs" (OuterVolumeSpecName: "logs") pod "6ae778b5-53f6-4fbf-b779-83a194d4da7d" (UID: "6ae778b5-53f6-4fbf-b779-83a194d4da7d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.561441 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-scripts" (OuterVolumeSpecName: "scripts") pod "6ae778b5-53f6-4fbf-b779-83a194d4da7d" (UID: "6ae778b5-53f6-4fbf-b779-83a194d4da7d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.600630 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ae778b5-53f6-4fbf-b779-83a194d4da7d" (UID: "6ae778b5-53f6-4fbf-b779-83a194d4da7d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.602566 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-config-data" (OuterVolumeSpecName: "config-data") pod "6ae778b5-53f6-4fbf-b779-83a194d4da7d" (UID: "6ae778b5-53f6-4fbf-b779-83a194d4da7d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.658156 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.658186 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.658197 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6k6r\" (UniqueName: \"kubernetes.io/projected/6ae778b5-53f6-4fbf-b779-83a194d4da7d-kube-api-access-h6k6r\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.658207 4950 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.658215 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ae778b5-53f6-4fbf-b779-83a194d4da7d-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.658224 4950 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6ae778b5-53f6-4fbf-b779-83a194d4da7d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.658233 4950 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ae778b5-53f6-4fbf-b779-83a194d4da7d-logs\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.734160 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-tpdmj" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.862968 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a4bb9e1a-0694-4e00-8ba6-f2765f82b983-db-sync-config-data\") pod \"a4bb9e1a-0694-4e00-8ba6-f2765f82b983\" (UID: \"a4bb9e1a-0694-4e00-8ba6-f2765f82b983\") " Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.863089 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4bb9e1a-0694-4e00-8ba6-f2765f82b983-combined-ca-bundle\") pod \"a4bb9e1a-0694-4e00-8ba6-f2765f82b983\" (UID: \"a4bb9e1a-0694-4e00-8ba6-f2765f82b983\") " Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.863173 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tjdf\" (UniqueName: \"kubernetes.io/projected/a4bb9e1a-0694-4e00-8ba6-f2765f82b983-kube-api-access-5tjdf\") pod \"a4bb9e1a-0694-4e00-8ba6-f2765f82b983\" (UID: \"a4bb9e1a-0694-4e00-8ba6-f2765f82b983\") " Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.869507 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4bb9e1a-0694-4e00-8ba6-f2765f82b983-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a4bb9e1a-0694-4e00-8ba6-f2765f82b983" (UID: "a4bb9e1a-0694-4e00-8ba6-f2765f82b983"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.869885 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.870978 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4bb9e1a-0694-4e00-8ba6-f2765f82b983-kube-api-access-5tjdf" (OuterVolumeSpecName: "kube-api-access-5tjdf") pod "a4bb9e1a-0694-4e00-8ba6-f2765f82b983" (UID: "a4bb9e1a-0694-4e00-8ba6-f2765f82b983"). InnerVolumeSpecName "kube-api-access-5tjdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.898590 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4bb9e1a-0694-4e00-8ba6-f2765f82b983-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4bb9e1a-0694-4e00-8ba6-f2765f82b983" (UID: "a4bb9e1a-0694-4e00-8ba6-f2765f82b983"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.965295 4950 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a4bb9e1a-0694-4e00-8ba6-f2765f82b983-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.965334 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4bb9e1a-0694-4e00-8ba6-f2765f82b983-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:39 crc kubenswrapper[4950]: I0318 20:26:39.965344 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tjdf\" (UniqueName: \"kubernetes.io/projected/a4bb9e1a-0694-4e00-8ba6-f2765f82b983-kube-api-access-5tjdf\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.408956 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-tpdmj" event={"ID":"a4bb9e1a-0694-4e00-8ba6-f2765f82b983","Type":"ContainerDied","Data":"5d841f1711b2d20eb9beeb857c1703a9d8c135fead221b2e15a23a8700b18a1a"} Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.410165 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d841f1711b2d20eb9beeb857c1703a9d8c135fead221b2e15a23a8700b18a1a" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.408990 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-tpdmj" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.411084 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a00a076-ca9d-46e9-b3e7-af13dcaca76e","Type":"ContainerStarted","Data":"18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508"} Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.411134 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.446161 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.457660 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.473331 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 18 20:26:40 crc kubenswrapper[4950]: E0318 20:26:40.473684 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4bb9e1a-0694-4e00-8ba6-f2765f82b983" containerName="barbican-db-sync" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.473703 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4bb9e1a-0694-4e00-8ba6-f2765f82b983" containerName="barbican-db-sync" Mar 18 20:26:40 crc kubenswrapper[4950]: E0318 20:26:40.473715 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ae778b5-53f6-4fbf-b779-83a194d4da7d" containerName="cinder-api-log" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.473722 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ae778b5-53f6-4fbf-b779-83a194d4da7d" containerName="cinder-api-log" Mar 18 20:26:40 crc kubenswrapper[4950]: E0318 20:26:40.473748 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ae778b5-53f6-4fbf-b779-83a194d4da7d" containerName="cinder-api" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.473754 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ae778b5-53f6-4fbf-b779-83a194d4da7d" containerName="cinder-api" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.473900 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ae778b5-53f6-4fbf-b779-83a194d4da7d" containerName="cinder-api" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.473918 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4bb9e1a-0694-4e00-8ba6-f2765f82b983" containerName="barbican-db-sync" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.473935 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ae778b5-53f6-4fbf-b779-83a194d4da7d" containerName="cinder-api-log" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.474746 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.476137 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.476581 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.479525 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.486398 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.572982 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bae85ae-0545-46ac-90fa-c8f967648ed4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.573256 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bae85ae-0545-46ac-90fa-c8f967648ed4-logs\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.573425 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnsdj\" (UniqueName: \"kubernetes.io/projected/1bae85ae-0545-46ac-90fa-c8f967648ed4-kube-api-access-bnsdj\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.573584 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bae85ae-0545-46ac-90fa-c8f967648ed4-scripts\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.573844 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bae85ae-0545-46ac-90fa-c8f967648ed4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.573927 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bae85ae-0545-46ac-90fa-c8f967648ed4-config-data\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.574524 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1bae85ae-0545-46ac-90fa-c8f967648ed4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.574594 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bae85ae-0545-46ac-90fa-c8f967648ed4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.574839 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1bae85ae-0545-46ac-90fa-c8f967648ed4-config-data-custom\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.676969 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bae85ae-0545-46ac-90fa-c8f967648ed4-scripts\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.677050 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bae85ae-0545-46ac-90fa-c8f967648ed4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.677077 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bae85ae-0545-46ac-90fa-c8f967648ed4-config-data\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.677093 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1bae85ae-0545-46ac-90fa-c8f967648ed4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.677116 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bae85ae-0545-46ac-90fa-c8f967648ed4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.677139 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1bae85ae-0545-46ac-90fa-c8f967648ed4-config-data-custom\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.677164 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bae85ae-0545-46ac-90fa-c8f967648ed4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.677179 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bae85ae-0545-46ac-90fa-c8f967648ed4-logs\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.677209 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1bae85ae-0545-46ac-90fa-c8f967648ed4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.677213 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnsdj\" (UniqueName: \"kubernetes.io/projected/1bae85ae-0545-46ac-90fa-c8f967648ed4-kube-api-access-bnsdj\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.677925 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bae85ae-0545-46ac-90fa-c8f967648ed4-logs\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.686076 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bae85ae-0545-46ac-90fa-c8f967648ed4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.686734 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1bae85ae-0545-46ac-90fa-c8f967648ed4-config-data-custom\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.687300 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bae85ae-0545-46ac-90fa-c8f967648ed4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.691056 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bae85ae-0545-46ac-90fa-c8f967648ed4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.693509 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-6c48df686c-lz5gl"] Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.700293 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6c48df686c-lz5gl" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.720366 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bae85ae-0545-46ac-90fa-c8f967648ed4-config-data\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.721261 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bae85ae-0545-46ac-90fa-c8f967648ed4-scripts\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.721764 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.721970 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-qsjx9" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.722224 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.745246 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6c48df686c-lz5gl"] Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.772041 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnsdj\" (UniqueName: \"kubernetes.io/projected/1bae85ae-0545-46ac-90fa-c8f967648ed4-kube-api-access-bnsdj\") pod \"cinder-api-0\" (UID: \"1bae85ae-0545-46ac-90fa-c8f967648ed4\") " pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.778490 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n27xr\" (UniqueName: \"kubernetes.io/projected/f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9-kube-api-access-n27xr\") pod \"barbican-worker-6c48df686c-lz5gl\" (UID: \"f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9\") " pod="openstack/barbican-worker-6c48df686c-lz5gl" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.778544 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9-logs\") pod \"barbican-worker-6c48df686c-lz5gl\" (UID: \"f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9\") " pod="openstack/barbican-worker-6c48df686c-lz5gl" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.778573 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9-config-data\") pod \"barbican-worker-6c48df686c-lz5gl\" (UID: \"f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9\") " pod="openstack/barbican-worker-6c48df686c-lz5gl" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.778614 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9-combined-ca-bundle\") pod \"barbican-worker-6c48df686c-lz5gl\" (UID: \"f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9\") " pod="openstack/barbican-worker-6c48df686c-lz5gl" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.778663 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9-config-data-custom\") pod \"barbican-worker-6c48df686c-lz5gl\" (UID: \"f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9\") " pod="openstack/barbican-worker-6c48df686c-lz5gl" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.782087 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-546478654b-mwfqm"] Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.783517 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-546478654b-mwfqm" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.806106 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.806714 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.834895 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-546478654b-mwfqm"] Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.898911 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n27xr\" (UniqueName: \"kubernetes.io/projected/f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9-kube-api-access-n27xr\") pod \"barbican-worker-6c48df686c-lz5gl\" (UID: \"f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9\") " pod="openstack/barbican-worker-6c48df686c-lz5gl" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.899125 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4fw2\" (UniqueName: \"kubernetes.io/projected/72fe734a-54a6-4391-a125-04ddd4c9c890-kube-api-access-r4fw2\") pod \"barbican-keystone-listener-546478654b-mwfqm\" (UID: \"72fe734a-54a6-4391-a125-04ddd4c9c890\") " pod="openstack/barbican-keystone-listener-546478654b-mwfqm" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.899205 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9-logs\") pod \"barbican-worker-6c48df686c-lz5gl\" (UID: \"f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9\") " pod="openstack/barbican-worker-6c48df686c-lz5gl" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.899292 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9-config-data\") pod \"barbican-worker-6c48df686c-lz5gl\" (UID: \"f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9\") " pod="openstack/barbican-worker-6c48df686c-lz5gl" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.899387 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72fe734a-54a6-4391-a125-04ddd4c9c890-combined-ca-bundle\") pod \"barbican-keystone-listener-546478654b-mwfqm\" (UID: \"72fe734a-54a6-4391-a125-04ddd4c9c890\") " pod="openstack/barbican-keystone-listener-546478654b-mwfqm" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.899502 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9-combined-ca-bundle\") pod \"barbican-worker-6c48df686c-lz5gl\" (UID: \"f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9\") " pod="openstack/barbican-worker-6c48df686c-lz5gl" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.899582 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/72fe734a-54a6-4391-a125-04ddd4c9c890-config-data-custom\") pod \"barbican-keystone-listener-546478654b-mwfqm\" (UID: \"72fe734a-54a6-4391-a125-04ddd4c9c890\") " pod="openstack/barbican-keystone-listener-546478654b-mwfqm" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.899684 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9-config-data-custom\") pod \"barbican-worker-6c48df686c-lz5gl\" (UID: \"f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9\") " pod="openstack/barbican-worker-6c48df686c-lz5gl" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.899772 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72fe734a-54a6-4391-a125-04ddd4c9c890-config-data\") pod \"barbican-keystone-listener-546478654b-mwfqm\" (UID: \"72fe734a-54a6-4391-a125-04ddd4c9c890\") " pod="openstack/barbican-keystone-listener-546478654b-mwfqm" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.899872 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72fe734a-54a6-4391-a125-04ddd4c9c890-logs\") pod \"barbican-keystone-listener-546478654b-mwfqm\" (UID: \"72fe734a-54a6-4391-a125-04ddd4c9c890\") " pod="openstack/barbican-keystone-listener-546478654b-mwfqm" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.900743 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9-logs\") pod \"barbican-worker-6c48df686c-lz5gl\" (UID: \"f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9\") " pod="openstack/barbican-worker-6c48df686c-lz5gl" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.905267 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6648966d55-b42h7"] Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.905497 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6648966d55-b42h7" podUID="292b3f8d-075a-4d1d-8aa9-56dd2c00943c" containerName="dnsmasq-dns" containerID="cri-o://9e39560939f2c6a63b846fe7631faa7dc91c92edf7fece72891f4dfc17d751a2" gracePeriod=10 Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.916049 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9-config-data-custom\") pod \"barbican-worker-6c48df686c-lz5gl\" (UID: \"f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9\") " pod="openstack/barbican-worker-6c48df686c-lz5gl" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.916788 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9-combined-ca-bundle\") pod \"barbican-worker-6c48df686c-lz5gl\" (UID: \"f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9\") " pod="openstack/barbican-worker-6c48df686c-lz5gl" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.937664 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9-config-data\") pod \"barbican-worker-6c48df686c-lz5gl\" (UID: \"f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9\") " pod="openstack/barbican-worker-6c48df686c-lz5gl" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.951016 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n27xr\" (UniqueName: \"kubernetes.io/projected/f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9-kube-api-access-n27xr\") pod \"barbican-worker-6c48df686c-lz5gl\" (UID: \"f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9\") " pod="openstack/barbican-worker-6c48df686c-lz5gl" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.964151 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6c48df686c-lz5gl" Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.977703 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78b7d887cf-gdzl7"] Mar 18 20:26:40 crc kubenswrapper[4950]: I0318 20:26:40.979307 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.016425 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78b7d887cf-gdzl7"] Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.018037 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4fw2\" (UniqueName: \"kubernetes.io/projected/72fe734a-54a6-4391-a125-04ddd4c9c890-kube-api-access-r4fw2\") pod \"barbican-keystone-listener-546478654b-mwfqm\" (UID: \"72fe734a-54a6-4391-a125-04ddd4c9c890\") " pod="openstack/barbican-keystone-listener-546478654b-mwfqm" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.018096 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72fe734a-54a6-4391-a125-04ddd4c9c890-combined-ca-bundle\") pod \"barbican-keystone-listener-546478654b-mwfqm\" (UID: \"72fe734a-54a6-4391-a125-04ddd4c9c890\") " pod="openstack/barbican-keystone-listener-546478654b-mwfqm" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.018143 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/72fe734a-54a6-4391-a125-04ddd4c9c890-config-data-custom\") pod \"barbican-keystone-listener-546478654b-mwfqm\" (UID: \"72fe734a-54a6-4391-a125-04ddd4c9c890\") " pod="openstack/barbican-keystone-listener-546478654b-mwfqm" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.018221 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72fe734a-54a6-4391-a125-04ddd4c9c890-config-data\") pod \"barbican-keystone-listener-546478654b-mwfqm\" (UID: \"72fe734a-54a6-4391-a125-04ddd4c9c890\") " pod="openstack/barbican-keystone-listener-546478654b-mwfqm" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.018245 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72fe734a-54a6-4391-a125-04ddd4c9c890-logs\") pod \"barbican-keystone-listener-546478654b-mwfqm\" (UID: \"72fe734a-54a6-4391-a125-04ddd4c9c890\") " pod="openstack/barbican-keystone-listener-546478654b-mwfqm" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.018601 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72fe734a-54a6-4391-a125-04ddd4c9c890-logs\") pod \"barbican-keystone-listener-546478654b-mwfqm\" (UID: \"72fe734a-54a6-4391-a125-04ddd4c9c890\") " pod="openstack/barbican-keystone-listener-546478654b-mwfqm" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.030206 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72fe734a-54a6-4391-a125-04ddd4c9c890-combined-ca-bundle\") pod \"barbican-keystone-listener-546478654b-mwfqm\" (UID: \"72fe734a-54a6-4391-a125-04ddd4c9c890\") " pod="openstack/barbican-keystone-listener-546478654b-mwfqm" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.030489 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72fe734a-54a6-4391-a125-04ddd4c9c890-config-data\") pod \"barbican-keystone-listener-546478654b-mwfqm\" (UID: \"72fe734a-54a6-4391-a125-04ddd4c9c890\") " pod="openstack/barbican-keystone-listener-546478654b-mwfqm" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.033033 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/72fe734a-54a6-4391-a125-04ddd4c9c890-config-data-custom\") pod \"barbican-keystone-listener-546478654b-mwfqm\" (UID: \"72fe734a-54a6-4391-a125-04ddd4c9c890\") " pod="openstack/barbican-keystone-listener-546478654b-mwfqm" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.070485 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-776bcc8684-8mp9k"] Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.087710 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4fw2\" (UniqueName: \"kubernetes.io/projected/72fe734a-54a6-4391-a125-04ddd4c9c890-kube-api-access-r4fw2\") pod \"barbican-keystone-listener-546478654b-mwfqm\" (UID: \"72fe734a-54a6-4391-a125-04ddd4c9c890\") " pod="openstack/barbican-keystone-listener-546478654b-mwfqm" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.098805 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.106629 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.119730 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-ovsdbserver-nb\") pod \"dnsmasq-dns-78b7d887cf-gdzl7\" (UID: \"0013b445-540c-472b-9ead-44958805929b\") " pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.119822 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-869b7\" (UniqueName: \"kubernetes.io/projected/0013b445-540c-472b-9ead-44958805929b-kube-api-access-869b7\") pod \"dnsmasq-dns-78b7d887cf-gdzl7\" (UID: \"0013b445-540c-472b-9ead-44958805929b\") " pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.119851 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-dns-svc\") pod \"dnsmasq-dns-78b7d887cf-gdzl7\" (UID: \"0013b445-540c-472b-9ead-44958805929b\") " pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.119917 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-ovsdbserver-sb\") pod \"dnsmasq-dns-78b7d887cf-gdzl7\" (UID: \"0013b445-540c-472b-9ead-44958805929b\") " pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.119989 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-config\") pod \"dnsmasq-dns-78b7d887cf-gdzl7\" (UID: \"0013b445-540c-472b-9ead-44958805929b\") " pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.190730 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-776bcc8684-8mp9k"] Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.225338 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-combined-ca-bundle\") pod \"barbican-api-776bcc8684-8mp9k\" (UID: \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\") " pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.225383 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-dns-svc\") pod \"dnsmasq-dns-78b7d887cf-gdzl7\" (UID: \"0013b445-540c-472b-9ead-44958805929b\") " pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.225403 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-logs\") pod \"barbican-api-776bcc8684-8mp9k\" (UID: \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\") " pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.225498 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-ovsdbserver-sb\") pod \"dnsmasq-dns-78b7d887cf-gdzl7\" (UID: \"0013b445-540c-472b-9ead-44958805929b\") " pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.225544 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-config-data-custom\") pod \"barbican-api-776bcc8684-8mp9k\" (UID: \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\") " pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.225573 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-config\") pod \"dnsmasq-dns-78b7d887cf-gdzl7\" (UID: \"0013b445-540c-472b-9ead-44958805929b\") " pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.225600 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-ovsdbserver-nb\") pod \"dnsmasq-dns-78b7d887cf-gdzl7\" (UID: \"0013b445-540c-472b-9ead-44958805929b\") " pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.225636 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8s2z2\" (UniqueName: \"kubernetes.io/projected/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-kube-api-access-8s2z2\") pod \"barbican-api-776bcc8684-8mp9k\" (UID: \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\") " pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.225656 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-config-data\") pod \"barbican-api-776bcc8684-8mp9k\" (UID: \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\") " pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.225676 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-869b7\" (UniqueName: \"kubernetes.io/projected/0013b445-540c-472b-9ead-44958805929b-kube-api-access-869b7\") pod \"dnsmasq-dns-78b7d887cf-gdzl7\" (UID: \"0013b445-540c-472b-9ead-44958805929b\") " pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.226750 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-dns-svc\") pod \"dnsmasq-dns-78b7d887cf-gdzl7\" (UID: \"0013b445-540c-472b-9ead-44958805929b\") " pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.227013 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-config\") pod \"dnsmasq-dns-78b7d887cf-gdzl7\" (UID: \"0013b445-540c-472b-9ead-44958805929b\") " pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.227265 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-ovsdbserver-nb\") pod \"dnsmasq-dns-78b7d887cf-gdzl7\" (UID: \"0013b445-540c-472b-9ead-44958805929b\") " pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.227585 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-ovsdbserver-sb\") pod \"dnsmasq-dns-78b7d887cf-gdzl7\" (UID: \"0013b445-540c-472b-9ead-44958805929b\") " pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.259047 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-869b7\" (UniqueName: \"kubernetes.io/projected/0013b445-540c-472b-9ead-44958805929b-kube-api-access-869b7\") pod \"dnsmasq-dns-78b7d887cf-gdzl7\" (UID: \"0013b445-540c-472b-9ead-44958805929b\") " pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.281070 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-546478654b-mwfqm" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.322920 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.327382 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-config-data-custom\") pod \"barbican-api-776bcc8684-8mp9k\" (UID: \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\") " pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.327488 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8s2z2\" (UniqueName: \"kubernetes.io/projected/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-kube-api-access-8s2z2\") pod \"barbican-api-776bcc8684-8mp9k\" (UID: \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\") " pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.327513 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-config-data\") pod \"barbican-api-776bcc8684-8mp9k\" (UID: \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\") " pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.327538 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-combined-ca-bundle\") pod \"barbican-api-776bcc8684-8mp9k\" (UID: \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\") " pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.327556 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-logs\") pod \"barbican-api-776bcc8684-8mp9k\" (UID: \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\") " pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.327995 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-logs\") pod \"barbican-api-776bcc8684-8mp9k\" (UID: \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\") " pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.332141 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-combined-ca-bundle\") pod \"barbican-api-776bcc8684-8mp9k\" (UID: \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\") " pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.332680 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-config-data-custom\") pod \"barbican-api-776bcc8684-8mp9k\" (UID: \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\") " pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.334251 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-config-data\") pod \"barbican-api-776bcc8684-8mp9k\" (UID: \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\") " pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.352906 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8s2z2\" (UniqueName: \"kubernetes.io/projected/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-kube-api-access-8s2z2\") pod \"barbican-api-776bcc8684-8mp9k\" (UID: \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\") " pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.446269 4950 generic.go:334] "Generic (PLEG): container finished" podID="292b3f8d-075a-4d1d-8aa9-56dd2c00943c" containerID="9e39560939f2c6a63b846fe7631faa7dc91c92edf7fece72891f4dfc17d751a2" exitCode=0 Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.447502 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6648966d55-b42h7" event={"ID":"292b3f8d-075a-4d1d-8aa9-56dd2c00943c","Type":"ContainerDied","Data":"9e39560939f2c6a63b846fe7631faa7dc91c92edf7fece72891f4dfc17d751a2"} Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.465074 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.504895 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ae778b5-53f6-4fbf-b779-83a194d4da7d" path="/var/lib/kubelet/pods/6ae778b5-53f6-4fbf-b779-83a194d4da7d/volumes" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.574895 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.846282 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6648966d55-b42h7" Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.975333 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-ovsdbserver-nb\") pod \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\" (UID: \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\") " Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.975388 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-config\") pod \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\" (UID: \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\") " Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.975445 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9zks\" (UniqueName: \"kubernetes.io/projected/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-kube-api-access-h9zks\") pod \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\" (UID: \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\") " Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.975570 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-ovsdbserver-sb\") pod \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\" (UID: \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\") " Mar 18 20:26:41 crc kubenswrapper[4950]: I0318 20:26:41.975621 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-dns-svc\") pod \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\" (UID: \"292b3f8d-075a-4d1d-8aa9-56dd2c00943c\") " Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.014540 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-kube-api-access-h9zks" (OuterVolumeSpecName: "kube-api-access-h9zks") pod "292b3f8d-075a-4d1d-8aa9-56dd2c00943c" (UID: "292b3f8d-075a-4d1d-8aa9-56dd2c00943c"). InnerVolumeSpecName "kube-api-access-h9zks". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.064554 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6c48df686c-lz5gl"] Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.072990 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "292b3f8d-075a-4d1d-8aa9-56dd2c00943c" (UID: "292b3f8d-075a-4d1d-8aa9-56dd2c00943c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.079580 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9zks\" (UniqueName: \"kubernetes.io/projected/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-kube-api-access-h9zks\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.079608 4950 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.082306 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-546478654b-mwfqm"] Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.225038 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-config" (OuterVolumeSpecName: "config") pod "292b3f8d-075a-4d1d-8aa9-56dd2c00943c" (UID: "292b3f8d-075a-4d1d-8aa9-56dd2c00943c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.259055 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "292b3f8d-075a-4d1d-8aa9-56dd2c00943c" (UID: "292b3f8d-075a-4d1d-8aa9-56dd2c00943c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.292014 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "292b3f8d-075a-4d1d-8aa9-56dd2c00943c" (UID: "292b3f8d-075a-4d1d-8aa9-56dd2c00943c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.294595 4950 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.294625 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.294633 4950 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/292b3f8d-075a-4d1d-8aa9-56dd2c00943c-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.423184 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78b7d887cf-gdzl7"] Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.487781 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6c48df686c-lz5gl" event={"ID":"f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9","Type":"ContainerStarted","Data":"fe7889be5fceb9aec44b3cb4a06d99d42337daafc80f2be47a24bf524e3fb31c"} Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.500301 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-546478654b-mwfqm" event={"ID":"72fe734a-54a6-4391-a125-04ddd4c9c890","Type":"ContainerStarted","Data":"e9c8323c0c14d25311d802225503f39427591d18f2310244d72d9f7dfcba01ca"} Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.507751 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1bae85ae-0545-46ac-90fa-c8f967648ed4","Type":"ContainerStarted","Data":"6b44ab413d2915534ae76456c21af11d46709990ac9b88f60b77cf12af644352"} Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.508823 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" event={"ID":"0013b445-540c-472b-9ead-44958805929b","Type":"ContainerStarted","Data":"c6fac8752917e54e8553c6e43ad7b4408d8b7d6f86f2a8443a53a0b24cacc620"} Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.515953 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6648966d55-b42h7" event={"ID":"292b3f8d-075a-4d1d-8aa9-56dd2c00943c","Type":"ContainerDied","Data":"927554ad0952af8811296c017ee1311035e9da9c03bfd58a0c23e295a7fb5f38"} Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.515991 4950 scope.go:117] "RemoveContainer" containerID="9e39560939f2c6a63b846fe7631faa7dc91c92edf7fece72891f4dfc17d751a2" Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.516107 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6648966d55-b42h7" Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.561753 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6648966d55-b42h7"] Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.572854 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6648966d55-b42h7"] Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.616922 4950 scope.go:117] "RemoveContainer" containerID="92cb7ef93cf371a845543fb38f5e1870998e9149cf186cce812098557089a974" Mar 18 20:26:42 crc kubenswrapper[4950]: I0318 20:26:42.725288 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-776bcc8684-8mp9k"] Mar 18 20:26:43 crc kubenswrapper[4950]: I0318 20:26:43.460029 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-c888c6fcb-jnt6f" podUID="ed97e535-9755-40ef-a82f-176f0eb7e31b" containerName="neutron-httpd" probeResult="failure" output="Get \"http://10.217.0.144:9696/\": dial tcp 10.217.0.144:9696: connect: connection refused" Mar 18 20:26:43 crc kubenswrapper[4950]: I0318 20:26:43.497903 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="292b3f8d-075a-4d1d-8aa9-56dd2c00943c" path="/var/lib/kubelet/pods/292b3f8d-075a-4d1d-8aa9-56dd2c00943c/volumes" Mar 18 20:26:43 crc kubenswrapper[4950]: I0318 20:26:43.545444 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1bae85ae-0545-46ac-90fa-c8f967648ed4","Type":"ContainerStarted","Data":"c9a201087aa494a78c16e030a3e47713798d561a74c0ada4cbccc002761f0831"} Mar 18 20:26:43 crc kubenswrapper[4950]: I0318 20:26:43.552080 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a00a076-ca9d-46e9-b3e7-af13dcaca76e","Type":"ContainerStarted","Data":"7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891"} Mar 18 20:26:43 crc kubenswrapper[4950]: I0318 20:26:43.553088 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 18 20:26:43 crc kubenswrapper[4950]: I0318 20:26:43.554310 4950 generic.go:334] "Generic (PLEG): container finished" podID="0013b445-540c-472b-9ead-44958805929b" containerID="77033a50291f56779c6983dc1cfa7d6d4a076aeb6a41ecb9042247ce5c227943" exitCode=0 Mar 18 20:26:43 crc kubenswrapper[4950]: I0318 20:26:43.554354 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" event={"ID":"0013b445-540c-472b-9ead-44958805929b","Type":"ContainerDied","Data":"77033a50291f56779c6983dc1cfa7d6d4a076aeb6a41ecb9042247ce5c227943"} Mar 18 20:26:43 crc kubenswrapper[4950]: I0318 20:26:43.576772 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.278006551 podStartE2EDuration="7.576755703s" podCreationTimestamp="2026-03-18 20:26:36 +0000 UTC" firstStartedPulling="2026-03-18 20:26:37.572729424 +0000 UTC m=+1210.813571292" lastFinishedPulling="2026-03-18 20:26:41.871478576 +0000 UTC m=+1215.112320444" observedRunningTime="2026-03-18 20:26:43.573812107 +0000 UTC m=+1216.814653975" watchObservedRunningTime="2026-03-18 20:26:43.576755703 +0000 UTC m=+1216.817597571" Mar 18 20:26:43 crc kubenswrapper[4950]: I0318 20:26:43.580479 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-776bcc8684-8mp9k" event={"ID":"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1","Type":"ContainerStarted","Data":"9584a8936196d9827442cd1eb57252fcb4c3fee9630c39e28cb477cca30362be"} Mar 18 20:26:43 crc kubenswrapper[4950]: I0318 20:26:43.580533 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-776bcc8684-8mp9k" event={"ID":"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1","Type":"ContainerStarted","Data":"896addbe1c7c1e9be3a785950f100544eb21fb5c2eb42a17124da60c933c03be"} Mar 18 20:26:43 crc kubenswrapper[4950]: I0318 20:26:43.580543 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-776bcc8684-8mp9k" event={"ID":"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1","Type":"ContainerStarted","Data":"7856acbf054d957f91306906724b142712d6e291d952913bd62b0223a8e1298c"} Mar 18 20:26:43 crc kubenswrapper[4950]: I0318 20:26:43.581387 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:26:43 crc kubenswrapper[4950]: I0318 20:26:43.581425 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:26:43 crc kubenswrapper[4950]: I0318 20:26:43.627842 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-776bcc8684-8mp9k" podStartSLOduration=2.627826468 podStartE2EDuration="2.627826468s" podCreationTimestamp="2026-03-18 20:26:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:26:43.626694849 +0000 UTC m=+1216.867536717" watchObservedRunningTime="2026-03-18 20:26:43.627826468 +0000 UTC m=+1216.868668336" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.288704 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7d5775f6d4-lshhl"] Mar 18 20:26:44 crc kubenswrapper[4950]: E0318 20:26:44.289443 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="292b3f8d-075a-4d1d-8aa9-56dd2c00943c" containerName="dnsmasq-dns" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.289465 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="292b3f8d-075a-4d1d-8aa9-56dd2c00943c" containerName="dnsmasq-dns" Mar 18 20:26:44 crc kubenswrapper[4950]: E0318 20:26:44.289490 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="292b3f8d-075a-4d1d-8aa9-56dd2c00943c" containerName="init" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.289497 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="292b3f8d-075a-4d1d-8aa9-56dd2c00943c" containerName="init" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.289666 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="292b3f8d-075a-4d1d-8aa9-56dd2c00943c" containerName="dnsmasq-dns" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.290494 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.297134 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.297362 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.318785 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7d5775f6d4-lshhl"] Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.432572 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f111df34-12e9-441b-a362-e94471ec8eaf-logs\") pod \"barbican-api-7d5775f6d4-lshhl\" (UID: \"f111df34-12e9-441b-a362-e94471ec8eaf\") " pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.432609 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f111df34-12e9-441b-a362-e94471ec8eaf-internal-tls-certs\") pod \"barbican-api-7d5775f6d4-lshhl\" (UID: \"f111df34-12e9-441b-a362-e94471ec8eaf\") " pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.432676 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f111df34-12e9-441b-a362-e94471ec8eaf-config-data\") pod \"barbican-api-7d5775f6d4-lshhl\" (UID: \"f111df34-12e9-441b-a362-e94471ec8eaf\") " pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.432705 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f111df34-12e9-441b-a362-e94471ec8eaf-config-data-custom\") pod \"barbican-api-7d5775f6d4-lshhl\" (UID: \"f111df34-12e9-441b-a362-e94471ec8eaf\") " pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.432749 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ths4\" (UniqueName: \"kubernetes.io/projected/f111df34-12e9-441b-a362-e94471ec8eaf-kube-api-access-8ths4\") pod \"barbican-api-7d5775f6d4-lshhl\" (UID: \"f111df34-12e9-441b-a362-e94471ec8eaf\") " pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.432780 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f111df34-12e9-441b-a362-e94471ec8eaf-public-tls-certs\") pod \"barbican-api-7d5775f6d4-lshhl\" (UID: \"f111df34-12e9-441b-a362-e94471ec8eaf\") " pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.432810 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f111df34-12e9-441b-a362-e94471ec8eaf-combined-ca-bundle\") pod \"barbican-api-7d5775f6d4-lshhl\" (UID: \"f111df34-12e9-441b-a362-e94471ec8eaf\") " pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.535065 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f111df34-12e9-441b-a362-e94471ec8eaf-config-data\") pod \"barbican-api-7d5775f6d4-lshhl\" (UID: \"f111df34-12e9-441b-a362-e94471ec8eaf\") " pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.535112 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f111df34-12e9-441b-a362-e94471ec8eaf-config-data-custom\") pod \"barbican-api-7d5775f6d4-lshhl\" (UID: \"f111df34-12e9-441b-a362-e94471ec8eaf\") " pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.535137 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ths4\" (UniqueName: \"kubernetes.io/projected/f111df34-12e9-441b-a362-e94471ec8eaf-kube-api-access-8ths4\") pod \"barbican-api-7d5775f6d4-lshhl\" (UID: \"f111df34-12e9-441b-a362-e94471ec8eaf\") " pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.535175 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f111df34-12e9-441b-a362-e94471ec8eaf-public-tls-certs\") pod \"barbican-api-7d5775f6d4-lshhl\" (UID: \"f111df34-12e9-441b-a362-e94471ec8eaf\") " pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.535206 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f111df34-12e9-441b-a362-e94471ec8eaf-combined-ca-bundle\") pod \"barbican-api-7d5775f6d4-lshhl\" (UID: \"f111df34-12e9-441b-a362-e94471ec8eaf\") " pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.535273 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f111df34-12e9-441b-a362-e94471ec8eaf-logs\") pod \"barbican-api-7d5775f6d4-lshhl\" (UID: \"f111df34-12e9-441b-a362-e94471ec8eaf\") " pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.535289 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f111df34-12e9-441b-a362-e94471ec8eaf-internal-tls-certs\") pod \"barbican-api-7d5775f6d4-lshhl\" (UID: \"f111df34-12e9-441b-a362-e94471ec8eaf\") " pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.536271 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f111df34-12e9-441b-a362-e94471ec8eaf-logs\") pod \"barbican-api-7d5775f6d4-lshhl\" (UID: \"f111df34-12e9-441b-a362-e94471ec8eaf\") " pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.553474 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f111df34-12e9-441b-a362-e94471ec8eaf-combined-ca-bundle\") pod \"barbican-api-7d5775f6d4-lshhl\" (UID: \"f111df34-12e9-441b-a362-e94471ec8eaf\") " pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.554443 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f111df34-12e9-441b-a362-e94471ec8eaf-public-tls-certs\") pod \"barbican-api-7d5775f6d4-lshhl\" (UID: \"f111df34-12e9-441b-a362-e94471ec8eaf\") " pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.557277 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f111df34-12e9-441b-a362-e94471ec8eaf-internal-tls-certs\") pod \"barbican-api-7d5775f6d4-lshhl\" (UID: \"f111df34-12e9-441b-a362-e94471ec8eaf\") " pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.559103 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f111df34-12e9-441b-a362-e94471ec8eaf-config-data\") pod \"barbican-api-7d5775f6d4-lshhl\" (UID: \"f111df34-12e9-441b-a362-e94471ec8eaf\") " pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.562112 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f111df34-12e9-441b-a362-e94471ec8eaf-config-data-custom\") pod \"barbican-api-7d5775f6d4-lshhl\" (UID: \"f111df34-12e9-441b-a362-e94471ec8eaf\") " pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.571086 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ths4\" (UniqueName: \"kubernetes.io/projected/f111df34-12e9-441b-a362-e94471ec8eaf-kube-api-access-8ths4\") pod \"barbican-api-7d5775f6d4-lshhl\" (UID: \"f111df34-12e9-441b-a362-e94471ec8eaf\") " pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.605951 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1bae85ae-0545-46ac-90fa-c8f967648ed4","Type":"ContainerStarted","Data":"0fd3b5aa0d3514bfc4f5fe027b83915dbef48df07316d9e05b20ce32ff98e494"} Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.607228 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.620490 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.652165 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.6521432019999995 podStartE2EDuration="4.652143202s" podCreationTimestamp="2026-03-18 20:26:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:26:44.638912562 +0000 UTC m=+1217.879754430" watchObservedRunningTime="2026-03-18 20:26:44.652143202 +0000 UTC m=+1217.892985070" Mar 18 20:26:44 crc kubenswrapper[4950]: I0318 20:26:44.975923 4950 scope.go:117] "RemoveContainer" containerID="54d11f46c8dc572a33f1b53b29578543919f5927e4093db85e04577390e80f45" Mar 18 20:26:45 crc kubenswrapper[4950]: I0318 20:26:45.462061 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 18 20:26:45 crc kubenswrapper[4950]: I0318 20:26:45.564853 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 18 20:26:45 crc kubenswrapper[4950]: I0318 20:26:45.613503 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="eac8a754-52b8-471c-bce1-85c449f801ba" containerName="cinder-scheduler" containerID="cri-o://42d560c87b7d449e8e9b9bb91455358f6f1c39aa6f8e2cb7378e5b8ac837ccfa" gracePeriod=30 Mar 18 20:26:45 crc kubenswrapper[4950]: I0318 20:26:45.614579 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="eac8a754-52b8-471c-bce1-85c449f801ba" containerName="probe" containerID="cri-o://c956a40169bc02e7c0273be7eebc398074072389554d2646b06f1a2f18bd2b83" gracePeriod=30 Mar 18 20:26:45 crc kubenswrapper[4950]: I0318 20:26:45.906461 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7d5775f6d4-lshhl"] Mar 18 20:26:46 crc kubenswrapper[4950]: I0318 20:26:46.329447 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:26:46 crc kubenswrapper[4950]: I0318 20:26:46.642048 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" event={"ID":"0013b445-540c-472b-9ead-44958805929b","Type":"ContainerStarted","Data":"243cc4d14f94fd3cd99d4d03c8a4c6451725b42cad9f6e9d0c523e978a85fddd"} Mar 18 20:26:46 crc kubenswrapper[4950]: I0318 20:26:46.642095 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" Mar 18 20:26:46 crc kubenswrapper[4950]: I0318 20:26:46.650210 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7d5775f6d4-lshhl" event={"ID":"f111df34-12e9-441b-a362-e94471ec8eaf","Type":"ContainerStarted","Data":"4cb03c8e26222f068fdf243a16d8308ad5477ca9c0a5767a109a6b1d58d83818"} Mar 18 20:26:46 crc kubenswrapper[4950]: I0318 20:26:46.650256 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7d5775f6d4-lshhl" event={"ID":"f111df34-12e9-441b-a362-e94471ec8eaf","Type":"ContainerStarted","Data":"5730b0e6747636b0c8f1672132a1df0619486a75fd66909149f9d6b6b76a1728"} Mar 18 20:26:46 crc kubenswrapper[4950]: I0318 20:26:46.650271 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7d5775f6d4-lshhl" event={"ID":"f111df34-12e9-441b-a362-e94471ec8eaf","Type":"ContainerStarted","Data":"d36a53cb4873059daba1b8691cb5f2b9cb715f84d118cc2f15370cfdaef4c62d"} Mar 18 20:26:46 crc kubenswrapper[4950]: I0318 20:26:46.650638 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:46 crc kubenswrapper[4950]: I0318 20:26:46.650756 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:46 crc kubenswrapper[4950]: I0318 20:26:46.657666 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6c48df686c-lz5gl" event={"ID":"f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9","Type":"ContainerStarted","Data":"6b81e7e57694eaaf3d23853344c14f6fe2c80d633b07c35a4ef73cccf8961e0a"} Mar 18 20:26:46 crc kubenswrapper[4950]: I0318 20:26:46.657708 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6c48df686c-lz5gl" event={"ID":"f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9","Type":"ContainerStarted","Data":"36b8c758fe653d9db72cdcdc4f066a92c2d308a3f01202695d4bca2fcbc0d716"} Mar 18 20:26:46 crc kubenswrapper[4950]: I0318 20:26:46.667496 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" podStartSLOduration=6.66739548 podStartE2EDuration="6.66739548s" podCreationTimestamp="2026-03-18 20:26:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:26:46.659890917 +0000 UTC m=+1219.900732785" watchObservedRunningTime="2026-03-18 20:26:46.66739548 +0000 UTC m=+1219.908237348" Mar 18 20:26:46 crc kubenswrapper[4950]: I0318 20:26:46.668566 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-546478654b-mwfqm" event={"ID":"72fe734a-54a6-4391-a125-04ddd4c9c890","Type":"ContainerStarted","Data":"8762d34115592e4c7567b01968ba813d32ebb76f5c524d6f1bfd3033442c1d5d"} Mar 18 20:26:46 crc kubenswrapper[4950]: I0318 20:26:46.668620 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-546478654b-mwfqm" event={"ID":"72fe734a-54a6-4391-a125-04ddd4c9c890","Type":"ContainerStarted","Data":"ff782d4340479cff69203700e61fb8a086012135024d4af2a9d033b42457814d"} Mar 18 20:26:46 crc kubenswrapper[4950]: I0318 20:26:46.676679 4950 generic.go:334] "Generic (PLEG): container finished" podID="eac8a754-52b8-471c-bce1-85c449f801ba" containerID="c956a40169bc02e7c0273be7eebc398074072389554d2646b06f1a2f18bd2b83" exitCode=0 Mar 18 20:26:46 crc kubenswrapper[4950]: I0318 20:26:46.677548 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"eac8a754-52b8-471c-bce1-85c449f801ba","Type":"ContainerDied","Data":"c956a40169bc02e7c0273be7eebc398074072389554d2646b06f1a2f18bd2b83"} Mar 18 20:26:46 crc kubenswrapper[4950]: I0318 20:26:46.690744 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-6c48df686c-lz5gl" podStartSLOduration=3.5087100810000003 podStartE2EDuration="6.690726441s" podCreationTimestamp="2026-03-18 20:26:40 +0000 UTC" firstStartedPulling="2026-03-18 20:26:42.139002234 +0000 UTC m=+1215.379844102" lastFinishedPulling="2026-03-18 20:26:45.321018594 +0000 UTC m=+1218.561860462" observedRunningTime="2026-03-18 20:26:46.678036394 +0000 UTC m=+1219.918878252" watchObservedRunningTime="2026-03-18 20:26:46.690726441 +0000 UTC m=+1219.931568309" Mar 18 20:26:46 crc kubenswrapper[4950]: I0318 20:26:46.712392 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7d5775f6d4-lshhl" podStartSLOduration=2.712369818 podStartE2EDuration="2.712369818s" podCreationTimestamp="2026-03-18 20:26:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:26:46.696553511 +0000 UTC m=+1219.937395379" watchObservedRunningTime="2026-03-18 20:26:46.712369818 +0000 UTC m=+1219.953211686" Mar 18 20:26:46 crc kubenswrapper[4950]: I0318 20:26:46.760293 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-546478654b-mwfqm" podStartSLOduration=3.587434258 podStartE2EDuration="6.760269311s" podCreationTimestamp="2026-03-18 20:26:40 +0000 UTC" firstStartedPulling="2026-03-18 20:26:42.14779017 +0000 UTC m=+1215.388632038" lastFinishedPulling="2026-03-18 20:26:45.320625223 +0000 UTC m=+1218.561467091" observedRunningTime="2026-03-18 20:26:46.717387637 +0000 UTC m=+1219.958229505" watchObservedRunningTime="2026-03-18 20:26:46.760269311 +0000 UTC m=+1220.001111179" Mar 18 20:26:47 crc kubenswrapper[4950]: E0318 20:26:47.608309 4950 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/99e2798c18957b8afdfd0ca3284cd5783b14fa113c2001284d80398d88a20cad/diff" to get inode usage: stat /var/lib/containers/storage/overlay/99e2798c18957b8afdfd0ca3284cd5783b14fa113c2001284d80398d88a20cad/diff: no such file or directory, extraDiskErr: Mar 18 20:26:47 crc kubenswrapper[4950]: I0318 20:26:47.705495 4950 generic.go:334] "Generic (PLEG): container finished" podID="eac8a754-52b8-471c-bce1-85c449f801ba" containerID="42d560c87b7d449e8e9b9bb91455358f6f1c39aa6f8e2cb7378e5b8ac837ccfa" exitCode=0 Mar 18 20:26:47 crc kubenswrapper[4950]: I0318 20:26:47.706261 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"eac8a754-52b8-471c-bce1-85c449f801ba","Type":"ContainerDied","Data":"42d560c87b7d449e8e9b9bb91455358f6f1c39aa6f8e2cb7378e5b8ac837ccfa"} Mar 18 20:26:47 crc kubenswrapper[4950]: I0318 20:26:47.955110 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.135896 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-config-data\") pod \"eac8a754-52b8-471c-bce1-85c449f801ba\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.135977 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-combined-ca-bundle\") pod \"eac8a754-52b8-471c-bce1-85c449f801ba\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.136005 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-config-data-custom\") pod \"eac8a754-52b8-471c-bce1-85c449f801ba\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.136029 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eac8a754-52b8-471c-bce1-85c449f801ba-etc-machine-id\") pod \"eac8a754-52b8-471c-bce1-85c449f801ba\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.136090 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpbnc\" (UniqueName: \"kubernetes.io/projected/eac8a754-52b8-471c-bce1-85c449f801ba-kube-api-access-vpbnc\") pod \"eac8a754-52b8-471c-bce1-85c449f801ba\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.136141 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-scripts\") pod \"eac8a754-52b8-471c-bce1-85c449f801ba\" (UID: \"eac8a754-52b8-471c-bce1-85c449f801ba\") " Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.136431 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eac8a754-52b8-471c-bce1-85c449f801ba-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "eac8a754-52b8-471c-bce1-85c449f801ba" (UID: "eac8a754-52b8-471c-bce1-85c449f801ba"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.136619 4950 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eac8a754-52b8-471c-bce1-85c449f801ba-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.154041 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-scripts" (OuterVolumeSpecName: "scripts") pod "eac8a754-52b8-471c-bce1-85c449f801ba" (UID: "eac8a754-52b8-471c-bce1-85c449f801ba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.178681 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eac8a754-52b8-471c-bce1-85c449f801ba-kube-api-access-vpbnc" (OuterVolumeSpecName: "kube-api-access-vpbnc") pod "eac8a754-52b8-471c-bce1-85c449f801ba" (UID: "eac8a754-52b8-471c-bce1-85c449f801ba"). InnerVolumeSpecName "kube-api-access-vpbnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.207558 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "eac8a754-52b8-471c-bce1-85c449f801ba" (UID: "eac8a754-52b8-471c-bce1-85c449f801ba"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.238095 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.238122 4950 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.238132 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpbnc\" (UniqueName: \"kubernetes.io/projected/eac8a754-52b8-471c-bce1-85c449f801ba-kube-api-access-vpbnc\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.301996 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eac8a754-52b8-471c-bce1-85c449f801ba" (UID: "eac8a754-52b8-471c-bce1-85c449f801ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.311585 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-config-data" (OuterVolumeSpecName: "config-data") pod "eac8a754-52b8-471c-bce1-85c449f801ba" (UID: "eac8a754-52b8-471c-bce1-85c449f801ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.339699 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.339740 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eac8a754-52b8-471c-bce1-85c449f801ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.715260 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"eac8a754-52b8-471c-bce1-85c449f801ba","Type":"ContainerDied","Data":"d3ccce4ca8ca5f25f58a935906e075ef2e4680759dd54bb310fc16bd73ae07ca"} Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.715306 4950 scope.go:117] "RemoveContainer" containerID="c956a40169bc02e7c0273be7eebc398074072389554d2646b06f1a2f18bd2b83" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.715424 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.756262 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.771028 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.772314 4950 scope.go:117] "RemoveContainer" containerID="42d560c87b7d449e8e9b9bb91455358f6f1c39aa6f8e2cb7378e5b8ac837ccfa" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.785434 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 18 20:26:48 crc kubenswrapper[4950]: E0318 20:26:48.785912 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eac8a754-52b8-471c-bce1-85c449f801ba" containerName="probe" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.785932 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="eac8a754-52b8-471c-bce1-85c449f801ba" containerName="probe" Mar 18 20:26:48 crc kubenswrapper[4950]: E0318 20:26:48.785979 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eac8a754-52b8-471c-bce1-85c449f801ba" containerName="cinder-scheduler" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.785987 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="eac8a754-52b8-471c-bce1-85c449f801ba" containerName="cinder-scheduler" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.786188 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="eac8a754-52b8-471c-bce1-85c449f801ba" containerName="cinder-scheduler" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.786202 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="eac8a754-52b8-471c-bce1-85c449f801ba" containerName="probe" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.787217 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.803228 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.822171 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.957245 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbb2e2de-55ef-4bc1-a8df-1428e18c664e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fbb2e2de-55ef-4bc1-a8df-1428e18c664e\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.957302 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbb2e2de-55ef-4bc1-a8df-1428e18c664e-scripts\") pod \"cinder-scheduler-0\" (UID: \"fbb2e2de-55ef-4bc1-a8df-1428e18c664e\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.957343 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbb2e2de-55ef-4bc1-a8df-1428e18c664e-config-data\") pod \"cinder-scheduler-0\" (UID: \"fbb2e2de-55ef-4bc1-a8df-1428e18c664e\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.957440 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76hk8\" (UniqueName: \"kubernetes.io/projected/fbb2e2de-55ef-4bc1-a8df-1428e18c664e-kube-api-access-76hk8\") pod \"cinder-scheduler-0\" (UID: \"fbb2e2de-55ef-4bc1-a8df-1428e18c664e\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.957625 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fbb2e2de-55ef-4bc1-a8df-1428e18c664e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fbb2e2de-55ef-4bc1-a8df-1428e18c664e\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:48 crc kubenswrapper[4950]: I0318 20:26:48.957670 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fbb2e2de-55ef-4bc1-a8df-1428e18c664e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fbb2e2de-55ef-4bc1-a8df-1428e18c664e\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:48 crc kubenswrapper[4950]: E0318 20:26:48.996742 4950 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/2b475ee87a24649c020c115a6f660d4e268d87926bf28082972a7d691d0f31de/diff" to get inode usage: stat /var/lib/containers/storage/overlay/2b475ee87a24649c020c115a6f660d4e268d87926bf28082972a7d691d0f31de/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_dnsmasq-dns-74b5448f7f-zdrjg_ad591d5a-4120-4285-9cca-417b19176795/dnsmasq-dns/0.log" to get inode usage: stat /var/log/pods/openstack_dnsmasq-dns-74b5448f7f-zdrjg_ad591d5a-4120-4285-9cca-417b19176795/dnsmasq-dns/0.log: no such file or directory Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.059239 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fbb2e2de-55ef-4bc1-a8df-1428e18c664e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fbb2e2de-55ef-4bc1-a8df-1428e18c664e\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.059299 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fbb2e2de-55ef-4bc1-a8df-1428e18c664e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fbb2e2de-55ef-4bc1-a8df-1428e18c664e\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.059349 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbb2e2de-55ef-4bc1-a8df-1428e18c664e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fbb2e2de-55ef-4bc1-a8df-1428e18c664e\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.059367 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbb2e2de-55ef-4bc1-a8df-1428e18c664e-scripts\") pod \"cinder-scheduler-0\" (UID: \"fbb2e2de-55ef-4bc1-a8df-1428e18c664e\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.059385 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbb2e2de-55ef-4bc1-a8df-1428e18c664e-config-data\") pod \"cinder-scheduler-0\" (UID: \"fbb2e2de-55ef-4bc1-a8df-1428e18c664e\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.059479 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76hk8\" (UniqueName: \"kubernetes.io/projected/fbb2e2de-55ef-4bc1-a8df-1428e18c664e-kube-api-access-76hk8\") pod \"cinder-scheduler-0\" (UID: \"fbb2e2de-55ef-4bc1-a8df-1428e18c664e\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.059819 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fbb2e2de-55ef-4bc1-a8df-1428e18c664e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fbb2e2de-55ef-4bc1-a8df-1428e18c664e\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.064548 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbb2e2de-55ef-4bc1-a8df-1428e18c664e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fbb2e2de-55ef-4bc1-a8df-1428e18c664e\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.065665 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbb2e2de-55ef-4bc1-a8df-1428e18c664e-scripts\") pod \"cinder-scheduler-0\" (UID: \"fbb2e2de-55ef-4bc1-a8df-1428e18c664e\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.066277 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fbb2e2de-55ef-4bc1-a8df-1428e18c664e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fbb2e2de-55ef-4bc1-a8df-1428e18c664e\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.068532 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbb2e2de-55ef-4bc1-a8df-1428e18c664e-config-data\") pod \"cinder-scheduler-0\" (UID: \"fbb2e2de-55ef-4bc1-a8df-1428e18c664e\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.080979 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76hk8\" (UniqueName: \"kubernetes.io/projected/fbb2e2de-55ef-4bc1-a8df-1428e18c664e-kube-api-access-76hk8\") pod \"cinder-scheduler-0\" (UID: \"fbb2e2de-55ef-4bc1-a8df-1428e18c664e\") " pod="openstack/cinder-scheduler-0" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.131008 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.446765 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.447089 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.517631 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eac8a754-52b8-471c-bce1-85c449f801ba" path="/var/lib/kubelet/pods/eac8a754-52b8-471c-bce1-85c449f801ba/volumes" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.699954 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.723932 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5889fd87f8-6w269"] Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.725776 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.739548 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fbb2e2de-55ef-4bc1-a8df-1428e18c664e","Type":"ContainerStarted","Data":"f4e819dd2959c724a283ab4e048ba0f6257835e4e777dcf995a871a45509b8f7"} Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.776468 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5889fd87f8-6w269"] Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.880506 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r67xz\" (UniqueName: \"kubernetes.io/projected/96409976-b947-4ba5-af6e-eeb0ae687b0b-kube-api-access-r67xz\") pod \"placement-5889fd87f8-6w269\" (UID: \"96409976-b947-4ba5-af6e-eeb0ae687b0b\") " pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.880570 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96409976-b947-4ba5-af6e-eeb0ae687b0b-scripts\") pod \"placement-5889fd87f8-6w269\" (UID: \"96409976-b947-4ba5-af6e-eeb0ae687b0b\") " pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.880604 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/96409976-b947-4ba5-af6e-eeb0ae687b0b-public-tls-certs\") pod \"placement-5889fd87f8-6w269\" (UID: \"96409976-b947-4ba5-af6e-eeb0ae687b0b\") " pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.880656 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96409976-b947-4ba5-af6e-eeb0ae687b0b-combined-ca-bundle\") pod \"placement-5889fd87f8-6w269\" (UID: \"96409976-b947-4ba5-af6e-eeb0ae687b0b\") " pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.880715 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96409976-b947-4ba5-af6e-eeb0ae687b0b-config-data\") pod \"placement-5889fd87f8-6w269\" (UID: \"96409976-b947-4ba5-af6e-eeb0ae687b0b\") " pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.880730 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96409976-b947-4ba5-af6e-eeb0ae687b0b-logs\") pod \"placement-5889fd87f8-6w269\" (UID: \"96409976-b947-4ba5-af6e-eeb0ae687b0b\") " pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.880845 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/96409976-b947-4ba5-af6e-eeb0ae687b0b-internal-tls-certs\") pod \"placement-5889fd87f8-6w269\" (UID: \"96409976-b947-4ba5-af6e-eeb0ae687b0b\") " pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.982053 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/96409976-b947-4ba5-af6e-eeb0ae687b0b-internal-tls-certs\") pod \"placement-5889fd87f8-6w269\" (UID: \"96409976-b947-4ba5-af6e-eeb0ae687b0b\") " pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.982233 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r67xz\" (UniqueName: \"kubernetes.io/projected/96409976-b947-4ba5-af6e-eeb0ae687b0b-kube-api-access-r67xz\") pod \"placement-5889fd87f8-6w269\" (UID: \"96409976-b947-4ba5-af6e-eeb0ae687b0b\") " pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.982308 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96409976-b947-4ba5-af6e-eeb0ae687b0b-scripts\") pod \"placement-5889fd87f8-6w269\" (UID: \"96409976-b947-4ba5-af6e-eeb0ae687b0b\") " pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.982383 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/96409976-b947-4ba5-af6e-eeb0ae687b0b-public-tls-certs\") pod \"placement-5889fd87f8-6w269\" (UID: \"96409976-b947-4ba5-af6e-eeb0ae687b0b\") " pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.982529 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96409976-b947-4ba5-af6e-eeb0ae687b0b-combined-ca-bundle\") pod \"placement-5889fd87f8-6w269\" (UID: \"96409976-b947-4ba5-af6e-eeb0ae687b0b\") " pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.982597 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96409976-b947-4ba5-af6e-eeb0ae687b0b-config-data\") pod \"placement-5889fd87f8-6w269\" (UID: \"96409976-b947-4ba5-af6e-eeb0ae687b0b\") " pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.982623 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96409976-b947-4ba5-af6e-eeb0ae687b0b-logs\") pod \"placement-5889fd87f8-6w269\" (UID: \"96409976-b947-4ba5-af6e-eeb0ae687b0b\") " pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.983111 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96409976-b947-4ba5-af6e-eeb0ae687b0b-logs\") pod \"placement-5889fd87f8-6w269\" (UID: \"96409976-b947-4ba5-af6e-eeb0ae687b0b\") " pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.986697 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/96409976-b947-4ba5-af6e-eeb0ae687b0b-internal-tls-certs\") pod \"placement-5889fd87f8-6w269\" (UID: \"96409976-b947-4ba5-af6e-eeb0ae687b0b\") " pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.987396 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96409976-b947-4ba5-af6e-eeb0ae687b0b-combined-ca-bundle\") pod \"placement-5889fd87f8-6w269\" (UID: \"96409976-b947-4ba5-af6e-eeb0ae687b0b\") " pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:49 crc kubenswrapper[4950]: I0318 20:26:49.997832 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96409976-b947-4ba5-af6e-eeb0ae687b0b-scripts\") pod \"placement-5889fd87f8-6w269\" (UID: \"96409976-b947-4ba5-af6e-eeb0ae687b0b\") " pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:50 crc kubenswrapper[4950]: I0318 20:26:50.003343 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r67xz\" (UniqueName: \"kubernetes.io/projected/96409976-b947-4ba5-af6e-eeb0ae687b0b-kube-api-access-r67xz\") pod \"placement-5889fd87f8-6w269\" (UID: \"96409976-b947-4ba5-af6e-eeb0ae687b0b\") " pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:50 crc kubenswrapper[4950]: I0318 20:26:50.008857 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/96409976-b947-4ba5-af6e-eeb0ae687b0b-public-tls-certs\") pod \"placement-5889fd87f8-6w269\" (UID: \"96409976-b947-4ba5-af6e-eeb0ae687b0b\") " pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:50 crc kubenswrapper[4950]: I0318 20:26:50.009480 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96409976-b947-4ba5-af6e-eeb0ae687b0b-config-data\") pod \"placement-5889fd87f8-6w269\" (UID: \"96409976-b947-4ba5-af6e-eeb0ae687b0b\") " pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:50 crc kubenswrapper[4950]: I0318 20:26:50.052486 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:50 crc kubenswrapper[4950]: I0318 20:26:50.636948 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5889fd87f8-6w269"] Mar 18 20:26:50 crc kubenswrapper[4950]: I0318 20:26:50.774487 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fbb2e2de-55ef-4bc1-a8df-1428e18c664e","Type":"ContainerStarted","Data":"282eafe57068276d17e6ddcbbb5fa197465a81f4a8d3a12f6005fd9038132352"} Mar 18 20:26:50 crc kubenswrapper[4950]: I0318 20:26:50.781214 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5889fd87f8-6w269" event={"ID":"96409976-b947-4ba5-af6e-eeb0ae687b0b","Type":"ContainerStarted","Data":"849340deb8c77d8011857e14642d88188d7b5f63a4eab1ab0b3044a02d51216a"} Mar 18 20:26:51 crc kubenswrapper[4950]: I0318 20:26:51.324562 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" Mar 18 20:26:51 crc kubenswrapper[4950]: I0318 20:26:51.418491 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75bf4db767-zbs9v"] Mar 18 20:26:51 crc kubenswrapper[4950]: I0318 20:26:51.418992 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" podUID="a95776b7-4a12-4de5-b60f-02d008344485" containerName="dnsmasq-dns" containerID="cri-o://8c8dfa2860fb1e62b894f7d9111f01263d57835534226ef39a1bca9a7368ddba" gracePeriod=10 Mar 18 20:26:51 crc kubenswrapper[4950]: I0318 20:26:51.806677 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fbb2e2de-55ef-4bc1-a8df-1428e18c664e","Type":"ContainerStarted","Data":"edc3b6ffd7fc933e4a2bc8e6539f069069fdb9c44c3b71e8ff9d054b19677741"} Mar 18 20:26:51 crc kubenswrapper[4950]: I0318 20:26:51.819388 4950 generic.go:334] "Generic (PLEG): container finished" podID="a95776b7-4a12-4de5-b60f-02d008344485" containerID="8c8dfa2860fb1e62b894f7d9111f01263d57835534226ef39a1bca9a7368ddba" exitCode=0 Mar 18 20:26:51 crc kubenswrapper[4950]: I0318 20:26:51.819469 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" event={"ID":"a95776b7-4a12-4de5-b60f-02d008344485","Type":"ContainerDied","Data":"8c8dfa2860fb1e62b894f7d9111f01263d57835534226ef39a1bca9a7368ddba"} Mar 18 20:26:51 crc kubenswrapper[4950]: I0318 20:26:51.830499 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5889fd87f8-6w269" event={"ID":"96409976-b947-4ba5-af6e-eeb0ae687b0b","Type":"ContainerStarted","Data":"d5c7ea65ef7b3f2bed2b638641cad8f077ce7de916b5b95bedf8884884c35741"} Mar 18 20:26:51 crc kubenswrapper[4950]: I0318 20:26:51.830552 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5889fd87f8-6w269" event={"ID":"96409976-b947-4ba5-af6e-eeb0ae687b0b","Type":"ContainerStarted","Data":"bdf80484d7d57e60c8007caf11101bf41cebe8452ab0f1050ee0ed845203db95"} Mar 18 20:26:51 crc kubenswrapper[4950]: I0318 20:26:51.831340 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:51 crc kubenswrapper[4950]: I0318 20:26:51.831361 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:26:51 crc kubenswrapper[4950]: I0318 20:26:51.846751 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.8467331959999997 podStartE2EDuration="3.846733196s" podCreationTimestamp="2026-03-18 20:26:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:26:51.839916141 +0000 UTC m=+1225.080758019" watchObservedRunningTime="2026-03-18 20:26:51.846733196 +0000 UTC m=+1225.087575064" Mar 18 20:26:52 crc kubenswrapper[4950]: I0318 20:26:52.318402 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" Mar 18 20:26:52 crc kubenswrapper[4950]: I0318 20:26:52.342807 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5889fd87f8-6w269" podStartSLOduration=3.342789178 podStartE2EDuration="3.342789178s" podCreationTimestamp="2026-03-18 20:26:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:26:51.948048515 +0000 UTC m=+1225.188890383" watchObservedRunningTime="2026-03-18 20:26:52.342789178 +0000 UTC m=+1225.583631046" Mar 18 20:26:52 crc kubenswrapper[4950]: I0318 20:26:52.447802 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-dns-svc\") pod \"a95776b7-4a12-4de5-b60f-02d008344485\" (UID: \"a95776b7-4a12-4de5-b60f-02d008344485\") " Mar 18 20:26:52 crc kubenswrapper[4950]: I0318 20:26:52.447956 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-ovsdbserver-nb\") pod \"a95776b7-4a12-4de5-b60f-02d008344485\" (UID: \"a95776b7-4a12-4de5-b60f-02d008344485\") " Mar 18 20:26:52 crc kubenswrapper[4950]: I0318 20:26:52.448729 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-config\") pod \"a95776b7-4a12-4de5-b60f-02d008344485\" (UID: \"a95776b7-4a12-4de5-b60f-02d008344485\") " Mar 18 20:26:52 crc kubenswrapper[4950]: I0318 20:26:52.448773 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-ovsdbserver-sb\") pod \"a95776b7-4a12-4de5-b60f-02d008344485\" (UID: \"a95776b7-4a12-4de5-b60f-02d008344485\") " Mar 18 20:26:52 crc kubenswrapper[4950]: I0318 20:26:52.448796 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nh2qg\" (UniqueName: \"kubernetes.io/projected/a95776b7-4a12-4de5-b60f-02d008344485-kube-api-access-nh2qg\") pod \"a95776b7-4a12-4de5-b60f-02d008344485\" (UID: \"a95776b7-4a12-4de5-b60f-02d008344485\") " Mar 18 20:26:52 crc kubenswrapper[4950]: I0318 20:26:52.496966 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a95776b7-4a12-4de5-b60f-02d008344485-kube-api-access-nh2qg" (OuterVolumeSpecName: "kube-api-access-nh2qg") pod "a95776b7-4a12-4de5-b60f-02d008344485" (UID: "a95776b7-4a12-4de5-b60f-02d008344485"). InnerVolumeSpecName "kube-api-access-nh2qg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:26:52 crc kubenswrapper[4950]: I0318 20:26:52.555546 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nh2qg\" (UniqueName: \"kubernetes.io/projected/a95776b7-4a12-4de5-b60f-02d008344485-kube-api-access-nh2qg\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:52 crc kubenswrapper[4950]: I0318 20:26:52.572962 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a95776b7-4a12-4de5-b60f-02d008344485" (UID: "a95776b7-4a12-4de5-b60f-02d008344485"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:26:52 crc kubenswrapper[4950]: I0318 20:26:52.602847 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-config" (OuterVolumeSpecName: "config") pod "a95776b7-4a12-4de5-b60f-02d008344485" (UID: "a95776b7-4a12-4de5-b60f-02d008344485"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:26:52 crc kubenswrapper[4950]: I0318 20:26:52.602962 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a95776b7-4a12-4de5-b60f-02d008344485" (UID: "a95776b7-4a12-4de5-b60f-02d008344485"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:26:52 crc kubenswrapper[4950]: I0318 20:26:52.636238 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a95776b7-4a12-4de5-b60f-02d008344485" (UID: "a95776b7-4a12-4de5-b60f-02d008344485"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:26:52 crc kubenswrapper[4950]: I0318 20:26:52.656771 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:52 crc kubenswrapper[4950]: I0318 20:26:52.656807 4950 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:52 crc kubenswrapper[4950]: I0318 20:26:52.656847 4950 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:52 crc kubenswrapper[4950]: I0318 20:26:52.656857 4950 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a95776b7-4a12-4de5-b60f-02d008344485-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:52 crc kubenswrapper[4950]: I0318 20:26:52.839857 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" event={"ID":"a95776b7-4a12-4de5-b60f-02d008344485","Type":"ContainerDied","Data":"dc760a0ee02c4c6778cec4f216b5ae8418671bdb044ea749af5a31f372f92640"} Mar 18 20:26:52 crc kubenswrapper[4950]: I0318 20:26:52.839914 4950 scope.go:117] "RemoveContainer" containerID="8c8dfa2860fb1e62b894f7d9111f01263d57835534226ef39a1bca9a7368ddba" Mar 18 20:26:52 crc kubenswrapper[4950]: I0318 20:26:52.840090 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75bf4db767-zbs9v" Mar 18 20:26:52 crc kubenswrapper[4950]: I0318 20:26:52.879327 4950 scope.go:117] "RemoveContainer" containerID="a52ada126793a6eb000243037a35dc25cd43e2dfca4ac4a2872b35cc11e88ae8" Mar 18 20:26:52 crc kubenswrapper[4950]: I0318 20:26:52.910744 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75bf4db767-zbs9v"] Mar 18 20:26:52 crc kubenswrapper[4950]: I0318 20:26:52.917558 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75bf4db767-zbs9v"] Mar 18 20:26:53 crc kubenswrapper[4950]: I0318 20:26:53.490931 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a95776b7-4a12-4de5-b60f-02d008344485" path="/var/lib/kubelet/pods/a95776b7-4a12-4de5-b60f-02d008344485/volumes" Mar 18 20:26:54 crc kubenswrapper[4950]: I0318 20:26:54.132130 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 18 20:26:54 crc kubenswrapper[4950]: I0318 20:26:54.488222 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:26:54 crc kubenswrapper[4950]: I0318 20:26:54.815632 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="1bae85ae-0545-46ac-90fa-c8f967648ed4" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.153:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 18 20:26:55 crc kubenswrapper[4950]: I0318 20:26:55.586149 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Mar 18 20:26:55 crc kubenswrapper[4950]: I0318 20:26:55.661939 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:26:56 crc kubenswrapper[4950]: I0318 20:26:56.741389 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-564b994b87-rbnt8" Mar 18 20:26:57 crc kubenswrapper[4950]: I0318 20:26:57.277004 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-d5c6d4b4c-6cdzp" Mar 18 20:26:57 crc kubenswrapper[4950]: I0318 20:26:57.356495 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5d979b4d49-bzq8q"] Mar 18 20:26:57 crc kubenswrapper[4950]: I0318 20:26:57.356743 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5d979b4d49-bzq8q" podUID="31bbaba1-cd09-469b-9b89-f7ff2213cd8e" containerName="neutron-api" containerID="cri-o://0c993414a215e6519fba3c4c77718278e352a22c3d3bad0350254d5c89624b76" gracePeriod=30 Mar 18 20:26:57 crc kubenswrapper[4950]: I0318 20:26:57.357175 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5d979b4d49-bzq8q" podUID="31bbaba1-cd09-469b-9b89-f7ff2213cd8e" containerName="neutron-httpd" containerID="cri-o://8c5841219d263c4c634cb638c391cc932d8c4bf30892f6dc7b784816023de26e" gracePeriod=30 Mar 18 20:26:57 crc kubenswrapper[4950]: W0318 20:26:57.500023 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7563774c_0a5c_4af9_802a_662f4e721f3c.slice/crio-da6d7c641e3ff94fdefa562e2753b88297237917dc8d3b63b84a609a7f7d2450.scope WatchSource:0}: Error finding container da6d7c641e3ff94fdefa562e2753b88297237917dc8d3b63b84a609a7f7d2450: Status 404 returned error can't find the container with id da6d7c641e3ff94fdefa562e2753b88297237917dc8d3b63b84a609a7f7d2450 Mar 18 20:26:57 crc kubenswrapper[4950]: W0318 20:26:57.510565 4950 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeac8a754_52b8_471c_bce1_85c449f801ba.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeac8a754_52b8_471c_bce1_85c449f801ba.slice: no such file or directory Mar 18 20:26:57 crc kubenswrapper[4950]: W0318 20:26:57.510776 4950 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod292b3f8d_075a_4d1d_8aa9_56dd2c00943c.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod292b3f8d_075a_4d1d_8aa9_56dd2c00943c.slice: no such file or directory Mar 18 20:26:57 crc kubenswrapper[4950]: W0318 20:26:57.510795 4950 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ae778b5_53f6_4fbf_b779_83a194d4da7d.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ae778b5_53f6_4fbf_b779_83a194d4da7d.slice: no such file or directory Mar 18 20:26:57 crc kubenswrapper[4950]: I0318 20:26:57.932445 4950 generic.go:334] "Generic (PLEG): container finished" podID="31bbaba1-cd09-469b-9b89-f7ff2213cd8e" containerID="8c5841219d263c4c634cb638c391cc932d8c4bf30892f6dc7b784816023de26e" exitCode=0 Mar 18 20:26:57 crc kubenswrapper[4950]: I0318 20:26:57.932514 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d979b4d49-bzq8q" event={"ID":"31bbaba1-cd09-469b-9b89-f7ff2213cd8e","Type":"ContainerDied","Data":"8c5841219d263c4c634cb638c391cc932d8c4bf30892f6dc7b784816023de26e"} Mar 18 20:26:57 crc kubenswrapper[4950]: I0318 20:26:57.938593 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-c888c6fcb-jnt6f_ed97e535-9755-40ef-a82f-176f0eb7e31b/neutron-api/0.log" Mar 18 20:26:57 crc kubenswrapper[4950]: I0318 20:26:57.938637 4950 generic.go:334] "Generic (PLEG): container finished" podID="ed97e535-9755-40ef-a82f-176f0eb7e31b" containerID="74f0ac325be8ead0f0e1eaebbcab54a265d706045bb6cc501683c742e2f9b28f" exitCode=137 Mar 18 20:26:57 crc kubenswrapper[4950]: I0318 20:26:57.938668 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c888c6fcb-jnt6f" event={"ID":"ed97e535-9755-40ef-a82f-176f0eb7e31b","Type":"ContainerDied","Data":"74f0ac325be8ead0f0e1eaebbcab54a265d706045bb6cc501683c742e2f9b28f"} Mar 18 20:26:58 crc kubenswrapper[4950]: I0318 20:26:58.018985 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:58 crc kubenswrapper[4950]: I0318 20:26:58.229554 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-c888c6fcb-jnt6f_ed97e535-9755-40ef-a82f-176f0eb7e31b/neutron-api/0.log" Mar 18 20:26:58 crc kubenswrapper[4950]: I0318 20:26:58.229913 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c888c6fcb-jnt6f" Mar 18 20:26:58 crc kubenswrapper[4950]: I0318 20:26:58.255886 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-config\") pod \"ed97e535-9755-40ef-a82f-176f0eb7e31b\" (UID: \"ed97e535-9755-40ef-a82f-176f0eb7e31b\") " Mar 18 20:26:58 crc kubenswrapper[4950]: I0318 20:26:58.255965 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ftgk\" (UniqueName: \"kubernetes.io/projected/ed97e535-9755-40ef-a82f-176f0eb7e31b-kube-api-access-6ftgk\") pod \"ed97e535-9755-40ef-a82f-176f0eb7e31b\" (UID: \"ed97e535-9755-40ef-a82f-176f0eb7e31b\") " Mar 18 20:26:58 crc kubenswrapper[4950]: I0318 20:26:58.256057 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-httpd-config\") pod \"ed97e535-9755-40ef-a82f-176f0eb7e31b\" (UID: \"ed97e535-9755-40ef-a82f-176f0eb7e31b\") " Mar 18 20:26:58 crc kubenswrapper[4950]: I0318 20:26:58.256150 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-ovndb-tls-certs\") pod \"ed97e535-9755-40ef-a82f-176f0eb7e31b\" (UID: \"ed97e535-9755-40ef-a82f-176f0eb7e31b\") " Mar 18 20:26:58 crc kubenswrapper[4950]: I0318 20:26:58.256244 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-combined-ca-bundle\") pod \"ed97e535-9755-40ef-a82f-176f0eb7e31b\" (UID: \"ed97e535-9755-40ef-a82f-176f0eb7e31b\") " Mar 18 20:26:58 crc kubenswrapper[4950]: I0318 20:26:58.276785 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "ed97e535-9755-40ef-a82f-176f0eb7e31b" (UID: "ed97e535-9755-40ef-a82f-176f0eb7e31b"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:58 crc kubenswrapper[4950]: I0318 20:26:58.288651 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed97e535-9755-40ef-a82f-176f0eb7e31b-kube-api-access-6ftgk" (OuterVolumeSpecName: "kube-api-access-6ftgk") pod "ed97e535-9755-40ef-a82f-176f0eb7e31b" (UID: "ed97e535-9755-40ef-a82f-176f0eb7e31b"). InnerVolumeSpecName "kube-api-access-6ftgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:26:58 crc kubenswrapper[4950]: I0318 20:26:58.359474 4950 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:58 crc kubenswrapper[4950]: I0318 20:26:58.359521 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ftgk\" (UniqueName: \"kubernetes.io/projected/ed97e535-9755-40ef-a82f-176f0eb7e31b-kube-api-access-6ftgk\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:58 crc kubenswrapper[4950]: I0318 20:26:58.364791 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed97e535-9755-40ef-a82f-176f0eb7e31b" (UID: "ed97e535-9755-40ef-a82f-176f0eb7e31b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:58 crc kubenswrapper[4950]: I0318 20:26:58.389679 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-config" (OuterVolumeSpecName: "config") pod "ed97e535-9755-40ef-a82f-176f0eb7e31b" (UID: "ed97e535-9755-40ef-a82f-176f0eb7e31b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:58 crc kubenswrapper[4950]: I0318 20:26:58.414132 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "ed97e535-9755-40ef-a82f-176f0eb7e31b" (UID: "ed97e535-9755-40ef-a82f-176f0eb7e31b"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:26:58 crc kubenswrapper[4950]: I0318 20:26:58.461374 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:58 crc kubenswrapper[4950]: I0318 20:26:58.461400 4950 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:58 crc kubenswrapper[4950]: I0318 20:26:58.461423 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed97e535-9755-40ef-a82f-176f0eb7e31b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:26:58 crc kubenswrapper[4950]: I0318 20:26:58.981475 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-c888c6fcb-jnt6f_ed97e535-9755-40ef-a82f-176f0eb7e31b/neutron-api/0.log" Mar 18 20:26:58 crc kubenswrapper[4950]: I0318 20:26:58.981534 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c888c6fcb-jnt6f" event={"ID":"ed97e535-9755-40ef-a82f-176f0eb7e31b","Type":"ContainerDied","Data":"901e81991ebfa5935dfb7f6c7620365fc8bb2119a1f1e2e42f586862cab0a475"} Mar 18 20:26:58 crc kubenswrapper[4950]: I0318 20:26:58.981569 4950 scope.go:117] "RemoveContainer" containerID="e3df6aff85a9f7990f7a5c74c826e0e7d768d6648c7249bf7e4b60b739cb95f6" Mar 18 20:26:58 crc kubenswrapper[4950]: I0318 20:26:58.981690 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c888c6fcb-jnt6f" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.028008 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-c888c6fcb-jnt6f"] Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.033245 4950 scope.go:117] "RemoveContainer" containerID="74f0ac325be8ead0f0e1eaebbcab54a265d706045bb6cc501683c742e2f9b28f" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.033937 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-c888c6fcb-jnt6f"] Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.206585 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-776bcc8684-8mp9k" podUID="33fe46ee-d1cc-4d44-b6c9-8f9e808086d1" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.157:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.244966 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-776bcc8684-8mp9k" podUID="33fe46ee-d1cc-4d44-b6c9-8f9e808086d1" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.157:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.245130 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7d5775f6d4-lshhl" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.364157 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-776bcc8684-8mp9k"] Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.364374 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-776bcc8684-8mp9k" podUID="33fe46ee-d1cc-4d44-b6c9-8f9e808086d1" containerName="barbican-api-log" containerID="cri-o://896addbe1c7c1e9be3a785950f100544eb21fb5c2eb42a17124da60c933c03be" gracePeriod=30 Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.364769 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-776bcc8684-8mp9k" podUID="33fe46ee-d1cc-4d44-b6c9-8f9e808086d1" containerName="barbican-api" containerID="cri-o://9584a8936196d9827442cd1eb57252fcb4c3fee9630c39e28cb477cca30362be" gracePeriod=30 Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.409626 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-776bcc8684-8mp9k" podUID="33fe46ee-d1cc-4d44-b6c9-8f9e808086d1" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.157:9311/healthcheck\": EOF" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.409806 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-776bcc8684-8mp9k" podUID="33fe46ee-d1cc-4d44-b6c9-8f9e808086d1" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.157:9311/healthcheck\": EOF" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.515097 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed97e535-9755-40ef-a82f-176f0eb7e31b" path="/var/lib/kubelet/pods/ed97e535-9755-40ef-a82f-176f0eb7e31b/volumes" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.686159 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 18 20:26:59 crc kubenswrapper[4950]: E0318 20:26:59.686517 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed97e535-9755-40ef-a82f-176f0eb7e31b" containerName="neutron-api" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.686532 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed97e535-9755-40ef-a82f-176f0eb7e31b" containerName="neutron-api" Mar 18 20:26:59 crc kubenswrapper[4950]: E0318 20:26:59.686549 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed97e535-9755-40ef-a82f-176f0eb7e31b" containerName="neutron-httpd" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.686556 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed97e535-9755-40ef-a82f-176f0eb7e31b" containerName="neutron-httpd" Mar 18 20:26:59 crc kubenswrapper[4950]: E0318 20:26:59.686577 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a95776b7-4a12-4de5-b60f-02d008344485" containerName="init" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.686583 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="a95776b7-4a12-4de5-b60f-02d008344485" containerName="init" Mar 18 20:26:59 crc kubenswrapper[4950]: E0318 20:26:59.686597 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a95776b7-4a12-4de5-b60f-02d008344485" containerName="dnsmasq-dns" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.686603 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="a95776b7-4a12-4de5-b60f-02d008344485" containerName="dnsmasq-dns" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.686764 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed97e535-9755-40ef-a82f-176f0eb7e31b" containerName="neutron-api" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.686795 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed97e535-9755-40ef-a82f-176f0eb7e31b" containerName="neutron-httpd" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.686810 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="a95776b7-4a12-4de5-b60f-02d008344485" containerName="dnsmasq-dns" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.690853 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.693720 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-f5c5r" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.694011 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.694183 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.699539 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.822025 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e0941268-2503-4b8e-afeb-0a655b52f505-openstack-config-secret\") pod \"openstackclient\" (UID: \"e0941268-2503-4b8e-afeb-0a655b52f505\") " pod="openstack/openstackclient" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.822074 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s69x5\" (UniqueName: \"kubernetes.io/projected/e0941268-2503-4b8e-afeb-0a655b52f505-kube-api-access-s69x5\") pod \"openstackclient\" (UID: \"e0941268-2503-4b8e-afeb-0a655b52f505\") " pod="openstack/openstackclient" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.822118 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e0941268-2503-4b8e-afeb-0a655b52f505-openstack-config\") pod \"openstackclient\" (UID: \"e0941268-2503-4b8e-afeb-0a655b52f505\") " pod="openstack/openstackclient" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.822140 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0941268-2503-4b8e-afeb-0a655b52f505-combined-ca-bundle\") pod \"openstackclient\" (UID: \"e0941268-2503-4b8e-afeb-0a655b52f505\") " pod="openstack/openstackclient" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.823083 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.924228 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e0941268-2503-4b8e-afeb-0a655b52f505-openstack-config-secret\") pod \"openstackclient\" (UID: \"e0941268-2503-4b8e-afeb-0a655b52f505\") " pod="openstack/openstackclient" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.924302 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s69x5\" (UniqueName: \"kubernetes.io/projected/e0941268-2503-4b8e-afeb-0a655b52f505-kube-api-access-s69x5\") pod \"openstackclient\" (UID: \"e0941268-2503-4b8e-afeb-0a655b52f505\") " pod="openstack/openstackclient" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.924365 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e0941268-2503-4b8e-afeb-0a655b52f505-openstack-config\") pod \"openstackclient\" (UID: \"e0941268-2503-4b8e-afeb-0a655b52f505\") " pod="openstack/openstackclient" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.924383 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0941268-2503-4b8e-afeb-0a655b52f505-combined-ca-bundle\") pod \"openstackclient\" (UID: \"e0941268-2503-4b8e-afeb-0a655b52f505\") " pod="openstack/openstackclient" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.925605 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e0941268-2503-4b8e-afeb-0a655b52f505-openstack-config\") pod \"openstackclient\" (UID: \"e0941268-2503-4b8e-afeb-0a655b52f505\") " pod="openstack/openstackclient" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.932880 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e0941268-2503-4b8e-afeb-0a655b52f505-openstack-config-secret\") pod \"openstackclient\" (UID: \"e0941268-2503-4b8e-afeb-0a655b52f505\") " pod="openstack/openstackclient" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.951543 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0941268-2503-4b8e-afeb-0a655b52f505-combined-ca-bundle\") pod \"openstackclient\" (UID: \"e0941268-2503-4b8e-afeb-0a655b52f505\") " pod="openstack/openstackclient" Mar 18 20:26:59 crc kubenswrapper[4950]: I0318 20:26:59.955350 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s69x5\" (UniqueName: \"kubernetes.io/projected/e0941268-2503-4b8e-afeb-0a655b52f505-kube-api-access-s69x5\") pod \"openstackclient\" (UID: \"e0941268-2503-4b8e-afeb-0a655b52f505\") " pod="openstack/openstackclient" Mar 18 20:27:00 crc kubenswrapper[4950]: I0318 20:27:00.008285 4950 generic.go:334] "Generic (PLEG): container finished" podID="33fe46ee-d1cc-4d44-b6c9-8f9e808086d1" containerID="896addbe1c7c1e9be3a785950f100544eb21fb5c2eb42a17124da60c933c03be" exitCode=143 Mar 18 20:27:00 crc kubenswrapper[4950]: I0318 20:27:00.008556 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-776bcc8684-8mp9k" event={"ID":"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1","Type":"ContainerDied","Data":"896addbe1c7c1e9be3a785950f100544eb21fb5c2eb42a17124da60c933c03be"} Mar 18 20:27:00 crc kubenswrapper[4950]: I0318 20:27:00.008597 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 18 20:27:00 crc kubenswrapper[4950]: I0318 20:27:00.536731 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 18 20:27:01 crc kubenswrapper[4950]: I0318 20:27:01.031269 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"e0941268-2503-4b8e-afeb-0a655b52f505","Type":"ContainerStarted","Data":"4526ec917cf8e7c6999893cb9f1d4552362108eb89c2314cfc1c4496847ea4c6"} Mar 18 20:27:03 crc kubenswrapper[4950]: E0318 20:27:03.885521 4950 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7563774c_0a5c_4af9_802a_662f4e721f3c.slice/crio-conmon-e609f6cce6ca82ac4a1ba760e29aadc780269f67316528034724eda1d330732d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7563774c_0a5c_4af9_802a_662f4e721f3c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7563774c_0a5c_4af9_802a_662f4e721f3c.slice/crio-conmon-da6d7c641e3ff94fdefa562e2753b88297237917dc8d3b63b84a609a7f7d2450.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82a7c2a4_9f57_460e_8bac_419f14222502.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7563774c_0a5c_4af9_802a_662f4e721f3c.slice/crio-a94ea562bd7c3e520d36e8bc6b0f62fe4b690bcbac96af7089b92400bea93599.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad591d5a_4120_4285_9cca_417b19176795.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4bb9e1a_0694_4e00_8ba6_f2765f82b983.slice/crio-5d841f1711b2d20eb9beeb857c1703a9d8c135fead221b2e15a23a8700b18a1a\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82a7c2a4_9f57_460e_8bac_419f14222502.slice/crio-97a26f08854fc0dfaabc62754c4b6d2acb95a35796f8ece839e9b7c0bd98db5a\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad591d5a_4120_4285_9cca_417b19176795.slice/crio-8c293bb8351af4475bcd2918ff2d3a397db802e08287df076b7ab111bbf95677\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4bb9e1a_0694_4e00_8ba6_f2765f82b983.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4bb9e1a_0694_4e00_8ba6_f2765f82b983.slice/crio-conmon-de9ab05a4141d7f72e470b7a98a89d526812c15d9a4174be30b8a51cc0d0d4b3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82a7c2a4_9f57_460e_8bac_419f14222502.slice/crio-conmon-08cf42bf1bb71302f9083c4aa0908a4cd8aed1c546a9aa614f9927b0b0ad5bf8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7563774c_0a5c_4af9_802a_662f4e721f3c.slice/crio-conmon-a94ea562bd7c3e520d36e8bc6b0f62fe4b690bcbac96af7089b92400bea93599.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7563774c_0a5c_4af9_802a_662f4e721f3c.slice/crio-de1b4d05b6433c8bacfa4f5808c26a249120f01056e842ff8ef61a2c85cdcfad.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4bb9e1a_0694_4e00_8ba6_f2765f82b983.slice/crio-de9ab05a4141d7f72e470b7a98a89d526812c15d9a4174be30b8a51cc0d0d4b3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7563774c_0a5c_4af9_802a_662f4e721f3c.slice/crio-conmon-de1b4d05b6433c8bacfa4f5808c26a249120f01056e842ff8ef61a2c85cdcfad.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7563774c_0a5c_4af9_802a_662f4e721f3c.slice/crio-7cbf38b3da364f13e1adf82ec0026977bf82c9432979f7f2f5fb3080546ffc16\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7563774c_0a5c_4af9_802a_662f4e721f3c.slice/crio-e609f6cce6ca82ac4a1ba760e29aadc780269f67316528034724eda1d330732d.scope\": RecentStats: unable to find data in memory cache]" Mar 18 20:27:04 crc kubenswrapper[4950]: I0318 20:27:04.055587 4950 generic.go:334] "Generic (PLEG): container finished" podID="31bbaba1-cd09-469b-9b89-f7ff2213cd8e" containerID="0c993414a215e6519fba3c4c77718278e352a22c3d3bad0350254d5c89624b76" exitCode=0 Mar 18 20:27:04 crc kubenswrapper[4950]: I0318 20:27:04.055641 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d979b4d49-bzq8q" event={"ID":"31bbaba1-cd09-469b-9b89-f7ff2213cd8e","Type":"ContainerDied","Data":"0c993414a215e6519fba3c4c77718278e352a22c3d3bad0350254d5c89624b76"} Mar 18 20:27:04 crc kubenswrapper[4950]: I0318 20:27:04.468713 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-776bcc8684-8mp9k" podUID="33fe46ee-d1cc-4d44-b6c9-8f9e808086d1" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.157:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 18 20:27:04 crc kubenswrapper[4950]: I0318 20:27:04.722708 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:27:04 crc kubenswrapper[4950]: I0318 20:27:04.848314 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-public-tls-certs\") pod \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " Mar 18 20:27:04 crc kubenswrapper[4950]: I0318 20:27:04.848425 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-combined-ca-bundle\") pod \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " Mar 18 20:27:04 crc kubenswrapper[4950]: I0318 20:27:04.848498 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-ovndb-tls-certs\") pod \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " Mar 18 20:27:04 crc kubenswrapper[4950]: I0318 20:27:04.848522 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-internal-tls-certs\") pod \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " Mar 18 20:27:04 crc kubenswrapper[4950]: I0318 20:27:04.848540 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcj7f\" (UniqueName: \"kubernetes.io/projected/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-kube-api-access-vcj7f\") pod \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " Mar 18 20:27:04 crc kubenswrapper[4950]: I0318 20:27:04.848584 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-httpd-config\") pod \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " Mar 18 20:27:04 crc kubenswrapper[4950]: I0318 20:27:04.848623 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-config\") pod \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\" (UID: \"31bbaba1-cd09-469b-9b89-f7ff2213cd8e\") " Mar 18 20:27:04 crc kubenswrapper[4950]: I0318 20:27:04.857609 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-kube-api-access-vcj7f" (OuterVolumeSpecName: "kube-api-access-vcj7f") pod "31bbaba1-cd09-469b-9b89-f7ff2213cd8e" (UID: "31bbaba1-cd09-469b-9b89-f7ff2213cd8e"). InnerVolumeSpecName "kube-api-access-vcj7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:27:04 crc kubenswrapper[4950]: I0318 20:27:04.897587 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "31bbaba1-cd09-469b-9b89-f7ff2213cd8e" (UID: "31bbaba1-cd09-469b-9b89-f7ff2213cd8e"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:04 crc kubenswrapper[4950]: I0318 20:27:04.950162 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcj7f\" (UniqueName: \"kubernetes.io/projected/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-kube-api-access-vcj7f\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:04 crc kubenswrapper[4950]: I0318 20:27:04.950193 4950 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.021272 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "31bbaba1-cd09-469b-9b89-f7ff2213cd8e" (UID: "31bbaba1-cd09-469b-9b89-f7ff2213cd8e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.021555 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "31bbaba1-cd09-469b-9b89-f7ff2213cd8e" (UID: "31bbaba1-cd09-469b-9b89-f7ff2213cd8e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.047858 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "31bbaba1-cd09-469b-9b89-f7ff2213cd8e" (UID: "31bbaba1-cd09-469b-9b89-f7ff2213cd8e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.051398 4950 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.051511 4950 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.051567 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.062555 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "31bbaba1-cd09-469b-9b89-f7ff2213cd8e" (UID: "31bbaba1-cd09-469b-9b89-f7ff2213cd8e"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.098917 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d979b4d49-bzq8q" event={"ID":"31bbaba1-cd09-469b-9b89-f7ff2213cd8e","Type":"ContainerDied","Data":"4883a046fb74a10d04787d9d53f3adeb5aadbda6f5aea920fd78682a805ac980"} Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.098962 4950 scope.go:117] "RemoveContainer" containerID="8c5841219d263c4c634cb638c391cc932d8c4bf30892f6dc7b784816023de26e" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.099070 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d979b4d49-bzq8q" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.119622 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-config" (OuterVolumeSpecName: "config") pod "31bbaba1-cd09-469b-9b89-f7ff2213cd8e" (UID: "31bbaba1-cd09-469b-9b89-f7ff2213cd8e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.152934 4950 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.152964 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/31bbaba1-cd09-469b-9b89-f7ff2213cd8e-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.153007 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-776bcc8684-8mp9k" podUID="33fe46ee-d1cc-4d44-b6c9-8f9e808086d1" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.157:9311/healthcheck\": read tcp 10.217.0.2:43310->10.217.0.157:9311: read: connection reset by peer" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.153271 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-776bcc8684-8mp9k" podUID="33fe46ee-d1cc-4d44-b6c9-8f9e808086d1" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.157:9311/healthcheck\": read tcp 10.217.0.2:58536->10.217.0.157:9311: read: connection reset by peer" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.153370 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.154317 4950 scope.go:117] "RemoveContainer" containerID="0c993414a215e6519fba3c4c77718278e352a22c3d3bad0350254d5c89624b76" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.447167 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5d979b4d49-bzq8q"] Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.456040 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5d979b4d49-bzq8q"] Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.520184 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31bbaba1-cd09-469b-9b89-f7ff2213cd8e" path="/var/lib/kubelet/pods/31bbaba1-cd09-469b-9b89-f7ff2213cd8e/volumes" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.556300 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.659779 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-logs\") pod \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\" (UID: \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\") " Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.659857 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-config-data-custom\") pod \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\" (UID: \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\") " Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.659928 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8s2z2\" (UniqueName: \"kubernetes.io/projected/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-kube-api-access-8s2z2\") pod \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\" (UID: \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\") " Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.660039 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-config-data\") pod \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\" (UID: \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\") " Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.660112 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-combined-ca-bundle\") pod \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\" (UID: \"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1\") " Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.661757 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-logs" (OuterVolumeSpecName: "logs") pod "33fe46ee-d1cc-4d44-b6c9-8f9e808086d1" (UID: "33fe46ee-d1cc-4d44-b6c9-8f9e808086d1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.677724 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-kube-api-access-8s2z2" (OuterVolumeSpecName: "kube-api-access-8s2z2") pod "33fe46ee-d1cc-4d44-b6c9-8f9e808086d1" (UID: "33fe46ee-d1cc-4d44-b6c9-8f9e808086d1"). InnerVolumeSpecName "kube-api-access-8s2z2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.681617 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "33fe46ee-d1cc-4d44-b6c9-8f9e808086d1" (UID: "33fe46ee-d1cc-4d44-b6c9-8f9e808086d1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.688551 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33fe46ee-d1cc-4d44-b6c9-8f9e808086d1" (UID: "33fe46ee-d1cc-4d44-b6c9-8f9e808086d1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.712242 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-config-data" (OuterVolumeSpecName: "config-data") pod "33fe46ee-d1cc-4d44-b6c9-8f9e808086d1" (UID: "33fe46ee-d1cc-4d44-b6c9-8f9e808086d1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.762229 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.762287 4950 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-logs\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.762299 4950 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.762308 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8s2z2\" (UniqueName: \"kubernetes.io/projected/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-kube-api-access-8s2z2\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:05 crc kubenswrapper[4950]: I0318 20:27:05.762318 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:06 crc kubenswrapper[4950]: I0318 20:27:06.117054 4950 generic.go:334] "Generic (PLEG): container finished" podID="33fe46ee-d1cc-4d44-b6c9-8f9e808086d1" containerID="9584a8936196d9827442cd1eb57252fcb4c3fee9630c39e28cb477cca30362be" exitCode=0 Mar 18 20:27:06 crc kubenswrapper[4950]: I0318 20:27:06.117116 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-776bcc8684-8mp9k" Mar 18 20:27:06 crc kubenswrapper[4950]: I0318 20:27:06.117142 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-776bcc8684-8mp9k" event={"ID":"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1","Type":"ContainerDied","Data":"9584a8936196d9827442cd1eb57252fcb4c3fee9630c39e28cb477cca30362be"} Mar 18 20:27:06 crc kubenswrapper[4950]: I0318 20:27:06.117641 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-776bcc8684-8mp9k" event={"ID":"33fe46ee-d1cc-4d44-b6c9-8f9e808086d1","Type":"ContainerDied","Data":"7856acbf054d957f91306906724b142712d6e291d952913bd62b0223a8e1298c"} Mar 18 20:27:06 crc kubenswrapper[4950]: I0318 20:27:06.117669 4950 scope.go:117] "RemoveContainer" containerID="9584a8936196d9827442cd1eb57252fcb4c3fee9630c39e28cb477cca30362be" Mar 18 20:27:06 crc kubenswrapper[4950]: I0318 20:27:06.159173 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-776bcc8684-8mp9k"] Mar 18 20:27:06 crc kubenswrapper[4950]: I0318 20:27:06.164992 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-776bcc8684-8mp9k"] Mar 18 20:27:06 crc kubenswrapper[4950]: I0318 20:27:06.166531 4950 scope.go:117] "RemoveContainer" containerID="896addbe1c7c1e9be3a785950f100544eb21fb5c2eb42a17124da60c933c03be" Mar 18 20:27:06 crc kubenswrapper[4950]: I0318 20:27:06.185609 4950 scope.go:117] "RemoveContainer" containerID="9584a8936196d9827442cd1eb57252fcb4c3fee9630c39e28cb477cca30362be" Mar 18 20:27:06 crc kubenswrapper[4950]: E0318 20:27:06.186115 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9584a8936196d9827442cd1eb57252fcb4c3fee9630c39e28cb477cca30362be\": container with ID starting with 9584a8936196d9827442cd1eb57252fcb4c3fee9630c39e28cb477cca30362be not found: ID does not exist" containerID="9584a8936196d9827442cd1eb57252fcb4c3fee9630c39e28cb477cca30362be" Mar 18 20:27:06 crc kubenswrapper[4950]: I0318 20:27:06.186170 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9584a8936196d9827442cd1eb57252fcb4c3fee9630c39e28cb477cca30362be"} err="failed to get container status \"9584a8936196d9827442cd1eb57252fcb4c3fee9630c39e28cb477cca30362be\": rpc error: code = NotFound desc = could not find container \"9584a8936196d9827442cd1eb57252fcb4c3fee9630c39e28cb477cca30362be\": container with ID starting with 9584a8936196d9827442cd1eb57252fcb4c3fee9630c39e28cb477cca30362be not found: ID does not exist" Mar 18 20:27:06 crc kubenswrapper[4950]: I0318 20:27:06.186210 4950 scope.go:117] "RemoveContainer" containerID="896addbe1c7c1e9be3a785950f100544eb21fb5c2eb42a17124da60c933c03be" Mar 18 20:27:06 crc kubenswrapper[4950]: E0318 20:27:06.186551 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"896addbe1c7c1e9be3a785950f100544eb21fb5c2eb42a17124da60c933c03be\": container with ID starting with 896addbe1c7c1e9be3a785950f100544eb21fb5c2eb42a17124da60c933c03be not found: ID does not exist" containerID="896addbe1c7c1e9be3a785950f100544eb21fb5c2eb42a17124da60c933c03be" Mar 18 20:27:06 crc kubenswrapper[4950]: I0318 20:27:06.186587 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"896addbe1c7c1e9be3a785950f100544eb21fb5c2eb42a17124da60c933c03be"} err="failed to get container status \"896addbe1c7c1e9be3a785950f100544eb21fb5c2eb42a17124da60c933c03be\": rpc error: code = NotFound desc = could not find container \"896addbe1c7c1e9be3a785950f100544eb21fb5c2eb42a17124da60c933c03be\": container with ID starting with 896addbe1c7c1e9be3a785950f100544eb21fb5c2eb42a17124da60c933c03be not found: ID does not exist" Mar 18 20:27:06 crc kubenswrapper[4950]: I0318 20:27:06.970036 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 18 20:27:07 crc kubenswrapper[4950]: I0318 20:27:07.491083 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33fe46ee-d1cc-4d44-b6c9-8f9e808086d1" path="/var/lib/kubelet/pods/33fe46ee-d1cc-4d44-b6c9-8f9e808086d1/volumes" Mar 18 20:27:13 crc kubenswrapper[4950]: I0318 20:27:13.091641 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:27:13 crc kubenswrapper[4950]: I0318 20:27:13.092307 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3a00a076-ca9d-46e9-b3e7-af13dcaca76e" containerName="ceilometer-central-agent" containerID="cri-o://15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca" gracePeriod=30 Mar 18 20:27:13 crc kubenswrapper[4950]: I0318 20:27:13.092433 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3a00a076-ca9d-46e9-b3e7-af13dcaca76e" containerName="ceilometer-notification-agent" containerID="cri-o://c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc" gracePeriod=30 Mar 18 20:27:13 crc kubenswrapper[4950]: I0318 20:27:13.092429 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3a00a076-ca9d-46e9-b3e7-af13dcaca76e" containerName="sg-core" containerID="cri-o://18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508" gracePeriod=30 Mar 18 20:27:13 crc kubenswrapper[4950]: I0318 20:27:13.092466 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3a00a076-ca9d-46e9-b3e7-af13dcaca76e" containerName="proxy-httpd" containerID="cri-o://7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891" gracePeriod=30 Mar 18 20:27:13 crc kubenswrapper[4950]: I0318 20:27:13.188038 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"e0941268-2503-4b8e-afeb-0a655b52f505","Type":"ContainerStarted","Data":"d5fb5fe56f8e389fa1a4f16889ce2796fb6aaf8aa9793bfd701cc5e5080f0e27"} Mar 18 20:27:13 crc kubenswrapper[4950]: I0318 20:27:13.209025 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.042237999 podStartE2EDuration="14.209002876s" podCreationTimestamp="2026-03-18 20:26:59 +0000 UTC" firstStartedPulling="2026-03-18 20:27:00.551804011 +0000 UTC m=+1233.792645879" lastFinishedPulling="2026-03-18 20:27:12.718568878 +0000 UTC m=+1245.959410756" observedRunningTime="2026-03-18 20:27:13.200767574 +0000 UTC m=+1246.441609442" watchObservedRunningTime="2026-03-18 20:27:13.209002876 +0000 UTC m=+1246.449844754" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.181587 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.197305 4950 generic.go:334] "Generic (PLEG): container finished" podID="3a00a076-ca9d-46e9-b3e7-af13dcaca76e" containerID="7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891" exitCode=0 Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.197332 4950 generic.go:334] "Generic (PLEG): container finished" podID="3a00a076-ca9d-46e9-b3e7-af13dcaca76e" containerID="18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508" exitCode=2 Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.197341 4950 generic.go:334] "Generic (PLEG): container finished" podID="3a00a076-ca9d-46e9-b3e7-af13dcaca76e" containerID="c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc" exitCode=0 Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.197348 4950 generic.go:334] "Generic (PLEG): container finished" podID="3a00a076-ca9d-46e9-b3e7-af13dcaca76e" containerID="15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca" exitCode=0 Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.197377 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.197380 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a00a076-ca9d-46e9-b3e7-af13dcaca76e","Type":"ContainerDied","Data":"7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891"} Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.197439 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a00a076-ca9d-46e9-b3e7-af13dcaca76e","Type":"ContainerDied","Data":"18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508"} Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.197454 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a00a076-ca9d-46e9-b3e7-af13dcaca76e","Type":"ContainerDied","Data":"c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc"} Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.197464 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a00a076-ca9d-46e9-b3e7-af13dcaca76e","Type":"ContainerDied","Data":"15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca"} Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.197474 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a00a076-ca9d-46e9-b3e7-af13dcaca76e","Type":"ContainerDied","Data":"de71a314bd6be6cf438243681b817c44b5906f260c505a93a236ea484a733d7e"} Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.197490 4950 scope.go:117] "RemoveContainer" containerID="7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.216823 4950 scope.go:117] "RemoveContainer" containerID="18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.270292 4950 scope.go:117] "RemoveContainer" containerID="c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.294018 4950 scope.go:117] "RemoveContainer" containerID="15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.333914 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-config-data\") pod \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.333981 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-run-httpd\") pod \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.334103 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sv7h\" (UniqueName: \"kubernetes.io/projected/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-kube-api-access-6sv7h\") pod \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.334120 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-sg-core-conf-yaml\") pod \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.334141 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-combined-ca-bundle\") pod \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.334157 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-scripts\") pod \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.334171 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-log-httpd\") pod \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\" (UID: \"3a00a076-ca9d-46e9-b3e7-af13dcaca76e\") " Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.339098 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3a00a076-ca9d-46e9-b3e7-af13dcaca76e" (UID: "3a00a076-ca9d-46e9-b3e7-af13dcaca76e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.346102 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3a00a076-ca9d-46e9-b3e7-af13dcaca76e" (UID: "3a00a076-ca9d-46e9-b3e7-af13dcaca76e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.361635 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-kube-api-access-6sv7h" (OuterVolumeSpecName: "kube-api-access-6sv7h") pod "3a00a076-ca9d-46e9-b3e7-af13dcaca76e" (UID: "3a00a076-ca9d-46e9-b3e7-af13dcaca76e"). InnerVolumeSpecName "kube-api-access-6sv7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.363975 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-scripts" (OuterVolumeSpecName: "scripts") pod "3a00a076-ca9d-46e9-b3e7-af13dcaca76e" (UID: "3a00a076-ca9d-46e9-b3e7-af13dcaca76e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.376772 4950 scope.go:117] "RemoveContainer" containerID="7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891" Mar 18 20:27:14 crc kubenswrapper[4950]: E0318 20:27:14.382820 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891\": container with ID starting with 7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891 not found: ID does not exist" containerID="7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.382874 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891"} err="failed to get container status \"7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891\": rpc error: code = NotFound desc = could not find container \"7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891\": container with ID starting with 7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891 not found: ID does not exist" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.382903 4950 scope.go:117] "RemoveContainer" containerID="18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508" Mar 18 20:27:14 crc kubenswrapper[4950]: E0318 20:27:14.385145 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508\": container with ID starting with 18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508 not found: ID does not exist" containerID="18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.385191 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508"} err="failed to get container status \"18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508\": rpc error: code = NotFound desc = could not find container \"18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508\": container with ID starting with 18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508 not found: ID does not exist" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.385216 4950 scope.go:117] "RemoveContainer" containerID="c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc" Mar 18 20:27:14 crc kubenswrapper[4950]: E0318 20:27:14.385677 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc\": container with ID starting with c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc not found: ID does not exist" containerID="c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.385708 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc"} err="failed to get container status \"c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc\": rpc error: code = NotFound desc = could not find container \"c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc\": container with ID starting with c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc not found: ID does not exist" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.385726 4950 scope.go:117] "RemoveContainer" containerID="15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca" Mar 18 20:27:14 crc kubenswrapper[4950]: E0318 20:27:14.385937 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca\": container with ID starting with 15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca not found: ID does not exist" containerID="15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.385960 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca"} err="failed to get container status \"15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca\": rpc error: code = NotFound desc = could not find container \"15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca\": container with ID starting with 15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca not found: ID does not exist" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.385978 4950 scope.go:117] "RemoveContainer" containerID="7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.386158 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891"} err="failed to get container status \"7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891\": rpc error: code = NotFound desc = could not find container \"7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891\": container with ID starting with 7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891 not found: ID does not exist" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.386180 4950 scope.go:117] "RemoveContainer" containerID="18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.386367 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508"} err="failed to get container status \"18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508\": rpc error: code = NotFound desc = could not find container \"18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508\": container with ID starting with 18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508 not found: ID does not exist" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.386392 4950 scope.go:117] "RemoveContainer" containerID="c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.387098 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc"} err="failed to get container status \"c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc\": rpc error: code = NotFound desc = could not find container \"c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc\": container with ID starting with c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc not found: ID does not exist" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.387123 4950 scope.go:117] "RemoveContainer" containerID="15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.387338 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca"} err="failed to get container status \"15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca\": rpc error: code = NotFound desc = could not find container \"15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca\": container with ID starting with 15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca not found: ID does not exist" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.387359 4950 scope.go:117] "RemoveContainer" containerID="7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.387894 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891"} err="failed to get container status \"7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891\": rpc error: code = NotFound desc = could not find container \"7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891\": container with ID starting with 7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891 not found: ID does not exist" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.387932 4950 scope.go:117] "RemoveContainer" containerID="18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.394559 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508"} err="failed to get container status \"18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508\": rpc error: code = NotFound desc = could not find container \"18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508\": container with ID starting with 18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508 not found: ID does not exist" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.394707 4950 scope.go:117] "RemoveContainer" containerID="c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.402543 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc"} err="failed to get container status \"c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc\": rpc error: code = NotFound desc = could not find container \"c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc\": container with ID starting with c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc not found: ID does not exist" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.402707 4950 scope.go:117] "RemoveContainer" containerID="15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.403930 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca"} err="failed to get container status \"15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca\": rpc error: code = NotFound desc = could not find container \"15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca\": container with ID starting with 15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca not found: ID does not exist" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.403981 4950 scope.go:117] "RemoveContainer" containerID="7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.409047 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891"} err="failed to get container status \"7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891\": rpc error: code = NotFound desc = could not find container \"7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891\": container with ID starting with 7b9786f1f16bf15ca7a0b2183733384c896bb71d44bf2c21d2e78eb370eea891 not found: ID does not exist" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.409108 4950 scope.go:117] "RemoveContainer" containerID="18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.410843 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508"} err="failed to get container status \"18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508\": rpc error: code = NotFound desc = could not find container \"18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508\": container with ID starting with 18ac3a38374da2b7f67a9cc573ea84d35c2d11bca2f885341806092206b74508 not found: ID does not exist" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.410877 4950 scope.go:117] "RemoveContainer" containerID="c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.411755 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc"} err="failed to get container status \"c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc\": rpc error: code = NotFound desc = could not find container \"c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc\": container with ID starting with c3720a75b2fa6f4b4609485450bebac3ba5879718973fbd796cda49a047d54fc not found: ID does not exist" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.411782 4950 scope.go:117] "RemoveContainer" containerID="15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.413176 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca"} err="failed to get container status \"15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca\": rpc error: code = NotFound desc = could not find container \"15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca\": container with ID starting with 15afde7629a359330ed07bbacfda802a0547ea96a45580dc2dacb2278eff72ca not found: ID does not exist" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.437671 4950 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.437721 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sv7h\" (UniqueName: \"kubernetes.io/projected/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-kube-api-access-6sv7h\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.437737 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.437749 4950 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.488528 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3a00a076-ca9d-46e9-b3e7-af13dcaca76e" (UID: "3a00a076-ca9d-46e9-b3e7-af13dcaca76e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.525670 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a00a076-ca9d-46e9-b3e7-af13dcaca76e" (UID: "3a00a076-ca9d-46e9-b3e7-af13dcaca76e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.538788 4950 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.538821 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.624618 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-config-data" (OuterVolumeSpecName: "config-data") pod "3a00a076-ca9d-46e9-b3e7-af13dcaca76e" (UID: "3a00a076-ca9d-46e9-b3e7-af13dcaca76e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.640125 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a00a076-ca9d-46e9-b3e7-af13dcaca76e-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.825516 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.832389 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.881370 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:27:14 crc kubenswrapper[4950]: E0318 20:27:14.881772 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a00a076-ca9d-46e9-b3e7-af13dcaca76e" containerName="proxy-httpd" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.881791 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a00a076-ca9d-46e9-b3e7-af13dcaca76e" containerName="proxy-httpd" Mar 18 20:27:14 crc kubenswrapper[4950]: E0318 20:27:14.881811 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a00a076-ca9d-46e9-b3e7-af13dcaca76e" containerName="sg-core" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.881817 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a00a076-ca9d-46e9-b3e7-af13dcaca76e" containerName="sg-core" Mar 18 20:27:14 crc kubenswrapper[4950]: E0318 20:27:14.881831 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33fe46ee-d1cc-4d44-b6c9-8f9e808086d1" containerName="barbican-api" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.881838 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="33fe46ee-d1cc-4d44-b6c9-8f9e808086d1" containerName="barbican-api" Mar 18 20:27:14 crc kubenswrapper[4950]: E0318 20:27:14.881850 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a00a076-ca9d-46e9-b3e7-af13dcaca76e" containerName="ceilometer-central-agent" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.881856 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a00a076-ca9d-46e9-b3e7-af13dcaca76e" containerName="ceilometer-central-agent" Mar 18 20:27:14 crc kubenswrapper[4950]: E0318 20:27:14.881865 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31bbaba1-cd09-469b-9b89-f7ff2213cd8e" containerName="neutron-api" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.881871 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="31bbaba1-cd09-469b-9b89-f7ff2213cd8e" containerName="neutron-api" Mar 18 20:27:14 crc kubenswrapper[4950]: E0318 20:27:14.881884 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a00a076-ca9d-46e9-b3e7-af13dcaca76e" containerName="ceilometer-notification-agent" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.881893 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a00a076-ca9d-46e9-b3e7-af13dcaca76e" containerName="ceilometer-notification-agent" Mar 18 20:27:14 crc kubenswrapper[4950]: E0318 20:27:14.881908 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31bbaba1-cd09-469b-9b89-f7ff2213cd8e" containerName="neutron-httpd" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.881917 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="31bbaba1-cd09-469b-9b89-f7ff2213cd8e" containerName="neutron-httpd" Mar 18 20:27:14 crc kubenswrapper[4950]: E0318 20:27:14.881939 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33fe46ee-d1cc-4d44-b6c9-8f9e808086d1" containerName="barbican-api-log" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.881948 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="33fe46ee-d1cc-4d44-b6c9-8f9e808086d1" containerName="barbican-api-log" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.882119 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="31bbaba1-cd09-469b-9b89-f7ff2213cd8e" containerName="neutron-httpd" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.882131 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="33fe46ee-d1cc-4d44-b6c9-8f9e808086d1" containerName="barbican-api-log" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.882166 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a00a076-ca9d-46e9-b3e7-af13dcaca76e" containerName="ceilometer-central-agent" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.882178 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="31bbaba1-cd09-469b-9b89-f7ff2213cd8e" containerName="neutron-api" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.882192 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="33fe46ee-d1cc-4d44-b6c9-8f9e808086d1" containerName="barbican-api" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.882207 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a00a076-ca9d-46e9-b3e7-af13dcaca76e" containerName="sg-core" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.882217 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a00a076-ca9d-46e9-b3e7-af13dcaca76e" containerName="ceilometer-notification-agent" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.882231 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a00a076-ca9d-46e9-b3e7-af13dcaca76e" containerName="proxy-httpd" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.883775 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.891095 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.891123 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 18 20:27:14 crc kubenswrapper[4950]: I0318 20:27:14.908691 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.051292 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-scripts\") pod \"ceilometer-0\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " pod="openstack/ceilometer-0" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.051357 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3feebfda-5171-4cee-9309-4b9c73c8026d-log-httpd\") pod \"ceilometer-0\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " pod="openstack/ceilometer-0" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.051380 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3feebfda-5171-4cee-9309-4b9c73c8026d-run-httpd\") pod \"ceilometer-0\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " pod="openstack/ceilometer-0" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.051405 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " pod="openstack/ceilometer-0" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.051449 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " pod="openstack/ceilometer-0" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.051492 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-config-data\") pod \"ceilometer-0\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " pod="openstack/ceilometer-0" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.051545 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtqpz\" (UniqueName: \"kubernetes.io/projected/3feebfda-5171-4cee-9309-4b9c73c8026d-kube-api-access-dtqpz\") pod \"ceilometer-0\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " pod="openstack/ceilometer-0" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.152503 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtqpz\" (UniqueName: \"kubernetes.io/projected/3feebfda-5171-4cee-9309-4b9c73c8026d-kube-api-access-dtqpz\") pod \"ceilometer-0\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " pod="openstack/ceilometer-0" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.152772 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-scripts\") pod \"ceilometer-0\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " pod="openstack/ceilometer-0" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.152881 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3feebfda-5171-4cee-9309-4b9c73c8026d-log-httpd\") pod \"ceilometer-0\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " pod="openstack/ceilometer-0" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.153122 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3feebfda-5171-4cee-9309-4b9c73c8026d-run-httpd\") pod \"ceilometer-0\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " pod="openstack/ceilometer-0" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.153585 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " pod="openstack/ceilometer-0" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.153524 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3feebfda-5171-4cee-9309-4b9c73c8026d-run-httpd\") pod \"ceilometer-0\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " pod="openstack/ceilometer-0" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.153361 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3feebfda-5171-4cee-9309-4b9c73c8026d-log-httpd\") pod \"ceilometer-0\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " pod="openstack/ceilometer-0" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.154024 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " pod="openstack/ceilometer-0" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.154218 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-config-data\") pod \"ceilometer-0\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " pod="openstack/ceilometer-0" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.156081 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-scripts\") pod \"ceilometer-0\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " pod="openstack/ceilometer-0" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.158801 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-config-data\") pod \"ceilometer-0\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " pod="openstack/ceilometer-0" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.159447 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " pod="openstack/ceilometer-0" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.160717 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " pod="openstack/ceilometer-0" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.179987 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtqpz\" (UniqueName: \"kubernetes.io/projected/3feebfda-5171-4cee-9309-4b9c73c8026d-kube-api-access-dtqpz\") pod \"ceilometer-0\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " pod="openstack/ceilometer-0" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.196808 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.488797 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a00a076-ca9d-46e9-b3e7-af13dcaca76e" path="/var/lib/kubelet/pods/3a00a076-ca9d-46e9-b3e7-af13dcaca76e/volumes" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.704332 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.750008 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-hdfzb"] Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.751258 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hdfzb" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.766999 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-hdfzb"] Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.836749 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-6t686"] Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.840548 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6t686" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.857173 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-6t686"] Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.863896 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5h5s\" (UniqueName: \"kubernetes.io/projected/90c9ab22-c07c-48df-afec-0d5439f06f96-kube-api-access-r5h5s\") pod \"nova-api-db-create-hdfzb\" (UID: \"90c9ab22-c07c-48df-afec-0d5439f06f96\") " pod="openstack/nova-api-db-create-hdfzb" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.864088 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90c9ab22-c07c-48df-afec-0d5439f06f96-operator-scripts\") pod \"nova-api-db-create-hdfzb\" (UID: \"90c9ab22-c07c-48df-afec-0d5439f06f96\") " pod="openstack/nova-api-db-create-hdfzb" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.943794 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-1118-account-create-update-6d4gl"] Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.945128 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1118-account-create-update-6d4gl" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.948378 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.954794 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-1118-account-create-update-6d4gl"] Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.965521 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90c9ab22-c07c-48df-afec-0d5439f06f96-operator-scripts\") pod \"nova-api-db-create-hdfzb\" (UID: \"90c9ab22-c07c-48df-afec-0d5439f06f96\") " pod="openstack/nova-api-db-create-hdfzb" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.965641 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t8fz\" (UniqueName: \"kubernetes.io/projected/15a8a30f-2d5d-4cbc-8fd6-71ec32c73712-kube-api-access-8t8fz\") pod \"nova-cell0-db-create-6t686\" (UID: \"15a8a30f-2d5d-4cbc-8fd6-71ec32c73712\") " pod="openstack/nova-cell0-db-create-6t686" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.965691 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15a8a30f-2d5d-4cbc-8fd6-71ec32c73712-operator-scripts\") pod \"nova-cell0-db-create-6t686\" (UID: \"15a8a30f-2d5d-4cbc-8fd6-71ec32c73712\") " pod="openstack/nova-cell0-db-create-6t686" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.965710 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5h5s\" (UniqueName: \"kubernetes.io/projected/90c9ab22-c07c-48df-afec-0d5439f06f96-kube-api-access-r5h5s\") pod \"nova-api-db-create-hdfzb\" (UID: \"90c9ab22-c07c-48df-afec-0d5439f06f96\") " pod="openstack/nova-api-db-create-hdfzb" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.966427 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90c9ab22-c07c-48df-afec-0d5439f06f96-operator-scripts\") pod \"nova-api-db-create-hdfzb\" (UID: \"90c9ab22-c07c-48df-afec-0d5439f06f96\") " pod="openstack/nova-api-db-create-hdfzb" Mar 18 20:27:15 crc kubenswrapper[4950]: I0318 20:27:15.998769 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5h5s\" (UniqueName: \"kubernetes.io/projected/90c9ab22-c07c-48df-afec-0d5439f06f96-kube-api-access-r5h5s\") pod \"nova-api-db-create-hdfzb\" (UID: \"90c9ab22-c07c-48df-afec-0d5439f06f96\") " pod="openstack/nova-api-db-create-hdfzb" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.045398 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-swn7d"] Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.047191 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-swn7d" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.068494 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5tx7\" (UniqueName: \"kubernetes.io/projected/3c107bf1-4431-444a-bb5b-2a44f4864e7e-kube-api-access-v5tx7\") pod \"nova-api-1118-account-create-update-6d4gl\" (UID: \"3c107bf1-4431-444a-bb5b-2a44f4864e7e\") " pod="openstack/nova-api-1118-account-create-update-6d4gl" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.068539 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t8fz\" (UniqueName: \"kubernetes.io/projected/15a8a30f-2d5d-4cbc-8fd6-71ec32c73712-kube-api-access-8t8fz\") pod \"nova-cell0-db-create-6t686\" (UID: \"15a8a30f-2d5d-4cbc-8fd6-71ec32c73712\") " pod="openstack/nova-cell0-db-create-6t686" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.068588 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15a8a30f-2d5d-4cbc-8fd6-71ec32c73712-operator-scripts\") pod \"nova-cell0-db-create-6t686\" (UID: \"15a8a30f-2d5d-4cbc-8fd6-71ec32c73712\") " pod="openstack/nova-cell0-db-create-6t686" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.068658 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c107bf1-4431-444a-bb5b-2a44f4864e7e-operator-scripts\") pod \"nova-api-1118-account-create-update-6d4gl\" (UID: \"3c107bf1-4431-444a-bb5b-2a44f4864e7e\") " pod="openstack/nova-api-1118-account-create-update-6d4gl" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.070992 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15a8a30f-2d5d-4cbc-8fd6-71ec32c73712-operator-scripts\") pod \"nova-cell0-db-create-6t686\" (UID: \"15a8a30f-2d5d-4cbc-8fd6-71ec32c73712\") " pod="openstack/nova-cell0-db-create-6t686" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.076674 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hdfzb" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.090349 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t8fz\" (UniqueName: \"kubernetes.io/projected/15a8a30f-2d5d-4cbc-8fd6-71ec32c73712-kube-api-access-8t8fz\") pod \"nova-cell0-db-create-6t686\" (UID: \"15a8a30f-2d5d-4cbc-8fd6-71ec32c73712\") " pod="openstack/nova-cell0-db-create-6t686" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.096544 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-swn7d"] Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.156765 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6t686" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.169878 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-91fe-account-create-update-kfnc8"] Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.169905 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj97w\" (UniqueName: \"kubernetes.io/projected/1c3408fa-8790-4354-b564-1c0978f35665-kube-api-access-nj97w\") pod \"nova-cell1-db-create-swn7d\" (UID: \"1c3408fa-8790-4354-b564-1c0978f35665\") " pod="openstack/nova-cell1-db-create-swn7d" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.169953 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c107bf1-4431-444a-bb5b-2a44f4864e7e-operator-scripts\") pod \"nova-api-1118-account-create-update-6d4gl\" (UID: \"3c107bf1-4431-444a-bb5b-2a44f4864e7e\") " pod="openstack/nova-api-1118-account-create-update-6d4gl" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.170094 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c3408fa-8790-4354-b564-1c0978f35665-operator-scripts\") pod \"nova-cell1-db-create-swn7d\" (UID: \"1c3408fa-8790-4354-b564-1c0978f35665\") " pod="openstack/nova-cell1-db-create-swn7d" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.170228 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5tx7\" (UniqueName: \"kubernetes.io/projected/3c107bf1-4431-444a-bb5b-2a44f4864e7e-kube-api-access-v5tx7\") pod \"nova-api-1118-account-create-update-6d4gl\" (UID: \"3c107bf1-4431-444a-bb5b-2a44f4864e7e\") " pod="openstack/nova-api-1118-account-create-update-6d4gl" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.170632 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c107bf1-4431-444a-bb5b-2a44f4864e7e-operator-scripts\") pod \"nova-api-1118-account-create-update-6d4gl\" (UID: \"3c107bf1-4431-444a-bb5b-2a44f4864e7e\") " pod="openstack/nova-api-1118-account-create-update-6d4gl" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.175597 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-91fe-account-create-update-kfnc8" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.184630 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.191962 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5tx7\" (UniqueName: \"kubernetes.io/projected/3c107bf1-4431-444a-bb5b-2a44f4864e7e-kube-api-access-v5tx7\") pod \"nova-api-1118-account-create-update-6d4gl\" (UID: \"3c107bf1-4431-444a-bb5b-2a44f4864e7e\") " pod="openstack/nova-api-1118-account-create-update-6d4gl" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.193260 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-91fe-account-create-update-kfnc8"] Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.241190 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3feebfda-5171-4cee-9309-4b9c73c8026d","Type":"ContainerStarted","Data":"137c044a36874640c5da15fd67f18f3ede19e99d94b9b0b17535868f5d90c726"} Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.262812 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1118-account-create-update-6d4gl" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.273138 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj97w\" (UniqueName: \"kubernetes.io/projected/1c3408fa-8790-4354-b564-1c0978f35665-kube-api-access-nj97w\") pod \"nova-cell1-db-create-swn7d\" (UID: \"1c3408fa-8790-4354-b564-1c0978f35665\") " pod="openstack/nova-cell1-db-create-swn7d" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.273247 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c3408fa-8790-4354-b564-1c0978f35665-operator-scripts\") pod \"nova-cell1-db-create-swn7d\" (UID: \"1c3408fa-8790-4354-b564-1c0978f35665\") " pod="openstack/nova-cell1-db-create-swn7d" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.273296 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgtc6\" (UniqueName: \"kubernetes.io/projected/9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3-kube-api-access-jgtc6\") pod \"nova-cell0-91fe-account-create-update-kfnc8\" (UID: \"9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3\") " pod="openstack/nova-cell0-91fe-account-create-update-kfnc8" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.273400 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3-operator-scripts\") pod \"nova-cell0-91fe-account-create-update-kfnc8\" (UID: \"9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3\") " pod="openstack/nova-cell0-91fe-account-create-update-kfnc8" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.274618 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c3408fa-8790-4354-b564-1c0978f35665-operator-scripts\") pod \"nova-cell1-db-create-swn7d\" (UID: \"1c3408fa-8790-4354-b564-1c0978f35665\") " pod="openstack/nova-cell1-db-create-swn7d" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.300251 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj97w\" (UniqueName: \"kubernetes.io/projected/1c3408fa-8790-4354-b564-1c0978f35665-kube-api-access-nj97w\") pod \"nova-cell1-db-create-swn7d\" (UID: \"1c3408fa-8790-4354-b564-1c0978f35665\") " pod="openstack/nova-cell1-db-create-swn7d" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.367242 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-swn7d" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.372524 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-dc64-account-create-update-tdrsv"] Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.376947 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-dc64-account-create-update-tdrsv" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.376995 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgtc6\" (UniqueName: \"kubernetes.io/projected/9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3-kube-api-access-jgtc6\") pod \"nova-cell0-91fe-account-create-update-kfnc8\" (UID: \"9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3\") " pod="openstack/nova-cell0-91fe-account-create-update-kfnc8" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.377116 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3-operator-scripts\") pod \"nova-cell0-91fe-account-create-update-kfnc8\" (UID: \"9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3\") " pod="openstack/nova-cell0-91fe-account-create-update-kfnc8" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.378251 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3-operator-scripts\") pod \"nova-cell0-91fe-account-create-update-kfnc8\" (UID: \"9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3\") " pod="openstack/nova-cell0-91fe-account-create-update-kfnc8" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.389377 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.395725 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-dc64-account-create-update-tdrsv"] Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.408779 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgtc6\" (UniqueName: \"kubernetes.io/projected/9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3-kube-api-access-jgtc6\") pod \"nova-cell0-91fe-account-create-update-kfnc8\" (UID: \"9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3\") " pod="openstack/nova-cell0-91fe-account-create-update-kfnc8" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.481499 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c21582df-e249-45d2-82de-9f0ab2592388-operator-scripts\") pod \"nova-cell1-dc64-account-create-update-tdrsv\" (UID: \"c21582df-e249-45d2-82de-9f0ab2592388\") " pod="openstack/nova-cell1-dc64-account-create-update-tdrsv" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.481826 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k7tj\" (UniqueName: \"kubernetes.io/projected/c21582df-e249-45d2-82de-9f0ab2592388-kube-api-access-7k7tj\") pod \"nova-cell1-dc64-account-create-update-tdrsv\" (UID: \"c21582df-e249-45d2-82de-9f0ab2592388\") " pod="openstack/nova-cell1-dc64-account-create-update-tdrsv" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.583480 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k7tj\" (UniqueName: \"kubernetes.io/projected/c21582df-e249-45d2-82de-9f0ab2592388-kube-api-access-7k7tj\") pod \"nova-cell1-dc64-account-create-update-tdrsv\" (UID: \"c21582df-e249-45d2-82de-9f0ab2592388\") " pod="openstack/nova-cell1-dc64-account-create-update-tdrsv" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.583809 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c21582df-e249-45d2-82de-9f0ab2592388-operator-scripts\") pod \"nova-cell1-dc64-account-create-update-tdrsv\" (UID: \"c21582df-e249-45d2-82de-9f0ab2592388\") " pod="openstack/nova-cell1-dc64-account-create-update-tdrsv" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.584576 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c21582df-e249-45d2-82de-9f0ab2592388-operator-scripts\") pod \"nova-cell1-dc64-account-create-update-tdrsv\" (UID: \"c21582df-e249-45d2-82de-9f0ab2592388\") " pod="openstack/nova-cell1-dc64-account-create-update-tdrsv" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.590687 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-91fe-account-create-update-kfnc8" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.600035 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k7tj\" (UniqueName: \"kubernetes.io/projected/c21582df-e249-45d2-82de-9f0ab2592388-kube-api-access-7k7tj\") pod \"nova-cell1-dc64-account-create-update-tdrsv\" (UID: \"c21582df-e249-45d2-82de-9f0ab2592388\") " pod="openstack/nova-cell1-dc64-account-create-update-tdrsv" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.671878 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-hdfzb"] Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.712911 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-dc64-account-create-update-tdrsv" Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.770136 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-1118-account-create-update-6d4gl"] Mar 18 20:27:16 crc kubenswrapper[4950]: I0318 20:27:16.855827 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-6t686"] Mar 18 20:27:17 crc kubenswrapper[4950]: I0318 20:27:17.101928 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-91fe-account-create-update-kfnc8"] Mar 18 20:27:17 crc kubenswrapper[4950]: I0318 20:27:17.116177 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-swn7d"] Mar 18 20:27:17 crc kubenswrapper[4950]: I0318 20:27:17.254698 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hdfzb" event={"ID":"90c9ab22-c07c-48df-afec-0d5439f06f96","Type":"ContainerStarted","Data":"7fc7429f018879091bb018b4a873c962ed82b644be5c1aa9828b1bedb37c2e25"} Mar 18 20:27:17 crc kubenswrapper[4950]: I0318 20:27:17.254736 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hdfzb" event={"ID":"90c9ab22-c07c-48df-afec-0d5439f06f96","Type":"ContainerStarted","Data":"d31e3f612a018676cce34b54d458fb7d15a904ba9b2ca800ce8d4301659738d7"} Mar 18 20:27:17 crc kubenswrapper[4950]: I0318 20:27:17.262122 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-91fe-account-create-update-kfnc8" event={"ID":"9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3","Type":"ContainerStarted","Data":"2888b0b56c4a6a497dbdbe4740e332049615552b3b22c4ec4f2efe64307535eb"} Mar 18 20:27:17 crc kubenswrapper[4950]: I0318 20:27:17.280298 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-hdfzb" podStartSLOduration=2.280269262 podStartE2EDuration="2.280269262s" podCreationTimestamp="2026-03-18 20:27:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:27:17.274933464 +0000 UTC m=+1250.515775332" watchObservedRunningTime="2026-03-18 20:27:17.280269262 +0000 UTC m=+1250.521111150" Mar 18 20:27:17 crc kubenswrapper[4950]: I0318 20:27:17.288048 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-swn7d" event={"ID":"1c3408fa-8790-4354-b564-1c0978f35665","Type":"ContainerStarted","Data":"ec5bd338a1b1533a5d0ec92b9b570b35b30297360b4f7791cd18fd78a63a6d45"} Mar 18 20:27:17 crc kubenswrapper[4950]: I0318 20:27:17.290003 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1118-account-create-update-6d4gl" event={"ID":"3c107bf1-4431-444a-bb5b-2a44f4864e7e","Type":"ContainerStarted","Data":"45e8537d70c5d73022541ed9f24875d9a618a2231a495244009b73ccefb10bd2"} Mar 18 20:27:17 crc kubenswrapper[4950]: I0318 20:27:17.290038 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1118-account-create-update-6d4gl" event={"ID":"3c107bf1-4431-444a-bb5b-2a44f4864e7e","Type":"ContainerStarted","Data":"1687d5a03d8f27ae6b231eb6fb3e4f9de335fd561f7427d0dedacda7b15006e5"} Mar 18 20:27:17 crc kubenswrapper[4950]: I0318 20:27:17.316017 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3feebfda-5171-4cee-9309-4b9c73c8026d","Type":"ContainerStarted","Data":"7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2"} Mar 18 20:27:17 crc kubenswrapper[4950]: I0318 20:27:17.320614 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-1118-account-create-update-6d4gl" podStartSLOduration=2.320590899 podStartE2EDuration="2.320590899s" podCreationTimestamp="2026-03-18 20:27:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:27:17.307525933 +0000 UTC m=+1250.548367791" watchObservedRunningTime="2026-03-18 20:27:17.320590899 +0000 UTC m=+1250.561432757" Mar 18 20:27:17 crc kubenswrapper[4950]: I0318 20:27:17.325962 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6t686" event={"ID":"15a8a30f-2d5d-4cbc-8fd6-71ec32c73712","Type":"ContainerStarted","Data":"2ee516b8896be1dad75e1893fbab2c77d87280fa229eee831c1b5e1067966a6c"} Mar 18 20:27:17 crc kubenswrapper[4950]: W0318 20:27:17.452535 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc21582df_e249_45d2_82de_9f0ab2592388.slice/crio-09039ab0364f58e0a8b57ae2ffda14df1ecedc6d483f526921f6e785e3d1fd1b WatchSource:0}: Error finding container 09039ab0364f58e0a8b57ae2ffda14df1ecedc6d483f526921f6e785e3d1fd1b: Status 404 returned error can't find the container with id 09039ab0364f58e0a8b57ae2ffda14df1ecedc6d483f526921f6e785e3d1fd1b Mar 18 20:27:17 crc kubenswrapper[4950]: I0318 20:27:17.452654 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-dc64-account-create-update-tdrsv"] Mar 18 20:27:18 crc kubenswrapper[4950]: I0318 20:27:18.346130 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3feebfda-5171-4cee-9309-4b9c73c8026d","Type":"ContainerStarted","Data":"63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c"} Mar 18 20:27:18 crc kubenswrapper[4950]: I0318 20:27:18.347863 4950 generic.go:334] "Generic (PLEG): container finished" podID="3c107bf1-4431-444a-bb5b-2a44f4864e7e" containerID="45e8537d70c5d73022541ed9f24875d9a618a2231a495244009b73ccefb10bd2" exitCode=0 Mar 18 20:27:18 crc kubenswrapper[4950]: I0318 20:27:18.347905 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1118-account-create-update-6d4gl" event={"ID":"3c107bf1-4431-444a-bb5b-2a44f4864e7e","Type":"ContainerDied","Data":"45e8537d70c5d73022541ed9f24875d9a618a2231a495244009b73ccefb10bd2"} Mar 18 20:27:18 crc kubenswrapper[4950]: I0318 20:27:18.354159 4950 generic.go:334] "Generic (PLEG): container finished" podID="c21582df-e249-45d2-82de-9f0ab2592388" containerID="17f18f1a155e9ef4b996c63aafaeda20d2259fcb48b022b2e9b7b4e3e7d1ff39" exitCode=0 Mar 18 20:27:18 crc kubenswrapper[4950]: I0318 20:27:18.354273 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-dc64-account-create-update-tdrsv" event={"ID":"c21582df-e249-45d2-82de-9f0ab2592388","Type":"ContainerDied","Data":"17f18f1a155e9ef4b996c63aafaeda20d2259fcb48b022b2e9b7b4e3e7d1ff39"} Mar 18 20:27:18 crc kubenswrapper[4950]: I0318 20:27:18.354334 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-dc64-account-create-update-tdrsv" event={"ID":"c21582df-e249-45d2-82de-9f0ab2592388","Type":"ContainerStarted","Data":"09039ab0364f58e0a8b57ae2ffda14df1ecedc6d483f526921f6e785e3d1fd1b"} Mar 18 20:27:18 crc kubenswrapper[4950]: I0318 20:27:18.366792 4950 generic.go:334] "Generic (PLEG): container finished" podID="15a8a30f-2d5d-4cbc-8fd6-71ec32c73712" containerID="2a8c4760decf0fe38f5a95798f9b66c27c17fa33491aa829a44dd89f1950b2ee" exitCode=0 Mar 18 20:27:18 crc kubenswrapper[4950]: I0318 20:27:18.366884 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6t686" event={"ID":"15a8a30f-2d5d-4cbc-8fd6-71ec32c73712","Type":"ContainerDied","Data":"2a8c4760decf0fe38f5a95798f9b66c27c17fa33491aa829a44dd89f1950b2ee"} Mar 18 20:27:18 crc kubenswrapper[4950]: I0318 20:27:18.382695 4950 generic.go:334] "Generic (PLEG): container finished" podID="90c9ab22-c07c-48df-afec-0d5439f06f96" containerID="7fc7429f018879091bb018b4a873c962ed82b644be5c1aa9828b1bedb37c2e25" exitCode=0 Mar 18 20:27:18 crc kubenswrapper[4950]: I0318 20:27:18.382775 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hdfzb" event={"ID":"90c9ab22-c07c-48df-afec-0d5439f06f96","Type":"ContainerDied","Data":"7fc7429f018879091bb018b4a873c962ed82b644be5c1aa9828b1bedb37c2e25"} Mar 18 20:27:18 crc kubenswrapper[4950]: I0318 20:27:18.392705 4950 generic.go:334] "Generic (PLEG): container finished" podID="9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3" containerID="eb35c59a9465151717c7e151ffbe1e5b8992d800790f92dd7e3a481dac5f00c0" exitCode=0 Mar 18 20:27:18 crc kubenswrapper[4950]: I0318 20:27:18.393082 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-91fe-account-create-update-kfnc8" event={"ID":"9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3","Type":"ContainerDied","Data":"eb35c59a9465151717c7e151ffbe1e5b8992d800790f92dd7e3a481dac5f00c0"} Mar 18 20:27:18 crc kubenswrapper[4950]: I0318 20:27:18.405628 4950 generic.go:334] "Generic (PLEG): container finished" podID="1c3408fa-8790-4354-b564-1c0978f35665" containerID="164cccb8565a81eb585d5aa123611b9f78b542e75607a9b63101fa870e906493" exitCode=0 Mar 18 20:27:18 crc kubenswrapper[4950]: I0318 20:27:18.405672 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-swn7d" event={"ID":"1c3408fa-8790-4354-b564-1c0978f35665","Type":"ContainerDied","Data":"164cccb8565a81eb585d5aa123611b9f78b542e75607a9b63101fa870e906493"} Mar 18 20:27:19 crc kubenswrapper[4950]: I0318 20:27:19.415594 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3feebfda-5171-4cee-9309-4b9c73c8026d","Type":"ContainerStarted","Data":"472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5"} Mar 18 20:27:19 crc kubenswrapper[4950]: I0318 20:27:19.822913 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-dc64-account-create-update-tdrsv" Mar 18 20:27:19 crc kubenswrapper[4950]: I0318 20:27:19.961907 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c21582df-e249-45d2-82de-9f0ab2592388-operator-scripts\") pod \"c21582df-e249-45d2-82de-9f0ab2592388\" (UID: \"c21582df-e249-45d2-82de-9f0ab2592388\") " Mar 18 20:27:19 crc kubenswrapper[4950]: I0318 20:27:19.961948 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7k7tj\" (UniqueName: \"kubernetes.io/projected/c21582df-e249-45d2-82de-9f0ab2592388-kube-api-access-7k7tj\") pod \"c21582df-e249-45d2-82de-9f0ab2592388\" (UID: \"c21582df-e249-45d2-82de-9f0ab2592388\") " Mar 18 20:27:19 crc kubenswrapper[4950]: I0318 20:27:19.964271 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c21582df-e249-45d2-82de-9f0ab2592388-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c21582df-e249-45d2-82de-9f0ab2592388" (UID: "c21582df-e249-45d2-82de-9f0ab2592388"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:27:19 crc kubenswrapper[4950]: I0318 20:27:19.970103 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c21582df-e249-45d2-82de-9f0ab2592388-kube-api-access-7k7tj" (OuterVolumeSpecName: "kube-api-access-7k7tj") pod "c21582df-e249-45d2-82de-9f0ab2592388" (UID: "c21582df-e249-45d2-82de-9f0ab2592388"). InnerVolumeSpecName "kube-api-access-7k7tj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.064695 4950 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c21582df-e249-45d2-82de-9f0ab2592388-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.064727 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7k7tj\" (UniqueName: \"kubernetes.io/projected/c21582df-e249-45d2-82de-9f0ab2592388-kube-api-access-7k7tj\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.099840 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-91fe-account-create-update-kfnc8" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.110770 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hdfzb" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.131925 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1118-account-create-update-6d4gl" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.140433 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6t686" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.146186 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-swn7d" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.267527 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15a8a30f-2d5d-4cbc-8fd6-71ec32c73712-operator-scripts\") pod \"15a8a30f-2d5d-4cbc-8fd6-71ec32c73712\" (UID: \"15a8a30f-2d5d-4cbc-8fd6-71ec32c73712\") " Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.267598 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5h5s\" (UniqueName: \"kubernetes.io/projected/90c9ab22-c07c-48df-afec-0d5439f06f96-kube-api-access-r5h5s\") pod \"90c9ab22-c07c-48df-afec-0d5439f06f96\" (UID: \"90c9ab22-c07c-48df-afec-0d5439f06f96\") " Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.267618 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nj97w\" (UniqueName: \"kubernetes.io/projected/1c3408fa-8790-4354-b564-1c0978f35665-kube-api-access-nj97w\") pod \"1c3408fa-8790-4354-b564-1c0978f35665\" (UID: \"1c3408fa-8790-4354-b564-1c0978f35665\") " Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.267645 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5tx7\" (UniqueName: \"kubernetes.io/projected/3c107bf1-4431-444a-bb5b-2a44f4864e7e-kube-api-access-v5tx7\") pod \"3c107bf1-4431-444a-bb5b-2a44f4864e7e\" (UID: \"3c107bf1-4431-444a-bb5b-2a44f4864e7e\") " Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.267717 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90c9ab22-c07c-48df-afec-0d5439f06f96-operator-scripts\") pod \"90c9ab22-c07c-48df-afec-0d5439f06f96\" (UID: \"90c9ab22-c07c-48df-afec-0d5439f06f96\") " Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.267766 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgtc6\" (UniqueName: \"kubernetes.io/projected/9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3-kube-api-access-jgtc6\") pod \"9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3\" (UID: \"9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3\") " Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.267802 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8t8fz\" (UniqueName: \"kubernetes.io/projected/15a8a30f-2d5d-4cbc-8fd6-71ec32c73712-kube-api-access-8t8fz\") pod \"15a8a30f-2d5d-4cbc-8fd6-71ec32c73712\" (UID: \"15a8a30f-2d5d-4cbc-8fd6-71ec32c73712\") " Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.267861 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3-operator-scripts\") pod \"9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3\" (UID: \"9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3\") " Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.267899 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c3408fa-8790-4354-b564-1c0978f35665-operator-scripts\") pod \"1c3408fa-8790-4354-b564-1c0978f35665\" (UID: \"1c3408fa-8790-4354-b564-1c0978f35665\") " Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.267924 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c107bf1-4431-444a-bb5b-2a44f4864e7e-operator-scripts\") pod \"3c107bf1-4431-444a-bb5b-2a44f4864e7e\" (UID: \"3c107bf1-4431-444a-bb5b-2a44f4864e7e\") " Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.268022 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15a8a30f-2d5d-4cbc-8fd6-71ec32c73712-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "15a8a30f-2d5d-4cbc-8fd6-71ec32c73712" (UID: "15a8a30f-2d5d-4cbc-8fd6-71ec32c73712"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.268277 4950 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15a8a30f-2d5d-4cbc-8fd6-71ec32c73712-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.268656 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c107bf1-4431-444a-bb5b-2a44f4864e7e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3c107bf1-4431-444a-bb5b-2a44f4864e7e" (UID: "3c107bf1-4431-444a-bb5b-2a44f4864e7e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.268678 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c3408fa-8790-4354-b564-1c0978f35665-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1c3408fa-8790-4354-b564-1c0978f35665" (UID: "1c3408fa-8790-4354-b564-1c0978f35665"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.268897 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90c9ab22-c07c-48df-afec-0d5439f06f96-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "90c9ab22-c07c-48df-afec-0d5439f06f96" (UID: "90c9ab22-c07c-48df-afec-0d5439f06f96"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.268964 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3" (UID: "9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.276569 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c107bf1-4431-444a-bb5b-2a44f4864e7e-kube-api-access-v5tx7" (OuterVolumeSpecName: "kube-api-access-v5tx7") pod "3c107bf1-4431-444a-bb5b-2a44f4864e7e" (UID: "3c107bf1-4431-444a-bb5b-2a44f4864e7e"). InnerVolumeSpecName "kube-api-access-v5tx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.276625 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c3408fa-8790-4354-b564-1c0978f35665-kube-api-access-nj97w" (OuterVolumeSpecName: "kube-api-access-nj97w") pod "1c3408fa-8790-4354-b564-1c0978f35665" (UID: "1c3408fa-8790-4354-b564-1c0978f35665"). InnerVolumeSpecName "kube-api-access-nj97w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.276679 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3-kube-api-access-jgtc6" (OuterVolumeSpecName: "kube-api-access-jgtc6") pod "9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3" (UID: "9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3"). InnerVolumeSpecName "kube-api-access-jgtc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.276697 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90c9ab22-c07c-48df-afec-0d5439f06f96-kube-api-access-r5h5s" (OuterVolumeSpecName: "kube-api-access-r5h5s") pod "90c9ab22-c07c-48df-afec-0d5439f06f96" (UID: "90c9ab22-c07c-48df-afec-0d5439f06f96"). InnerVolumeSpecName "kube-api-access-r5h5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.280071 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15a8a30f-2d5d-4cbc-8fd6-71ec32c73712-kube-api-access-8t8fz" (OuterVolumeSpecName: "kube-api-access-8t8fz") pod "15a8a30f-2d5d-4cbc-8fd6-71ec32c73712" (UID: "15a8a30f-2d5d-4cbc-8fd6-71ec32c73712"). InnerVolumeSpecName "kube-api-access-8t8fz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.370138 4950 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90c9ab22-c07c-48df-afec-0d5439f06f96-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.370379 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgtc6\" (UniqueName: \"kubernetes.io/projected/9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3-kube-api-access-jgtc6\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.370521 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8t8fz\" (UniqueName: \"kubernetes.io/projected/15a8a30f-2d5d-4cbc-8fd6-71ec32c73712-kube-api-access-8t8fz\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.370615 4950 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.370689 4950 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c3408fa-8790-4354-b564-1c0978f35665-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.370761 4950 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c107bf1-4431-444a-bb5b-2a44f4864e7e-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.370832 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nj97w\" (UniqueName: \"kubernetes.io/projected/1c3408fa-8790-4354-b564-1c0978f35665-kube-api-access-nj97w\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.370910 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5h5s\" (UniqueName: \"kubernetes.io/projected/90c9ab22-c07c-48df-afec-0d5439f06f96-kube-api-access-r5h5s\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.370990 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5tx7\" (UniqueName: \"kubernetes.io/projected/3c107bf1-4431-444a-bb5b-2a44f4864e7e-kube-api-access-v5tx7\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.448360 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1118-account-create-update-6d4gl" event={"ID":"3c107bf1-4431-444a-bb5b-2a44f4864e7e","Type":"ContainerDied","Data":"1687d5a03d8f27ae6b231eb6fb3e4f9de335fd561f7427d0dedacda7b15006e5"} Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.449737 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1687d5a03d8f27ae6b231eb6fb3e4f9de335fd561f7427d0dedacda7b15006e5" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.449942 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1118-account-create-update-6d4gl" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.459547 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-dc64-account-create-update-tdrsv" event={"ID":"c21582df-e249-45d2-82de-9f0ab2592388","Type":"ContainerDied","Data":"09039ab0364f58e0a8b57ae2ffda14df1ecedc6d483f526921f6e785e3d1fd1b"} Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.459590 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09039ab0364f58e0a8b57ae2ffda14df1ecedc6d483f526921f6e785e3d1fd1b" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.459676 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-dc64-account-create-update-tdrsv" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.463059 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6t686" event={"ID":"15a8a30f-2d5d-4cbc-8fd6-71ec32c73712","Type":"ContainerDied","Data":"2ee516b8896be1dad75e1893fbab2c77d87280fa229eee831c1b5e1067966a6c"} Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.463100 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ee516b8896be1dad75e1893fbab2c77d87280fa229eee831c1b5e1067966a6c" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.463196 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6t686" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.483069 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hdfzb" event={"ID":"90c9ab22-c07c-48df-afec-0d5439f06f96","Type":"ContainerDied","Data":"d31e3f612a018676cce34b54d458fb7d15a904ba9b2ca800ce8d4301659738d7"} Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.483111 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d31e3f612a018676cce34b54d458fb7d15a904ba9b2ca800ce8d4301659738d7" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.483277 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hdfzb" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.486385 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-91fe-account-create-update-kfnc8" event={"ID":"9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3","Type":"ContainerDied","Data":"2888b0b56c4a6a497dbdbe4740e332049615552b3b22c4ec4f2efe64307535eb"} Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.486437 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2888b0b56c4a6a497dbdbe4740e332049615552b3b22c4ec4f2efe64307535eb" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.486547 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-91fe-account-create-update-kfnc8" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.501662 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-swn7d" event={"ID":"1c3408fa-8790-4354-b564-1c0978f35665","Type":"ContainerDied","Data":"ec5bd338a1b1533a5d0ec92b9b570b35b30297360b4f7791cd18fd78a63a6d45"} Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.501893 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec5bd338a1b1533a5d0ec92b9b570b35b30297360b4f7791cd18fd78a63a6d45" Mar 18 20:27:20 crc kubenswrapper[4950]: I0318 20:27:20.501709 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-swn7d" Mar 18 20:27:21 crc kubenswrapper[4950]: I0318 20:27:21.324927 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:27:21 crc kubenswrapper[4950]: I0318 20:27:21.448071 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5889fd87f8-6w269" Mar 18 20:27:21 crc kubenswrapper[4950]: I0318 20:27:21.531382 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3feebfda-5171-4cee-9309-4b9c73c8026d","Type":"ContainerStarted","Data":"3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971"} Mar 18 20:27:21 crc kubenswrapper[4950]: I0318 20:27:21.531459 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 18 20:27:21 crc kubenswrapper[4950]: I0318 20:27:21.539346 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-67f985b6b-xsntq"] Mar 18 20:27:21 crc kubenswrapper[4950]: I0318 20:27:21.541325 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-67f985b6b-xsntq" podUID="cebbda36-bfd8-4115-bdaf-616051b00eb0" containerName="placement-api" containerID="cri-o://071eed9b9bf77cd21c4fb507848fedfa3b7301e0abdc706543d2fea57787a0d5" gracePeriod=30 Mar 18 20:27:21 crc kubenswrapper[4950]: I0318 20:27:21.541305 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-67f985b6b-xsntq" podUID="cebbda36-bfd8-4115-bdaf-616051b00eb0" containerName="placement-log" containerID="cri-o://37b7d0561b5fbb0d11bf9f3c9faf271dff5effbb99739498ac66dc22afa16116" gracePeriod=30 Mar 18 20:27:21 crc kubenswrapper[4950]: I0318 20:27:21.558523 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.805934167 podStartE2EDuration="7.558504745s" podCreationTimestamp="2026-03-18 20:27:14 +0000 UTC" firstStartedPulling="2026-03-18 20:27:15.712039013 +0000 UTC m=+1248.952880881" lastFinishedPulling="2026-03-18 20:27:20.464609591 +0000 UTC m=+1253.705451459" observedRunningTime="2026-03-18 20:27:21.557326785 +0000 UTC m=+1254.798168653" watchObservedRunningTime="2026-03-18 20:27:21.558504745 +0000 UTC m=+1254.799346603" Mar 18 20:27:22 crc kubenswrapper[4950]: I0318 20:27:22.537237 4950 generic.go:334] "Generic (PLEG): container finished" podID="cebbda36-bfd8-4115-bdaf-616051b00eb0" containerID="37b7d0561b5fbb0d11bf9f3c9faf271dff5effbb99739498ac66dc22afa16116" exitCode=143 Mar 18 20:27:22 crc kubenswrapper[4950]: I0318 20:27:22.538133 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67f985b6b-xsntq" event={"ID":"cebbda36-bfd8-4115-bdaf-616051b00eb0","Type":"ContainerDied","Data":"37b7d0561b5fbb0d11bf9f3c9faf271dff5effbb99739498ac66dc22afa16116"} Mar 18 20:27:23 crc kubenswrapper[4950]: I0318 20:27:23.984259 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:27:24 crc kubenswrapper[4950]: I0318 20:27:24.572222 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3feebfda-5171-4cee-9309-4b9c73c8026d" containerName="ceilometer-central-agent" containerID="cri-o://7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2" gracePeriod=30 Mar 18 20:27:24 crc kubenswrapper[4950]: I0318 20:27:24.572284 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3feebfda-5171-4cee-9309-4b9c73c8026d" containerName="sg-core" containerID="cri-o://472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5" gracePeriod=30 Mar 18 20:27:24 crc kubenswrapper[4950]: I0318 20:27:24.572327 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3feebfda-5171-4cee-9309-4b9c73c8026d" containerName="ceilometer-notification-agent" containerID="cri-o://63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c" gracePeriod=30 Mar 18 20:27:24 crc kubenswrapper[4950]: I0318 20:27:24.572447 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3feebfda-5171-4cee-9309-4b9c73c8026d" containerName="proxy-httpd" containerID="cri-o://3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971" gracePeriod=30 Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.230554 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.390661 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtqpz\" (UniqueName: \"kubernetes.io/projected/3feebfda-5171-4cee-9309-4b9c73c8026d-kube-api-access-dtqpz\") pod \"3feebfda-5171-4cee-9309-4b9c73c8026d\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.390760 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3feebfda-5171-4cee-9309-4b9c73c8026d-run-httpd\") pod \"3feebfda-5171-4cee-9309-4b9c73c8026d\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.391198 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3feebfda-5171-4cee-9309-4b9c73c8026d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3feebfda-5171-4cee-9309-4b9c73c8026d" (UID: "3feebfda-5171-4cee-9309-4b9c73c8026d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.391272 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-sg-core-conf-yaml\") pod \"3feebfda-5171-4cee-9309-4b9c73c8026d\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.391331 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-scripts\") pod \"3feebfda-5171-4cee-9309-4b9c73c8026d\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.391360 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-config-data\") pod \"3feebfda-5171-4cee-9309-4b9c73c8026d\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.391687 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-combined-ca-bundle\") pod \"3feebfda-5171-4cee-9309-4b9c73c8026d\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.391750 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3feebfda-5171-4cee-9309-4b9c73c8026d-log-httpd\") pod \"3feebfda-5171-4cee-9309-4b9c73c8026d\" (UID: \"3feebfda-5171-4cee-9309-4b9c73c8026d\") " Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.392271 4950 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3feebfda-5171-4cee-9309-4b9c73c8026d-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.392965 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3feebfda-5171-4cee-9309-4b9c73c8026d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3feebfda-5171-4cee-9309-4b9c73c8026d" (UID: "3feebfda-5171-4cee-9309-4b9c73c8026d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.417837 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-scripts" (OuterVolumeSpecName: "scripts") pod "3feebfda-5171-4cee-9309-4b9c73c8026d" (UID: "3feebfda-5171-4cee-9309-4b9c73c8026d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.417947 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3feebfda-5171-4cee-9309-4b9c73c8026d-kube-api-access-dtqpz" (OuterVolumeSpecName: "kube-api-access-dtqpz") pod "3feebfda-5171-4cee-9309-4b9c73c8026d" (UID: "3feebfda-5171-4cee-9309-4b9c73c8026d"). InnerVolumeSpecName "kube-api-access-dtqpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.422185 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3feebfda-5171-4cee-9309-4b9c73c8026d" (UID: "3feebfda-5171-4cee-9309-4b9c73c8026d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.455360 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.493706 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtqpz\" (UniqueName: \"kubernetes.io/projected/3feebfda-5171-4cee-9309-4b9c73c8026d-kube-api-access-dtqpz\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.493733 4950 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.493744 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.493753 4950 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3feebfda-5171-4cee-9309-4b9c73c8026d-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.527743 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3feebfda-5171-4cee-9309-4b9c73c8026d" (UID: "3feebfda-5171-4cee-9309-4b9c73c8026d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.536613 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-config-data" (OuterVolumeSpecName: "config-data") pod "3feebfda-5171-4cee-9309-4b9c73c8026d" (UID: "3feebfda-5171-4cee-9309-4b9c73c8026d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.588733 4950 generic.go:334] "Generic (PLEG): container finished" podID="3feebfda-5171-4cee-9309-4b9c73c8026d" containerID="3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971" exitCode=0 Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.588768 4950 generic.go:334] "Generic (PLEG): container finished" podID="3feebfda-5171-4cee-9309-4b9c73c8026d" containerID="472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5" exitCode=2 Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.588778 4950 generic.go:334] "Generic (PLEG): container finished" podID="3feebfda-5171-4cee-9309-4b9c73c8026d" containerID="63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c" exitCode=0 Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.588787 4950 generic.go:334] "Generic (PLEG): container finished" podID="3feebfda-5171-4cee-9309-4b9c73c8026d" containerID="7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2" exitCode=0 Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.588812 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.588836 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3feebfda-5171-4cee-9309-4b9c73c8026d","Type":"ContainerDied","Data":"3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971"} Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.588864 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3feebfda-5171-4cee-9309-4b9c73c8026d","Type":"ContainerDied","Data":"472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5"} Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.588876 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3feebfda-5171-4cee-9309-4b9c73c8026d","Type":"ContainerDied","Data":"63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c"} Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.588887 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3feebfda-5171-4cee-9309-4b9c73c8026d","Type":"ContainerDied","Data":"7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2"} Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.588897 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3feebfda-5171-4cee-9309-4b9c73c8026d","Type":"ContainerDied","Data":"137c044a36874640c5da15fd67f18f3ede19e99d94b9b0b17535868f5d90c726"} Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.588912 4950 scope.go:117] "RemoveContainer" containerID="3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.593002 4950 generic.go:334] "Generic (PLEG): container finished" podID="cebbda36-bfd8-4115-bdaf-616051b00eb0" containerID="071eed9b9bf77cd21c4fb507848fedfa3b7301e0abdc706543d2fea57787a0d5" exitCode=0 Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.593041 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67f985b6b-xsntq" event={"ID":"cebbda36-bfd8-4115-bdaf-616051b00eb0","Type":"ContainerDied","Data":"071eed9b9bf77cd21c4fb507848fedfa3b7301e0abdc706543d2fea57787a0d5"} Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.593061 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67f985b6b-xsntq" event={"ID":"cebbda36-bfd8-4115-bdaf-616051b00eb0","Type":"ContainerDied","Data":"366be91b14609f8f748dc76b9a1376386ea107b218e76a95421561e8e5f46136"} Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.593101 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-67f985b6b-xsntq" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.594289 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cebbda36-bfd8-4115-bdaf-616051b00eb0-logs\") pod \"cebbda36-bfd8-4115-bdaf-616051b00eb0\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.594466 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-combined-ca-bundle\") pod \"cebbda36-bfd8-4115-bdaf-616051b00eb0\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.594551 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-config-data\") pod \"cebbda36-bfd8-4115-bdaf-616051b00eb0\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.594614 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-internal-tls-certs\") pod \"cebbda36-bfd8-4115-bdaf-616051b00eb0\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.594712 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sq8gh\" (UniqueName: \"kubernetes.io/projected/cebbda36-bfd8-4115-bdaf-616051b00eb0-kube-api-access-sq8gh\") pod \"cebbda36-bfd8-4115-bdaf-616051b00eb0\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.594773 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-public-tls-certs\") pod \"cebbda36-bfd8-4115-bdaf-616051b00eb0\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.594798 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-scripts\") pod \"cebbda36-bfd8-4115-bdaf-616051b00eb0\" (UID: \"cebbda36-bfd8-4115-bdaf-616051b00eb0\") " Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.595372 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.595394 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3feebfda-5171-4cee-9309-4b9c73c8026d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.596676 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cebbda36-bfd8-4115-bdaf-616051b00eb0-logs" (OuterVolumeSpecName: "logs") pod "cebbda36-bfd8-4115-bdaf-616051b00eb0" (UID: "cebbda36-bfd8-4115-bdaf-616051b00eb0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.598883 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cebbda36-bfd8-4115-bdaf-616051b00eb0-kube-api-access-sq8gh" (OuterVolumeSpecName: "kube-api-access-sq8gh") pod "cebbda36-bfd8-4115-bdaf-616051b00eb0" (UID: "cebbda36-bfd8-4115-bdaf-616051b00eb0"). InnerVolumeSpecName "kube-api-access-sq8gh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.610398 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-scripts" (OuterVolumeSpecName: "scripts") pod "cebbda36-bfd8-4115-bdaf-616051b00eb0" (UID: "cebbda36-bfd8-4115-bdaf-616051b00eb0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.626885 4950 scope.go:117] "RemoveContainer" containerID="472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.638205 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.655449 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.666060 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cebbda36-bfd8-4115-bdaf-616051b00eb0" (UID: "cebbda36-bfd8-4115-bdaf-616051b00eb0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.670185 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:27:25 crc kubenswrapper[4950]: E0318 20:27:25.670682 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3feebfda-5171-4cee-9309-4b9c73c8026d" containerName="proxy-httpd" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.670709 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="3feebfda-5171-4cee-9309-4b9c73c8026d" containerName="proxy-httpd" Mar 18 20:27:25 crc kubenswrapper[4950]: E0318 20:27:25.670727 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cebbda36-bfd8-4115-bdaf-616051b00eb0" containerName="placement-api" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.670736 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="cebbda36-bfd8-4115-bdaf-616051b00eb0" containerName="placement-api" Mar 18 20:27:25 crc kubenswrapper[4950]: E0318 20:27:25.670752 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c3408fa-8790-4354-b564-1c0978f35665" containerName="mariadb-database-create" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.670762 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c3408fa-8790-4354-b564-1c0978f35665" containerName="mariadb-database-create" Mar 18 20:27:25 crc kubenswrapper[4950]: E0318 20:27:25.670779 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3feebfda-5171-4cee-9309-4b9c73c8026d" containerName="sg-core" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.670788 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="3feebfda-5171-4cee-9309-4b9c73c8026d" containerName="sg-core" Mar 18 20:27:25 crc kubenswrapper[4950]: E0318 20:27:25.670803 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15a8a30f-2d5d-4cbc-8fd6-71ec32c73712" containerName="mariadb-database-create" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.670812 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="15a8a30f-2d5d-4cbc-8fd6-71ec32c73712" containerName="mariadb-database-create" Mar 18 20:27:25 crc kubenswrapper[4950]: E0318 20:27:25.670829 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90c9ab22-c07c-48df-afec-0d5439f06f96" containerName="mariadb-database-create" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.670837 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="90c9ab22-c07c-48df-afec-0d5439f06f96" containerName="mariadb-database-create" Mar 18 20:27:25 crc kubenswrapper[4950]: E0318 20:27:25.670852 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c107bf1-4431-444a-bb5b-2a44f4864e7e" containerName="mariadb-account-create-update" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.670859 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c107bf1-4431-444a-bb5b-2a44f4864e7e" containerName="mariadb-account-create-update" Mar 18 20:27:25 crc kubenswrapper[4950]: E0318 20:27:25.670878 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3feebfda-5171-4cee-9309-4b9c73c8026d" containerName="ceilometer-notification-agent" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.670885 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="3feebfda-5171-4cee-9309-4b9c73c8026d" containerName="ceilometer-notification-agent" Mar 18 20:27:25 crc kubenswrapper[4950]: E0318 20:27:25.670903 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3" containerName="mariadb-account-create-update" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.670912 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3" containerName="mariadb-account-create-update" Mar 18 20:27:25 crc kubenswrapper[4950]: E0318 20:27:25.670927 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3feebfda-5171-4cee-9309-4b9c73c8026d" containerName="ceilometer-central-agent" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.670934 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="3feebfda-5171-4cee-9309-4b9c73c8026d" containerName="ceilometer-central-agent" Mar 18 20:27:25 crc kubenswrapper[4950]: E0318 20:27:25.670944 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cebbda36-bfd8-4115-bdaf-616051b00eb0" containerName="placement-log" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.670951 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="cebbda36-bfd8-4115-bdaf-616051b00eb0" containerName="placement-log" Mar 18 20:27:25 crc kubenswrapper[4950]: E0318 20:27:25.670962 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c21582df-e249-45d2-82de-9f0ab2592388" containerName="mariadb-account-create-update" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.670969 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="c21582df-e249-45d2-82de-9f0ab2592388" containerName="mariadb-account-create-update" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.671151 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="3feebfda-5171-4cee-9309-4b9c73c8026d" containerName="ceilometer-notification-agent" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.671167 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="cebbda36-bfd8-4115-bdaf-616051b00eb0" containerName="placement-api" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.671181 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="c21582df-e249-45d2-82de-9f0ab2592388" containerName="mariadb-account-create-update" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.671207 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="cebbda36-bfd8-4115-bdaf-616051b00eb0" containerName="placement-log" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.671216 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c107bf1-4431-444a-bb5b-2a44f4864e7e" containerName="mariadb-account-create-update" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.671231 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="3feebfda-5171-4cee-9309-4b9c73c8026d" containerName="ceilometer-central-agent" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.671241 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3" containerName="mariadb-account-create-update" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.671253 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="90c9ab22-c07c-48df-afec-0d5439f06f96" containerName="mariadb-database-create" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.671263 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c3408fa-8790-4354-b564-1c0978f35665" containerName="mariadb-database-create" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.671274 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="3feebfda-5171-4cee-9309-4b9c73c8026d" containerName="proxy-httpd" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.671293 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="15a8a30f-2d5d-4cbc-8fd6-71ec32c73712" containerName="mariadb-database-create" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.671308 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="3feebfda-5171-4cee-9309-4b9c73c8026d" containerName="sg-core" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.673698 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.676073 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.676903 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.682939 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.697818 4950 scope.go:117] "RemoveContainer" containerID="63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c" Mar 18 20:27:25 crc kubenswrapper[4950]: E0318 20:27:25.699058 4950 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3feebfda_5171_4cee_9309_4b9c73c8026d.slice\": RecentStats: unable to find data in memory cache]" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.701754 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.701787 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sq8gh\" (UniqueName: \"kubernetes.io/projected/cebbda36-bfd8-4115-bdaf-616051b00eb0-kube-api-access-sq8gh\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.701797 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.701805 4950 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cebbda36-bfd8-4115-bdaf-616051b00eb0-logs\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.793570 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "cebbda36-bfd8-4115-bdaf-616051b00eb0" (UID: "cebbda36-bfd8-4115-bdaf-616051b00eb0"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.796141 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-config-data" (OuterVolumeSpecName: "config-data") pod "cebbda36-bfd8-4115-bdaf-616051b00eb0" (UID: "cebbda36-bfd8-4115-bdaf-616051b00eb0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.796356 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "cebbda36-bfd8-4115-bdaf-616051b00eb0" (UID: "cebbda36-bfd8-4115-bdaf-616051b00eb0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.803482 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-scripts\") pod \"ceilometer-0\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " pod="openstack/ceilometer-0" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.803656 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " pod="openstack/ceilometer-0" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.803748 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " pod="openstack/ceilometer-0" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.803836 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63712b81-a3d5-47d4-82bf-840c6e21d7a4-run-httpd\") pod \"ceilometer-0\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " pod="openstack/ceilometer-0" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.803910 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63712b81-a3d5-47d4-82bf-840c6e21d7a4-log-httpd\") pod \"ceilometer-0\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " pod="openstack/ceilometer-0" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.803999 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g58tv\" (UniqueName: \"kubernetes.io/projected/63712b81-a3d5-47d4-82bf-840c6e21d7a4-kube-api-access-g58tv\") pod \"ceilometer-0\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " pod="openstack/ceilometer-0" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.804108 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-config-data\") pod \"ceilometer-0\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " pod="openstack/ceilometer-0" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.804300 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.804372 4950 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.804509 4950 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cebbda36-bfd8-4115-bdaf-616051b00eb0-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.873530 4950 scope.go:117] "RemoveContainer" containerID="7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.897301 4950 scope.go:117] "RemoveContainer" containerID="3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971" Mar 18 20:27:25 crc kubenswrapper[4950]: E0318 20:27:25.897812 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971\": container with ID starting with 3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971 not found: ID does not exist" containerID="3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.897955 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971"} err="failed to get container status \"3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971\": rpc error: code = NotFound desc = could not find container \"3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971\": container with ID starting with 3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971 not found: ID does not exist" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.898092 4950 scope.go:117] "RemoveContainer" containerID="472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5" Mar 18 20:27:25 crc kubenswrapper[4950]: E0318 20:27:25.899583 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5\": container with ID starting with 472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5 not found: ID does not exist" containerID="472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.899716 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5"} err="failed to get container status \"472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5\": rpc error: code = NotFound desc = could not find container \"472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5\": container with ID starting with 472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5 not found: ID does not exist" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.899794 4950 scope.go:117] "RemoveContainer" containerID="63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c" Mar 18 20:27:25 crc kubenswrapper[4950]: E0318 20:27:25.900079 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c\": container with ID starting with 63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c not found: ID does not exist" containerID="63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.900160 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c"} err="failed to get container status \"63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c\": rpc error: code = NotFound desc = could not find container \"63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c\": container with ID starting with 63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c not found: ID does not exist" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.900240 4950 scope.go:117] "RemoveContainer" containerID="7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2" Mar 18 20:27:25 crc kubenswrapper[4950]: E0318 20:27:25.900558 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2\": container with ID starting with 7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2 not found: ID does not exist" containerID="7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.900665 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2"} err="failed to get container status \"7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2\": rpc error: code = NotFound desc = could not find container \"7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2\": container with ID starting with 7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2 not found: ID does not exist" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.900751 4950 scope.go:117] "RemoveContainer" containerID="3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.901039 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971"} err="failed to get container status \"3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971\": rpc error: code = NotFound desc = could not find container \"3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971\": container with ID starting with 3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971 not found: ID does not exist" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.901128 4950 scope.go:117] "RemoveContainer" containerID="472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.901475 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5"} err="failed to get container status \"472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5\": rpc error: code = NotFound desc = could not find container \"472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5\": container with ID starting with 472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5 not found: ID does not exist" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.901565 4950 scope.go:117] "RemoveContainer" containerID="63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.901807 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c"} err="failed to get container status \"63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c\": rpc error: code = NotFound desc = could not find container \"63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c\": container with ID starting with 63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c not found: ID does not exist" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.901885 4950 scope.go:117] "RemoveContainer" containerID="7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.902161 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2"} err="failed to get container status \"7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2\": rpc error: code = NotFound desc = could not find container \"7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2\": container with ID starting with 7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2 not found: ID does not exist" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.902259 4950 scope.go:117] "RemoveContainer" containerID="3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.904273 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971"} err="failed to get container status \"3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971\": rpc error: code = NotFound desc = could not find container \"3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971\": container with ID starting with 3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971 not found: ID does not exist" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.904430 4950 scope.go:117] "RemoveContainer" containerID="472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.905368 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-scripts\") pod \"ceilometer-0\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " pod="openstack/ceilometer-0" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.905512 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " pod="openstack/ceilometer-0" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.905582 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " pod="openstack/ceilometer-0" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.905672 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63712b81-a3d5-47d4-82bf-840c6e21d7a4-run-httpd\") pod \"ceilometer-0\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " pod="openstack/ceilometer-0" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.905758 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63712b81-a3d5-47d4-82bf-840c6e21d7a4-log-httpd\") pod \"ceilometer-0\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " pod="openstack/ceilometer-0" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.905820 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g58tv\" (UniqueName: \"kubernetes.io/projected/63712b81-a3d5-47d4-82bf-840c6e21d7a4-kube-api-access-g58tv\") pod \"ceilometer-0\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " pod="openstack/ceilometer-0" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.905913 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-config-data\") pod \"ceilometer-0\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " pod="openstack/ceilometer-0" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.906675 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5"} err="failed to get container status \"472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5\": rpc error: code = NotFound desc = could not find container \"472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5\": container with ID starting with 472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5 not found: ID does not exist" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.906734 4950 scope.go:117] "RemoveContainer" containerID="63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.907130 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c"} err="failed to get container status \"63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c\": rpc error: code = NotFound desc = could not find container \"63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c\": container with ID starting with 63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c not found: ID does not exist" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.907151 4950 scope.go:117] "RemoveContainer" containerID="7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.907573 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2"} err="failed to get container status \"7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2\": rpc error: code = NotFound desc = could not find container \"7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2\": container with ID starting with 7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2 not found: ID does not exist" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.907629 4950 scope.go:117] "RemoveContainer" containerID="3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.907861 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971"} err="failed to get container status \"3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971\": rpc error: code = NotFound desc = could not find container \"3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971\": container with ID starting with 3fff1eabc647d1cb9f43a6fcd3ce9e3afefc4f9efc25284024c9b5613b25b971 not found: ID does not exist" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.907901 4950 scope.go:117] "RemoveContainer" containerID="472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.908571 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63712b81-a3d5-47d4-82bf-840c6e21d7a4-log-httpd\") pod \"ceilometer-0\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " pod="openstack/ceilometer-0" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.908715 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5"} err="failed to get container status \"472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5\": rpc error: code = NotFound desc = could not find container \"472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5\": container with ID starting with 472ea74bbb80d0f7de8891f128882684b928dba5b0d201a7f8aff1bf1b12d3a5 not found: ID does not exist" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.908795 4950 scope.go:117] "RemoveContainer" containerID="63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.909286 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c"} err="failed to get container status \"63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c\": rpc error: code = NotFound desc = could not find container \"63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c\": container with ID starting with 63458e726f472946e53bed09dea0bde9cef0f3c49bbfc60881357d9f7fe52b1c not found: ID does not exist" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.909372 4950 scope.go:117] "RemoveContainer" containerID="7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.909377 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63712b81-a3d5-47d4-82bf-840c6e21d7a4-run-httpd\") pod \"ceilometer-0\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " pod="openstack/ceilometer-0" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.909982 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " pod="openstack/ceilometer-0" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.918473 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2"} err="failed to get container status \"7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2\": rpc error: code = NotFound desc = could not find container \"7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2\": container with ID starting with 7b8a269499e14d49cc5c084cdd227ae083c746a1b0a55bdebac6f026b866d6c2 not found: ID does not exist" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.918616 4950 scope.go:117] "RemoveContainer" containerID="071eed9b9bf77cd21c4fb507848fedfa3b7301e0abdc706543d2fea57787a0d5" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.919051 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " pod="openstack/ceilometer-0" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.920809 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-scripts\") pod \"ceilometer-0\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " pod="openstack/ceilometer-0" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.921304 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-config-data\") pod \"ceilometer-0\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " pod="openstack/ceilometer-0" Mar 18 20:27:25 crc kubenswrapper[4950]: I0318 20:27:25.934054 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g58tv\" (UniqueName: \"kubernetes.io/projected/63712b81-a3d5-47d4-82bf-840c6e21d7a4-kube-api-access-g58tv\") pod \"ceilometer-0\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " pod="openstack/ceilometer-0" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.034893 4950 scope.go:117] "RemoveContainer" containerID="37b7d0561b5fbb0d11bf9f3c9faf271dff5effbb99739498ac66dc22afa16116" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.071176 4950 scope.go:117] "RemoveContainer" containerID="071eed9b9bf77cd21c4fb507848fedfa3b7301e0abdc706543d2fea57787a0d5" Mar 18 20:27:26 crc kubenswrapper[4950]: E0318 20:27:26.071611 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"071eed9b9bf77cd21c4fb507848fedfa3b7301e0abdc706543d2fea57787a0d5\": container with ID starting with 071eed9b9bf77cd21c4fb507848fedfa3b7301e0abdc706543d2fea57787a0d5 not found: ID does not exist" containerID="071eed9b9bf77cd21c4fb507848fedfa3b7301e0abdc706543d2fea57787a0d5" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.071640 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"071eed9b9bf77cd21c4fb507848fedfa3b7301e0abdc706543d2fea57787a0d5"} err="failed to get container status \"071eed9b9bf77cd21c4fb507848fedfa3b7301e0abdc706543d2fea57787a0d5\": rpc error: code = NotFound desc = could not find container \"071eed9b9bf77cd21c4fb507848fedfa3b7301e0abdc706543d2fea57787a0d5\": container with ID starting with 071eed9b9bf77cd21c4fb507848fedfa3b7301e0abdc706543d2fea57787a0d5 not found: ID does not exist" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.071663 4950 scope.go:117] "RemoveContainer" containerID="37b7d0561b5fbb0d11bf9f3c9faf271dff5effbb99739498ac66dc22afa16116" Mar 18 20:27:26 crc kubenswrapper[4950]: E0318 20:27:26.071906 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37b7d0561b5fbb0d11bf9f3c9faf271dff5effbb99739498ac66dc22afa16116\": container with ID starting with 37b7d0561b5fbb0d11bf9f3c9faf271dff5effbb99739498ac66dc22afa16116 not found: ID does not exist" containerID="37b7d0561b5fbb0d11bf9f3c9faf271dff5effbb99739498ac66dc22afa16116" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.071926 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37b7d0561b5fbb0d11bf9f3c9faf271dff5effbb99739498ac66dc22afa16116"} err="failed to get container status \"37b7d0561b5fbb0d11bf9f3c9faf271dff5effbb99739498ac66dc22afa16116\": rpc error: code = NotFound desc = could not find container \"37b7d0561b5fbb0d11bf9f3c9faf271dff5effbb99739498ac66dc22afa16116\": container with ID starting with 37b7d0561b5fbb0d11bf9f3c9faf271dff5effbb99739498ac66dc22afa16116 not found: ID does not exist" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.073530 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-67f985b6b-xsntq"] Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.092042 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-67f985b6b-xsntq"] Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.104617 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5wz42"] Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.105742 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-5wz42" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.107615 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.109913 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.110342 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-v8gmw" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.115975 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5wz42"] Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.177697 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.221646 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4958284-3b28-4eff-bf68-16e2160710d1-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-5wz42\" (UID: \"d4958284-3b28-4eff-bf68-16e2160710d1\") " pod="openstack/nova-cell0-conductor-db-sync-5wz42" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.221721 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4958284-3b28-4eff-bf68-16e2160710d1-scripts\") pod \"nova-cell0-conductor-db-sync-5wz42\" (UID: \"d4958284-3b28-4eff-bf68-16e2160710d1\") " pod="openstack/nova-cell0-conductor-db-sync-5wz42" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.221815 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smrsg\" (UniqueName: \"kubernetes.io/projected/d4958284-3b28-4eff-bf68-16e2160710d1-kube-api-access-smrsg\") pod \"nova-cell0-conductor-db-sync-5wz42\" (UID: \"d4958284-3b28-4eff-bf68-16e2160710d1\") " pod="openstack/nova-cell0-conductor-db-sync-5wz42" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.221847 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4958284-3b28-4eff-bf68-16e2160710d1-config-data\") pod \"nova-cell0-conductor-db-sync-5wz42\" (UID: \"d4958284-3b28-4eff-bf68-16e2160710d1\") " pod="openstack/nova-cell0-conductor-db-sync-5wz42" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.323483 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smrsg\" (UniqueName: \"kubernetes.io/projected/d4958284-3b28-4eff-bf68-16e2160710d1-kube-api-access-smrsg\") pod \"nova-cell0-conductor-db-sync-5wz42\" (UID: \"d4958284-3b28-4eff-bf68-16e2160710d1\") " pod="openstack/nova-cell0-conductor-db-sync-5wz42" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.323857 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4958284-3b28-4eff-bf68-16e2160710d1-config-data\") pod \"nova-cell0-conductor-db-sync-5wz42\" (UID: \"d4958284-3b28-4eff-bf68-16e2160710d1\") " pod="openstack/nova-cell0-conductor-db-sync-5wz42" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.323907 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4958284-3b28-4eff-bf68-16e2160710d1-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-5wz42\" (UID: \"d4958284-3b28-4eff-bf68-16e2160710d1\") " pod="openstack/nova-cell0-conductor-db-sync-5wz42" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.323952 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4958284-3b28-4eff-bf68-16e2160710d1-scripts\") pod \"nova-cell0-conductor-db-sync-5wz42\" (UID: \"d4958284-3b28-4eff-bf68-16e2160710d1\") " pod="openstack/nova-cell0-conductor-db-sync-5wz42" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.329149 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4958284-3b28-4eff-bf68-16e2160710d1-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-5wz42\" (UID: \"d4958284-3b28-4eff-bf68-16e2160710d1\") " pod="openstack/nova-cell0-conductor-db-sync-5wz42" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.331111 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4958284-3b28-4eff-bf68-16e2160710d1-scripts\") pod \"nova-cell0-conductor-db-sync-5wz42\" (UID: \"d4958284-3b28-4eff-bf68-16e2160710d1\") " pod="openstack/nova-cell0-conductor-db-sync-5wz42" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.334553 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4958284-3b28-4eff-bf68-16e2160710d1-config-data\") pod \"nova-cell0-conductor-db-sync-5wz42\" (UID: \"d4958284-3b28-4eff-bf68-16e2160710d1\") " pod="openstack/nova-cell0-conductor-db-sync-5wz42" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.343047 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smrsg\" (UniqueName: \"kubernetes.io/projected/d4958284-3b28-4eff-bf68-16e2160710d1-kube-api-access-smrsg\") pod \"nova-cell0-conductor-db-sync-5wz42\" (UID: \"d4958284-3b28-4eff-bf68-16e2160710d1\") " pod="openstack/nova-cell0-conductor-db-sync-5wz42" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.427493 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-5wz42" Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.649288 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:27:26 crc kubenswrapper[4950]: W0318 20:27:26.661243 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63712b81_a3d5_47d4_82bf_840c6e21d7a4.slice/crio-a7e250b5af11d31a6b9d29c6a5c254deb22871636e8e7f0e3a1409b19361081d WatchSource:0}: Error finding container a7e250b5af11d31a6b9d29c6a5c254deb22871636e8e7f0e3a1409b19361081d: Status 404 returned error can't find the container with id a7e250b5af11d31a6b9d29c6a5c254deb22871636e8e7f0e3a1409b19361081d Mar 18 20:27:26 crc kubenswrapper[4950]: I0318 20:27:26.721947 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5wz42"] Mar 18 20:27:26 crc kubenswrapper[4950]: W0318 20:27:26.728214 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4958284_3b28_4eff_bf68_16e2160710d1.slice/crio-b31db11ea1a5f8d7c4608ac9f2544211ab6a7963db1feee5df1f5908d0c7a42f WatchSource:0}: Error finding container b31db11ea1a5f8d7c4608ac9f2544211ab6a7963db1feee5df1f5908d0c7a42f: Status 404 returned error can't find the container with id b31db11ea1a5f8d7c4608ac9f2544211ab6a7963db1feee5df1f5908d0c7a42f Mar 18 20:27:27 crc kubenswrapper[4950]: I0318 20:27:27.399556 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:27:27 crc kubenswrapper[4950]: I0318 20:27:27.495159 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3feebfda-5171-4cee-9309-4b9c73c8026d" path="/var/lib/kubelet/pods/3feebfda-5171-4cee-9309-4b9c73c8026d/volumes" Mar 18 20:27:27 crc kubenswrapper[4950]: I0318 20:27:27.496205 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cebbda36-bfd8-4115-bdaf-616051b00eb0" path="/var/lib/kubelet/pods/cebbda36-bfd8-4115-bdaf-616051b00eb0/volumes" Mar 18 20:27:27 crc kubenswrapper[4950]: I0318 20:27:27.617111 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-5wz42" event={"ID":"d4958284-3b28-4eff-bf68-16e2160710d1","Type":"ContainerStarted","Data":"b31db11ea1a5f8d7c4608ac9f2544211ab6a7963db1feee5df1f5908d0c7a42f"} Mar 18 20:27:27 crc kubenswrapper[4950]: I0318 20:27:27.619698 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63712b81-a3d5-47d4-82bf-840c6e21d7a4","Type":"ContainerStarted","Data":"7e1603ad7ab1a76e4e31de3ebc38f287b7e264b42c6d0a363eff00114cc96c45"} Mar 18 20:27:27 crc kubenswrapper[4950]: I0318 20:27:27.619771 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63712b81-a3d5-47d4-82bf-840c6e21d7a4","Type":"ContainerStarted","Data":"a7e250b5af11d31a6b9d29c6a5c254deb22871636e8e7f0e3a1409b19361081d"} Mar 18 20:27:28 crc kubenswrapper[4950]: I0318 20:27:28.631086 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63712b81-a3d5-47d4-82bf-840c6e21d7a4","Type":"ContainerStarted","Data":"3f056d86dba09f734bcfb4061f63d07b79afe4c7c92a4963af8cb90cb83346f4"} Mar 18 20:27:29 crc kubenswrapper[4950]: I0318 20:27:29.641181 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63712b81-a3d5-47d4-82bf-840c6e21d7a4","Type":"ContainerStarted","Data":"96f30879e79b3bf96f10dd7924b19e0da53f18f4d384f453b12e6ad9c6c84cfb"} Mar 18 20:27:35 crc kubenswrapper[4950]: I0318 20:27:35.690685 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-5wz42" event={"ID":"d4958284-3b28-4eff-bf68-16e2160710d1","Type":"ContainerStarted","Data":"a96279bdd3cc64e412ca56c34ddd8aad698c5e309382d4ce2300837520cebaab"} Mar 18 20:27:35 crc kubenswrapper[4950]: I0318 20:27:35.692899 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63712b81-a3d5-47d4-82bf-840c6e21d7a4","Type":"ContainerStarted","Data":"fc4e5553e36115b387543d76999d22ff7c0e8db181d2cebb7614644d2203da08"} Mar 18 20:27:35 crc kubenswrapper[4950]: I0318 20:27:35.693012 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="63712b81-a3d5-47d4-82bf-840c6e21d7a4" containerName="ceilometer-central-agent" containerID="cri-o://7e1603ad7ab1a76e4e31de3ebc38f287b7e264b42c6d0a363eff00114cc96c45" gracePeriod=30 Mar 18 20:27:35 crc kubenswrapper[4950]: I0318 20:27:35.693041 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 18 20:27:35 crc kubenswrapper[4950]: I0318 20:27:35.693051 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="63712b81-a3d5-47d4-82bf-840c6e21d7a4" containerName="sg-core" containerID="cri-o://96f30879e79b3bf96f10dd7924b19e0da53f18f4d384f453b12e6ad9c6c84cfb" gracePeriod=30 Mar 18 20:27:35 crc kubenswrapper[4950]: I0318 20:27:35.693084 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="63712b81-a3d5-47d4-82bf-840c6e21d7a4" containerName="proxy-httpd" containerID="cri-o://fc4e5553e36115b387543d76999d22ff7c0e8db181d2cebb7614644d2203da08" gracePeriod=30 Mar 18 20:27:35 crc kubenswrapper[4950]: I0318 20:27:35.693078 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="63712b81-a3d5-47d4-82bf-840c6e21d7a4" containerName="ceilometer-notification-agent" containerID="cri-o://3f056d86dba09f734bcfb4061f63d07b79afe4c7c92a4963af8cb90cb83346f4" gracePeriod=30 Mar 18 20:27:35 crc kubenswrapper[4950]: I0318 20:27:35.720681 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-5wz42" podStartSLOduration=2.001921571 podStartE2EDuration="10.720664798s" podCreationTimestamp="2026-03-18 20:27:25 +0000 UTC" firstStartedPulling="2026-03-18 20:27:26.734167937 +0000 UTC m=+1259.975009805" lastFinishedPulling="2026-03-18 20:27:35.452911164 +0000 UTC m=+1268.693753032" observedRunningTime="2026-03-18 20:27:35.713319669 +0000 UTC m=+1268.954161537" watchObservedRunningTime="2026-03-18 20:27:35.720664798 +0000 UTC m=+1268.961506666" Mar 18 20:27:36 crc kubenswrapper[4950]: I0318 20:27:36.706243 4950 generic.go:334] "Generic (PLEG): container finished" podID="63712b81-a3d5-47d4-82bf-840c6e21d7a4" containerID="96f30879e79b3bf96f10dd7924b19e0da53f18f4d384f453b12e6ad9c6c84cfb" exitCode=2 Mar 18 20:27:36 crc kubenswrapper[4950]: I0318 20:27:36.706509 4950 generic.go:334] "Generic (PLEG): container finished" podID="63712b81-a3d5-47d4-82bf-840c6e21d7a4" containerID="3f056d86dba09f734bcfb4061f63d07b79afe4c7c92a4963af8cb90cb83346f4" exitCode=0 Mar 18 20:27:36 crc kubenswrapper[4950]: I0318 20:27:36.706523 4950 generic.go:334] "Generic (PLEG): container finished" podID="63712b81-a3d5-47d4-82bf-840c6e21d7a4" containerID="7e1603ad7ab1a76e4e31de3ebc38f287b7e264b42c6d0a363eff00114cc96c45" exitCode=0 Mar 18 20:27:36 crc kubenswrapper[4950]: I0318 20:27:36.706340 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63712b81-a3d5-47d4-82bf-840c6e21d7a4","Type":"ContainerDied","Data":"96f30879e79b3bf96f10dd7924b19e0da53f18f4d384f453b12e6ad9c6c84cfb"} Mar 18 20:27:36 crc kubenswrapper[4950]: I0318 20:27:36.706788 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63712b81-a3d5-47d4-82bf-840c6e21d7a4","Type":"ContainerDied","Data":"3f056d86dba09f734bcfb4061f63d07b79afe4c7c92a4963af8cb90cb83346f4"} Mar 18 20:27:36 crc kubenswrapper[4950]: I0318 20:27:36.706816 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63712b81-a3d5-47d4-82bf-840c6e21d7a4","Type":"ContainerDied","Data":"7e1603ad7ab1a76e4e31de3ebc38f287b7e264b42c6d0a363eff00114cc96c45"} Mar 18 20:27:46 crc kubenswrapper[4950]: I0318 20:27:46.805369 4950 generic.go:334] "Generic (PLEG): container finished" podID="d4958284-3b28-4eff-bf68-16e2160710d1" containerID="a96279bdd3cc64e412ca56c34ddd8aad698c5e309382d4ce2300837520cebaab" exitCode=0 Mar 18 20:27:46 crc kubenswrapper[4950]: I0318 20:27:46.805527 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-5wz42" event={"ID":"d4958284-3b28-4eff-bf68-16e2160710d1","Type":"ContainerDied","Data":"a96279bdd3cc64e412ca56c34ddd8aad698c5e309382d4ce2300837520cebaab"} Mar 18 20:27:46 crc kubenswrapper[4950]: I0318 20:27:46.842196 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=13.063675617 podStartE2EDuration="21.842175062s" podCreationTimestamp="2026-03-18 20:27:25 +0000 UTC" firstStartedPulling="2026-03-18 20:27:26.663929339 +0000 UTC m=+1259.904771207" lastFinishedPulling="2026-03-18 20:27:35.442428784 +0000 UTC m=+1268.683270652" observedRunningTime="2026-03-18 20:27:35.740573221 +0000 UTC m=+1268.981415109" watchObservedRunningTime="2026-03-18 20:27:46.842175062 +0000 UTC m=+1280.083016940" Mar 18 20:27:48 crc kubenswrapper[4950]: I0318 20:27:48.241546 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-5wz42" Mar 18 20:27:48 crc kubenswrapper[4950]: I0318 20:27:48.340282 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4958284-3b28-4eff-bf68-16e2160710d1-config-data\") pod \"d4958284-3b28-4eff-bf68-16e2160710d1\" (UID: \"d4958284-3b28-4eff-bf68-16e2160710d1\") " Mar 18 20:27:48 crc kubenswrapper[4950]: I0318 20:27:48.340402 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4958284-3b28-4eff-bf68-16e2160710d1-combined-ca-bundle\") pod \"d4958284-3b28-4eff-bf68-16e2160710d1\" (UID: \"d4958284-3b28-4eff-bf68-16e2160710d1\") " Mar 18 20:27:48 crc kubenswrapper[4950]: I0318 20:27:48.340568 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smrsg\" (UniqueName: \"kubernetes.io/projected/d4958284-3b28-4eff-bf68-16e2160710d1-kube-api-access-smrsg\") pod \"d4958284-3b28-4eff-bf68-16e2160710d1\" (UID: \"d4958284-3b28-4eff-bf68-16e2160710d1\") " Mar 18 20:27:48 crc kubenswrapper[4950]: I0318 20:27:48.340599 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4958284-3b28-4eff-bf68-16e2160710d1-scripts\") pod \"d4958284-3b28-4eff-bf68-16e2160710d1\" (UID: \"d4958284-3b28-4eff-bf68-16e2160710d1\") " Mar 18 20:27:48 crc kubenswrapper[4950]: I0318 20:27:48.348149 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4958284-3b28-4eff-bf68-16e2160710d1-scripts" (OuterVolumeSpecName: "scripts") pod "d4958284-3b28-4eff-bf68-16e2160710d1" (UID: "d4958284-3b28-4eff-bf68-16e2160710d1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:48 crc kubenswrapper[4950]: I0318 20:27:48.357707 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4958284-3b28-4eff-bf68-16e2160710d1-kube-api-access-smrsg" (OuterVolumeSpecName: "kube-api-access-smrsg") pod "d4958284-3b28-4eff-bf68-16e2160710d1" (UID: "d4958284-3b28-4eff-bf68-16e2160710d1"). InnerVolumeSpecName "kube-api-access-smrsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:27:48 crc kubenswrapper[4950]: I0318 20:27:48.372870 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4958284-3b28-4eff-bf68-16e2160710d1-config-data" (OuterVolumeSpecName: "config-data") pod "d4958284-3b28-4eff-bf68-16e2160710d1" (UID: "d4958284-3b28-4eff-bf68-16e2160710d1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:48 crc kubenswrapper[4950]: I0318 20:27:48.373712 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4958284-3b28-4eff-bf68-16e2160710d1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4958284-3b28-4eff-bf68-16e2160710d1" (UID: "d4958284-3b28-4eff-bf68-16e2160710d1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:27:48 crc kubenswrapper[4950]: I0318 20:27:48.441847 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smrsg\" (UniqueName: \"kubernetes.io/projected/d4958284-3b28-4eff-bf68-16e2160710d1-kube-api-access-smrsg\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:48 crc kubenswrapper[4950]: I0318 20:27:48.441875 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4958284-3b28-4eff-bf68-16e2160710d1-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:48 crc kubenswrapper[4950]: I0318 20:27:48.441884 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4958284-3b28-4eff-bf68-16e2160710d1-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:48 crc kubenswrapper[4950]: I0318 20:27:48.441892 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4958284-3b28-4eff-bf68-16e2160710d1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:27:48 crc kubenswrapper[4950]: I0318 20:27:48.833775 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-5wz42" event={"ID":"d4958284-3b28-4eff-bf68-16e2160710d1","Type":"ContainerDied","Data":"b31db11ea1a5f8d7c4608ac9f2544211ab6a7963db1feee5df1f5908d0c7a42f"} Mar 18 20:27:48 crc kubenswrapper[4950]: I0318 20:27:48.833838 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b31db11ea1a5f8d7c4608ac9f2544211ab6a7963db1feee5df1f5908d0c7a42f" Mar 18 20:27:48 crc kubenswrapper[4950]: I0318 20:27:48.833882 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-5wz42" Mar 18 20:27:49 crc kubenswrapper[4950]: I0318 20:27:49.026293 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 18 20:27:49 crc kubenswrapper[4950]: E0318 20:27:49.026689 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4958284-3b28-4eff-bf68-16e2160710d1" containerName="nova-cell0-conductor-db-sync" Mar 18 20:27:49 crc kubenswrapper[4950]: I0318 20:27:49.026707 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4958284-3b28-4eff-bf68-16e2160710d1" containerName="nova-cell0-conductor-db-sync" Mar 18 20:27:49 crc kubenswrapper[4950]: I0318 20:27:49.026879 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4958284-3b28-4eff-bf68-16e2160710d1" containerName="nova-cell0-conductor-db-sync" Mar 18 20:27:49 crc kubenswrapper[4950]: I0318 20:27:49.027381 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 18 20:27:49 crc kubenswrapper[4950]: I0318 20:27:49.029547 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 18 20:27:49 crc kubenswrapper[4950]: I0318 20:27:49.032914 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-v8gmw" Mar 18 20:27:49 crc kubenswrapper[4950]: I0318 20:27:49.052118 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fbcf2da-2f82-4065-85e5-9f701c550afb-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5fbcf2da-2f82-4065-85e5-9f701c550afb\") " pod="openstack/nova-cell0-conductor-0" Mar 18 20:27:49 crc kubenswrapper[4950]: I0318 20:27:49.052191 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fbcf2da-2f82-4065-85e5-9f701c550afb-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5fbcf2da-2f82-4065-85e5-9f701c550afb\") " pod="openstack/nova-cell0-conductor-0" Mar 18 20:27:49 crc kubenswrapper[4950]: I0318 20:27:49.052254 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsvz4\" (UniqueName: \"kubernetes.io/projected/5fbcf2da-2f82-4065-85e5-9f701c550afb-kube-api-access-rsvz4\") pod \"nova-cell0-conductor-0\" (UID: \"5fbcf2da-2f82-4065-85e5-9f701c550afb\") " pod="openstack/nova-cell0-conductor-0" Mar 18 20:27:49 crc kubenswrapper[4950]: I0318 20:27:49.053528 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 18 20:27:49 crc kubenswrapper[4950]: I0318 20:27:49.153514 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fbcf2da-2f82-4065-85e5-9f701c550afb-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5fbcf2da-2f82-4065-85e5-9f701c550afb\") " pod="openstack/nova-cell0-conductor-0" Mar 18 20:27:49 crc kubenswrapper[4950]: I0318 20:27:49.153902 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fbcf2da-2f82-4065-85e5-9f701c550afb-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5fbcf2da-2f82-4065-85e5-9f701c550afb\") " pod="openstack/nova-cell0-conductor-0" Mar 18 20:27:49 crc kubenswrapper[4950]: I0318 20:27:49.153963 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsvz4\" (UniqueName: \"kubernetes.io/projected/5fbcf2da-2f82-4065-85e5-9f701c550afb-kube-api-access-rsvz4\") pod \"nova-cell0-conductor-0\" (UID: \"5fbcf2da-2f82-4065-85e5-9f701c550afb\") " pod="openstack/nova-cell0-conductor-0" Mar 18 20:27:49 crc kubenswrapper[4950]: I0318 20:27:49.163947 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fbcf2da-2f82-4065-85e5-9f701c550afb-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5fbcf2da-2f82-4065-85e5-9f701c550afb\") " pod="openstack/nova-cell0-conductor-0" Mar 18 20:27:49 crc kubenswrapper[4950]: I0318 20:27:49.170526 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fbcf2da-2f82-4065-85e5-9f701c550afb-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5fbcf2da-2f82-4065-85e5-9f701c550afb\") " pod="openstack/nova-cell0-conductor-0" Mar 18 20:27:49 crc kubenswrapper[4950]: I0318 20:27:49.172386 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsvz4\" (UniqueName: \"kubernetes.io/projected/5fbcf2da-2f82-4065-85e5-9f701c550afb-kube-api-access-rsvz4\") pod \"nova-cell0-conductor-0\" (UID: \"5fbcf2da-2f82-4065-85e5-9f701c550afb\") " pod="openstack/nova-cell0-conductor-0" Mar 18 20:27:49 crc kubenswrapper[4950]: I0318 20:27:49.343180 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 18 20:27:49 crc kubenswrapper[4950]: W0318 20:27:49.810716 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5fbcf2da_2f82_4065_85e5_9f701c550afb.slice/crio-847bcde45fad7912226fd38e560cc47b1c3b1c805fcf3c2784121b2d2427779b WatchSource:0}: Error finding container 847bcde45fad7912226fd38e560cc47b1c3b1c805fcf3c2784121b2d2427779b: Status 404 returned error can't find the container with id 847bcde45fad7912226fd38e560cc47b1c3b1c805fcf3c2784121b2d2427779b Mar 18 20:27:49 crc kubenswrapper[4950]: I0318 20:27:49.819934 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 18 20:27:49 crc kubenswrapper[4950]: I0318 20:27:49.843133 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5fbcf2da-2f82-4065-85e5-9f701c550afb","Type":"ContainerStarted","Data":"847bcde45fad7912226fd38e560cc47b1c3b1c805fcf3c2784121b2d2427779b"} Mar 18 20:27:50 crc kubenswrapper[4950]: I0318 20:27:50.852675 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5fbcf2da-2f82-4065-85e5-9f701c550afb","Type":"ContainerStarted","Data":"e62b651f75c2232ef33bee0d252280a46f952745981b659c803a433c932f4db5"} Mar 18 20:27:50 crc kubenswrapper[4950]: I0318 20:27:50.853200 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Mar 18 20:27:50 crc kubenswrapper[4950]: I0318 20:27:50.879115 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.879092973 podStartE2EDuration="2.879092973s" podCreationTimestamp="2026-03-18 20:27:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:27:50.879012211 +0000 UTC m=+1284.119854079" watchObservedRunningTime="2026-03-18 20:27:50.879092973 +0000 UTC m=+1284.119934841" Mar 18 20:27:54 crc kubenswrapper[4950]: I0318 20:27:54.377094 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Mar 18 20:27:54 crc kubenswrapper[4950]: I0318 20:27:54.877274 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-n6btl"] Mar 18 20:27:54 crc kubenswrapper[4950]: I0318 20:27:54.878852 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-n6btl" Mar 18 20:27:54 crc kubenswrapper[4950]: I0318 20:27:54.882365 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Mar 18 20:27:54 crc kubenswrapper[4950]: I0318 20:27:54.884233 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Mar 18 20:27:54 crc kubenswrapper[4950]: I0318 20:27:54.888748 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-n6btl"] Mar 18 20:27:54 crc kubenswrapper[4950]: I0318 20:27:54.959110 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-scripts\") pod \"nova-cell0-cell-mapping-n6btl\" (UID: \"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f\") " pod="openstack/nova-cell0-cell-mapping-n6btl" Mar 18 20:27:54 crc kubenswrapper[4950]: I0318 20:27:54.959169 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txcrn\" (UniqueName: \"kubernetes.io/projected/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-kube-api-access-txcrn\") pod \"nova-cell0-cell-mapping-n6btl\" (UID: \"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f\") " pod="openstack/nova-cell0-cell-mapping-n6btl" Mar 18 20:27:54 crc kubenswrapper[4950]: I0318 20:27:54.959196 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-config-data\") pod \"nova-cell0-cell-mapping-n6btl\" (UID: \"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f\") " pod="openstack/nova-cell0-cell-mapping-n6btl" Mar 18 20:27:54 crc kubenswrapper[4950]: I0318 20:27:54.959227 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-n6btl\" (UID: \"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f\") " pod="openstack/nova-cell0-cell-mapping-n6btl" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.060069 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txcrn\" (UniqueName: \"kubernetes.io/projected/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-kube-api-access-txcrn\") pod \"nova-cell0-cell-mapping-n6btl\" (UID: \"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f\") " pod="openstack/nova-cell0-cell-mapping-n6btl" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.060107 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-config-data\") pod \"nova-cell0-cell-mapping-n6btl\" (UID: \"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f\") " pod="openstack/nova-cell0-cell-mapping-n6btl" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.060130 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-n6btl\" (UID: \"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f\") " pod="openstack/nova-cell0-cell-mapping-n6btl" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.060245 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-scripts\") pod \"nova-cell0-cell-mapping-n6btl\" (UID: \"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f\") " pod="openstack/nova-cell0-cell-mapping-n6btl" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.069357 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-config-data\") pod \"nova-cell0-cell-mapping-n6btl\" (UID: \"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f\") " pod="openstack/nova-cell0-cell-mapping-n6btl" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.073178 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-scripts\") pod \"nova-cell0-cell-mapping-n6btl\" (UID: \"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f\") " pod="openstack/nova-cell0-cell-mapping-n6btl" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.077702 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.079018 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.081259 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txcrn\" (UniqueName: \"kubernetes.io/projected/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-kube-api-access-txcrn\") pod \"nova-cell0-cell-mapping-n6btl\" (UID: \"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f\") " pod="openstack/nova-cell0-cell-mapping-n6btl" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.084077 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.086684 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-n6btl\" (UID: \"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f\") " pod="openstack/nova-cell0-cell-mapping-n6btl" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.104074 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.161762 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgmhc\" (UniqueName: \"kubernetes.io/projected/0217d09c-db1e-45b6-bc64-c3eca777e7dc-kube-api-access-vgmhc\") pod \"nova-api-0\" (UID: \"0217d09c-db1e-45b6-bc64-c3eca777e7dc\") " pod="openstack/nova-api-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.161823 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0217d09c-db1e-45b6-bc64-c3eca777e7dc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0217d09c-db1e-45b6-bc64-c3eca777e7dc\") " pod="openstack/nova-api-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.161879 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0217d09c-db1e-45b6-bc64-c3eca777e7dc-config-data\") pod \"nova-api-0\" (UID: \"0217d09c-db1e-45b6-bc64-c3eca777e7dc\") " pod="openstack/nova-api-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.161945 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0217d09c-db1e-45b6-bc64-c3eca777e7dc-logs\") pod \"nova-api-0\" (UID: \"0217d09c-db1e-45b6-bc64-c3eca777e7dc\") " pod="openstack/nova-api-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.205766 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.209596 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.214542 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.221828 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-n6btl" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.224104 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.265355 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0217d09c-db1e-45b6-bc64-c3eca777e7dc-logs\") pod \"nova-api-0\" (UID: \"0217d09c-db1e-45b6-bc64-c3eca777e7dc\") " pod="openstack/nova-api-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.265446 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgmhc\" (UniqueName: \"kubernetes.io/projected/0217d09c-db1e-45b6-bc64-c3eca777e7dc-kube-api-access-vgmhc\") pod \"nova-api-0\" (UID: \"0217d09c-db1e-45b6-bc64-c3eca777e7dc\") " pod="openstack/nova-api-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.265511 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0217d09c-db1e-45b6-bc64-c3eca777e7dc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0217d09c-db1e-45b6-bc64-c3eca777e7dc\") " pod="openstack/nova-api-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.265586 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0217d09c-db1e-45b6-bc64-c3eca777e7dc-config-data\") pod \"nova-api-0\" (UID: \"0217d09c-db1e-45b6-bc64-c3eca777e7dc\") " pod="openstack/nova-api-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.266801 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0217d09c-db1e-45b6-bc64-c3eca777e7dc-logs\") pod \"nova-api-0\" (UID: \"0217d09c-db1e-45b6-bc64-c3eca777e7dc\") " pod="openstack/nova-api-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.270263 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0217d09c-db1e-45b6-bc64-c3eca777e7dc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0217d09c-db1e-45b6-bc64-c3eca777e7dc\") " pod="openstack/nova-api-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.277223 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0217d09c-db1e-45b6-bc64-c3eca777e7dc-config-data\") pod \"nova-api-0\" (UID: \"0217d09c-db1e-45b6-bc64-c3eca777e7dc\") " pod="openstack/nova-api-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.339706 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.342991 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.350980 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.357009 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgmhc\" (UniqueName: \"kubernetes.io/projected/0217d09c-db1e-45b6-bc64-c3eca777e7dc-kube-api-access-vgmhc\") pod \"nova-api-0\" (UID: \"0217d09c-db1e-45b6-bc64-c3eca777e7dc\") " pod="openstack/nova-api-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.367571 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc721dc9-f559-47b2-acca-3ca09cf54e27-logs\") pod \"nova-metadata-0\" (UID: \"fc721dc9-f559-47b2-acca-3ca09cf54e27\") " pod="openstack/nova-metadata-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.367646 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc721dc9-f559-47b2-acca-3ca09cf54e27-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fc721dc9-f559-47b2-acca-3ca09cf54e27\") " pod="openstack/nova-metadata-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.367743 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ts7nt\" (UniqueName: \"kubernetes.io/projected/fc721dc9-f559-47b2-acca-3ca09cf54e27-kube-api-access-ts7nt\") pod \"nova-metadata-0\" (UID: \"fc721dc9-f559-47b2-acca-3ca09cf54e27\") " pod="openstack/nova-metadata-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.367771 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc721dc9-f559-47b2-acca-3ca09cf54e27-config-data\") pod \"nova-metadata-0\" (UID: \"fc721dc9-f559-47b2-acca-3ca09cf54e27\") " pod="openstack/nova-metadata-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.375176 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.422052 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.423105 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.425632 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.459375 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.463245 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.486559 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pzxt\" (UniqueName: \"kubernetes.io/projected/d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65-kube-api-access-4pzxt\") pod \"nova-scheduler-0\" (UID: \"d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65\") " pod="openstack/nova-scheduler-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.486599 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/861b3002-ed64-4b45-b1fa-dc6eabc6a611-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"861b3002-ed64-4b45-b1fa-dc6eabc6a611\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.486640 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpl8j\" (UniqueName: \"kubernetes.io/projected/861b3002-ed64-4b45-b1fa-dc6eabc6a611-kube-api-access-rpl8j\") pod \"nova-cell1-novncproxy-0\" (UID: \"861b3002-ed64-4b45-b1fa-dc6eabc6a611\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.486664 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/861b3002-ed64-4b45-b1fa-dc6eabc6a611-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"861b3002-ed64-4b45-b1fa-dc6eabc6a611\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.486698 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ts7nt\" (UniqueName: \"kubernetes.io/projected/fc721dc9-f559-47b2-acca-3ca09cf54e27-kube-api-access-ts7nt\") pod \"nova-metadata-0\" (UID: \"fc721dc9-f559-47b2-acca-3ca09cf54e27\") " pod="openstack/nova-metadata-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.486723 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc721dc9-f559-47b2-acca-3ca09cf54e27-config-data\") pod \"nova-metadata-0\" (UID: \"fc721dc9-f559-47b2-acca-3ca09cf54e27\") " pod="openstack/nova-metadata-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.486743 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65-config-data\") pod \"nova-scheduler-0\" (UID: \"d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65\") " pod="openstack/nova-scheduler-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.486760 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65\") " pod="openstack/nova-scheduler-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.486783 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc721dc9-f559-47b2-acca-3ca09cf54e27-logs\") pod \"nova-metadata-0\" (UID: \"fc721dc9-f559-47b2-acca-3ca09cf54e27\") " pod="openstack/nova-metadata-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.486819 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc721dc9-f559-47b2-acca-3ca09cf54e27-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fc721dc9-f559-47b2-acca-3ca09cf54e27\") " pod="openstack/nova-metadata-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.493632 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc721dc9-f559-47b2-acca-3ca09cf54e27-config-data\") pod \"nova-metadata-0\" (UID: \"fc721dc9-f559-47b2-acca-3ca09cf54e27\") " pod="openstack/nova-metadata-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.495898 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc721dc9-f559-47b2-acca-3ca09cf54e27-logs\") pod \"nova-metadata-0\" (UID: \"fc721dc9-f559-47b2-acca-3ca09cf54e27\") " pod="openstack/nova-metadata-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.501042 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc721dc9-f559-47b2-acca-3ca09cf54e27-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fc721dc9-f559-47b2-acca-3ca09cf54e27\") " pod="openstack/nova-metadata-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.530010 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ts7nt\" (UniqueName: \"kubernetes.io/projected/fc721dc9-f559-47b2-acca-3ca09cf54e27-kube-api-access-ts7nt\") pod \"nova-metadata-0\" (UID: \"fc721dc9-f559-47b2-acca-3ca09cf54e27\") " pod="openstack/nova-metadata-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.551119 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c6bb4bf9-h85pn"] Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.557900 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.560357 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c6bb4bf9-h85pn"] Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.590260 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpl8j\" (UniqueName: \"kubernetes.io/projected/861b3002-ed64-4b45-b1fa-dc6eabc6a611-kube-api-access-rpl8j\") pod \"nova-cell1-novncproxy-0\" (UID: \"861b3002-ed64-4b45-b1fa-dc6eabc6a611\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.590608 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/861b3002-ed64-4b45-b1fa-dc6eabc6a611-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"861b3002-ed64-4b45-b1fa-dc6eabc6a611\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.590703 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65-config-data\") pod \"nova-scheduler-0\" (UID: \"d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65\") " pod="openstack/nova-scheduler-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.590725 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65\") " pod="openstack/nova-scheduler-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.590789 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pzxt\" (UniqueName: \"kubernetes.io/projected/d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65-kube-api-access-4pzxt\") pod \"nova-scheduler-0\" (UID: \"d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65\") " pod="openstack/nova-scheduler-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.590818 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/861b3002-ed64-4b45-b1fa-dc6eabc6a611-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"861b3002-ed64-4b45-b1fa-dc6eabc6a611\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.597721 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/861b3002-ed64-4b45-b1fa-dc6eabc6a611-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"861b3002-ed64-4b45-b1fa-dc6eabc6a611\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.604835 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65-config-data\") pod \"nova-scheduler-0\" (UID: \"d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65\") " pod="openstack/nova-scheduler-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.605498 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65\") " pod="openstack/nova-scheduler-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.615069 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/861b3002-ed64-4b45-b1fa-dc6eabc6a611-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"861b3002-ed64-4b45-b1fa-dc6eabc6a611\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.619999 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpl8j\" (UniqueName: \"kubernetes.io/projected/861b3002-ed64-4b45-b1fa-dc6eabc6a611-kube-api-access-rpl8j\") pod \"nova-cell1-novncproxy-0\" (UID: \"861b3002-ed64-4b45-b1fa-dc6eabc6a611\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.621071 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pzxt\" (UniqueName: \"kubernetes.io/projected/d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65-kube-api-access-4pzxt\") pod \"nova-scheduler-0\" (UID: \"d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65\") " pod="openstack/nova-scheduler-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.692648 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-ovsdbserver-sb\") pod \"dnsmasq-dns-75c6bb4bf9-h85pn\" (UID: \"e56878e7-f76d-48b6-b760-7f045d214b31\") " pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.692688 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-ovsdbserver-nb\") pod \"dnsmasq-dns-75c6bb4bf9-h85pn\" (UID: \"e56878e7-f76d-48b6-b760-7f045d214b31\") " pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.692817 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-dns-svc\") pod \"dnsmasq-dns-75c6bb4bf9-h85pn\" (UID: \"e56878e7-f76d-48b6-b760-7f045d214b31\") " pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.692848 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-config\") pod \"dnsmasq-dns-75c6bb4bf9-h85pn\" (UID: \"e56878e7-f76d-48b6-b760-7f045d214b31\") " pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.692886 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn4rt\" (UniqueName: \"kubernetes.io/projected/e56878e7-f76d-48b6-b760-7f045d214b31-kube-api-access-zn4rt\") pod \"dnsmasq-dns-75c6bb4bf9-h85pn\" (UID: \"e56878e7-f76d-48b6-b760-7f045d214b31\") " pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.719010 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.794831 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-dns-svc\") pod \"dnsmasq-dns-75c6bb4bf9-h85pn\" (UID: \"e56878e7-f76d-48b6-b760-7f045d214b31\") " pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.794884 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-config\") pod \"dnsmasq-dns-75c6bb4bf9-h85pn\" (UID: \"e56878e7-f76d-48b6-b760-7f045d214b31\") " pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.794922 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn4rt\" (UniqueName: \"kubernetes.io/projected/e56878e7-f76d-48b6-b760-7f045d214b31-kube-api-access-zn4rt\") pod \"dnsmasq-dns-75c6bb4bf9-h85pn\" (UID: \"e56878e7-f76d-48b6-b760-7f045d214b31\") " pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.794959 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-ovsdbserver-sb\") pod \"dnsmasq-dns-75c6bb4bf9-h85pn\" (UID: \"e56878e7-f76d-48b6-b760-7f045d214b31\") " pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.794978 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-ovsdbserver-nb\") pod \"dnsmasq-dns-75c6bb4bf9-h85pn\" (UID: \"e56878e7-f76d-48b6-b760-7f045d214b31\") " pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.796540 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-config\") pod \"dnsmasq-dns-75c6bb4bf9-h85pn\" (UID: \"e56878e7-f76d-48b6-b760-7f045d214b31\") " pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.796697 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-dns-svc\") pod \"dnsmasq-dns-75c6bb4bf9-h85pn\" (UID: \"e56878e7-f76d-48b6-b760-7f045d214b31\") " pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.797019 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.797057 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-ovsdbserver-sb\") pod \"dnsmasq-dns-75c6bb4bf9-h85pn\" (UID: \"e56878e7-f76d-48b6-b760-7f045d214b31\") " pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.800765 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-ovsdbserver-nb\") pod \"dnsmasq-dns-75c6bb4bf9-h85pn\" (UID: \"e56878e7-f76d-48b6-b760-7f045d214b31\") " pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.813909 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn4rt\" (UniqueName: \"kubernetes.io/projected/e56878e7-f76d-48b6-b760-7f045d214b31-kube-api-access-zn4rt\") pod \"dnsmasq-dns-75c6bb4bf9-h85pn\" (UID: \"e56878e7-f76d-48b6-b760-7f045d214b31\") " pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.830895 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 18 20:27:55 crc kubenswrapper[4950]: I0318 20:27:55.908338 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.025633 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-n6btl"] Mar 18 20:27:56 crc kubenswrapper[4950]: W0318 20:27:56.061320 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c71dbcb_5a73_4c85_8a1f_fa389a0e284f.slice/crio-b63e54fb95815beecad31e64b2d7392d7bbbaecdc3036f209febbb4ab576686c WatchSource:0}: Error finding container b63e54fb95815beecad31e64b2d7392d7bbbaecdc3036f209febbb4ab576686c: Status 404 returned error can't find the container with id b63e54fb95815beecad31e64b2d7392d7bbbaecdc3036f209febbb4ab576686c Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.212584 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.217099 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="63712b81-a3d5-47d4-82bf-840c6e21d7a4" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.260950 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 18 20:27:56 crc kubenswrapper[4950]: W0318 20:27:56.346453 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2ad0b9f_caf6_4cc7_bd82_6145c4cbaf65.slice/crio-6f2822f65df5bffa4cabe7fee26ad23f67964917d15200cc4fd5f4b42350ea09 WatchSource:0}: Error finding container 6f2822f65df5bffa4cabe7fee26ad23f67964917d15200cc4fd5f4b42350ea09: Status 404 returned error can't find the container with id 6f2822f65df5bffa4cabe7fee26ad23f67964917d15200cc4fd5f4b42350ea09 Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.707002 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 18 20:27:56 crc kubenswrapper[4950]: W0318 20:27:56.730041 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod861b3002_ed64_4b45_b1fa_dc6eabc6a611.slice/crio-af556040a5d747b0db1decfb629b321a2b82da507ca31f08943b0e8af2929376 WatchSource:0}: Error finding container af556040a5d747b0db1decfb629b321a2b82da507ca31f08943b0e8af2929376: Status 404 returned error can't find the container with id af556040a5d747b0db1decfb629b321a2b82da507ca31f08943b0e8af2929376 Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.734124 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.742816 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c6bb4bf9-h85pn"] Mar 18 20:27:56 crc kubenswrapper[4950]: W0318 20:27:56.747296 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc721dc9_f559_47b2_acca_3ca09cf54e27.slice/crio-62024613fd6026238d95b25e641b4b79f5f82559a95c71e8626c11e455bf3f03 WatchSource:0}: Error finding container 62024613fd6026238d95b25e641b4b79f5f82559a95c71e8626c11e455bf3f03: Status 404 returned error can't find the container with id 62024613fd6026238d95b25e641b4b79f5f82559a95c71e8626c11e455bf3f03 Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.820671 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lzx4p"] Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.821830 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lzx4p" Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.824616 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.824878 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.833602 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff62857a-136b-4b75-abf8-655bab5c8984-scripts\") pod \"nova-cell1-conductor-db-sync-lzx4p\" (UID: \"ff62857a-136b-4b75-abf8-655bab5c8984\") " pod="openstack/nova-cell1-conductor-db-sync-lzx4p" Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.833699 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff62857a-136b-4b75-abf8-655bab5c8984-config-data\") pod \"nova-cell1-conductor-db-sync-lzx4p\" (UID: \"ff62857a-136b-4b75-abf8-655bab5c8984\") " pod="openstack/nova-cell1-conductor-db-sync-lzx4p" Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.833730 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rwsw\" (UniqueName: \"kubernetes.io/projected/ff62857a-136b-4b75-abf8-655bab5c8984-kube-api-access-7rwsw\") pod \"nova-cell1-conductor-db-sync-lzx4p\" (UID: \"ff62857a-136b-4b75-abf8-655bab5c8984\") " pod="openstack/nova-cell1-conductor-db-sync-lzx4p" Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.833810 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff62857a-136b-4b75-abf8-655bab5c8984-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-lzx4p\" (UID: \"ff62857a-136b-4b75-abf8-655bab5c8984\") " pod="openstack/nova-cell1-conductor-db-sync-lzx4p" Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.846603 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lzx4p"] Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.919331 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65","Type":"ContainerStarted","Data":"6f2822f65df5bffa4cabe7fee26ad23f67964917d15200cc4fd5f4b42350ea09"} Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.920481 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"861b3002-ed64-4b45-b1fa-dc6eabc6a611","Type":"ContainerStarted","Data":"af556040a5d747b0db1decfb629b321a2b82da507ca31f08943b0e8af2929376"} Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.921453 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc721dc9-f559-47b2-acca-3ca09cf54e27","Type":"ContainerStarted","Data":"62024613fd6026238d95b25e641b4b79f5f82559a95c71e8626c11e455bf3f03"} Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.922590 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-n6btl" event={"ID":"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f","Type":"ContainerStarted","Data":"0e584954e3a8b88f9cc604d88832e1d8a7009b70353dc58272712cf61225203c"} Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.922613 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-n6btl" event={"ID":"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f","Type":"ContainerStarted","Data":"b63e54fb95815beecad31e64b2d7392d7bbbaecdc3036f209febbb4ab576686c"} Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.924582 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0217d09c-db1e-45b6-bc64-c3eca777e7dc","Type":"ContainerStarted","Data":"98f18e55bb38a673453f674909d7ae1ec80c77ea73d96072a71b85c36d647ead"} Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.926158 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" event={"ID":"e56878e7-f76d-48b6-b760-7f045d214b31","Type":"ContainerStarted","Data":"3744ecbd0c76c69e44bee2529488df2bd0b13ac0e5659344ec43f4bd14aace9f"} Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.934581 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff62857a-136b-4b75-abf8-655bab5c8984-config-data\") pod \"nova-cell1-conductor-db-sync-lzx4p\" (UID: \"ff62857a-136b-4b75-abf8-655bab5c8984\") " pod="openstack/nova-cell1-conductor-db-sync-lzx4p" Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.934633 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rwsw\" (UniqueName: \"kubernetes.io/projected/ff62857a-136b-4b75-abf8-655bab5c8984-kube-api-access-7rwsw\") pod \"nova-cell1-conductor-db-sync-lzx4p\" (UID: \"ff62857a-136b-4b75-abf8-655bab5c8984\") " pod="openstack/nova-cell1-conductor-db-sync-lzx4p" Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.934743 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff62857a-136b-4b75-abf8-655bab5c8984-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-lzx4p\" (UID: \"ff62857a-136b-4b75-abf8-655bab5c8984\") " pod="openstack/nova-cell1-conductor-db-sync-lzx4p" Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.934828 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff62857a-136b-4b75-abf8-655bab5c8984-scripts\") pod \"nova-cell1-conductor-db-sync-lzx4p\" (UID: \"ff62857a-136b-4b75-abf8-655bab5c8984\") " pod="openstack/nova-cell1-conductor-db-sync-lzx4p" Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.944301 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff62857a-136b-4b75-abf8-655bab5c8984-scripts\") pod \"nova-cell1-conductor-db-sync-lzx4p\" (UID: \"ff62857a-136b-4b75-abf8-655bab5c8984\") " pod="openstack/nova-cell1-conductor-db-sync-lzx4p" Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.946968 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff62857a-136b-4b75-abf8-655bab5c8984-config-data\") pod \"nova-cell1-conductor-db-sync-lzx4p\" (UID: \"ff62857a-136b-4b75-abf8-655bab5c8984\") " pod="openstack/nova-cell1-conductor-db-sync-lzx4p" Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.950327 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-n6btl" podStartSLOduration=2.950314213 podStartE2EDuration="2.950314213s" podCreationTimestamp="2026-03-18 20:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:27:56.943852697 +0000 UTC m=+1290.184694585" watchObservedRunningTime="2026-03-18 20:27:56.950314213 +0000 UTC m=+1290.191156081" Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.952785 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff62857a-136b-4b75-abf8-655bab5c8984-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-lzx4p\" (UID: \"ff62857a-136b-4b75-abf8-655bab5c8984\") " pod="openstack/nova-cell1-conductor-db-sync-lzx4p" Mar 18 20:27:56 crc kubenswrapper[4950]: I0318 20:27:56.960969 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rwsw\" (UniqueName: \"kubernetes.io/projected/ff62857a-136b-4b75-abf8-655bab5c8984-kube-api-access-7rwsw\") pod \"nova-cell1-conductor-db-sync-lzx4p\" (UID: \"ff62857a-136b-4b75-abf8-655bab5c8984\") " pod="openstack/nova-cell1-conductor-db-sync-lzx4p" Mar 18 20:27:57 crc kubenswrapper[4950]: I0318 20:27:57.140764 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lzx4p" Mar 18 20:27:57 crc kubenswrapper[4950]: I0318 20:27:57.648511 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lzx4p"] Mar 18 20:27:57 crc kubenswrapper[4950]: I0318 20:27:57.937299 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lzx4p" event={"ID":"ff62857a-136b-4b75-abf8-655bab5c8984","Type":"ContainerStarted","Data":"9b8cd39ebaa272935fa8e4cd424226e975751123cef5c91404d6c296fc9242b4"} Mar 18 20:27:57 crc kubenswrapper[4950]: I0318 20:27:57.937634 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lzx4p" event={"ID":"ff62857a-136b-4b75-abf8-655bab5c8984","Type":"ContainerStarted","Data":"38eb01a41c0da6a858cd5db49e4dc60f76864a5d13e1fbb92c5b25cb16129b18"} Mar 18 20:27:57 crc kubenswrapper[4950]: I0318 20:27:57.940748 4950 generic.go:334] "Generic (PLEG): container finished" podID="e56878e7-f76d-48b6-b760-7f045d214b31" containerID="e3e6920bc9551fc817f39f8bce6232ff1cfebfebda7b624eed583416cbe4e955" exitCode=0 Mar 18 20:27:57 crc kubenswrapper[4950]: I0318 20:27:57.941861 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" event={"ID":"e56878e7-f76d-48b6-b760-7f045d214b31","Type":"ContainerDied","Data":"e3e6920bc9551fc817f39f8bce6232ff1cfebfebda7b624eed583416cbe4e955"} Mar 18 20:27:57 crc kubenswrapper[4950]: I0318 20:27:57.987103 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-lzx4p" podStartSLOduration=1.987088108 podStartE2EDuration="1.987088108s" podCreationTimestamp="2026-03-18 20:27:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:27:57.983779773 +0000 UTC m=+1291.224621631" watchObservedRunningTime="2026-03-18 20:27:57.987088108 +0000 UTC m=+1291.227929976" Mar 18 20:27:58 crc kubenswrapper[4950]: I0318 20:27:58.978661 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" event={"ID":"e56878e7-f76d-48b6-b760-7f045d214b31","Type":"ContainerStarted","Data":"a3962b4da8a143dd1b9742fb29395bfd12a2d9a771e9097dee95a2586f9c1f6e"} Mar 18 20:27:58 crc kubenswrapper[4950]: I0318 20:27:58.979165 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" Mar 18 20:27:59 crc kubenswrapper[4950]: I0318 20:27:59.019320 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" podStartSLOduration=4.019299765 podStartE2EDuration="4.019299765s" podCreationTimestamp="2026-03-18 20:27:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:27:58.999052624 +0000 UTC m=+1292.239894492" watchObservedRunningTime="2026-03-18 20:27:59.019299765 +0000 UTC m=+1292.260141633" Mar 18 20:27:59 crc kubenswrapper[4950]: I0318 20:27:59.067630 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 18 20:27:59 crc kubenswrapper[4950]: I0318 20:27:59.089173 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 18 20:28:00 crc kubenswrapper[4950]: I0318 20:28:00.133532 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564428-7k4b9"] Mar 18 20:28:00 crc kubenswrapper[4950]: I0318 20:28:00.150037 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564428-7k4b9" Mar 18 20:28:00 crc kubenswrapper[4950]: I0318 20:28:00.150229 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564428-7k4b9"] Mar 18 20:28:00 crc kubenswrapper[4950]: I0318 20:28:00.152730 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:28:00 crc kubenswrapper[4950]: I0318 20:28:00.152974 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:28:00 crc kubenswrapper[4950]: I0318 20:28:00.153893 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:28:00 crc kubenswrapper[4950]: I0318 20:28:00.312910 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n46v8\" (UniqueName: \"kubernetes.io/projected/d2548eb6-5b6d-4ffd-9655-a766eecf2128-kube-api-access-n46v8\") pod \"auto-csr-approver-29564428-7k4b9\" (UID: \"d2548eb6-5b6d-4ffd-9655-a766eecf2128\") " pod="openshift-infra/auto-csr-approver-29564428-7k4b9" Mar 18 20:28:00 crc kubenswrapper[4950]: I0318 20:28:00.414844 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n46v8\" (UniqueName: \"kubernetes.io/projected/d2548eb6-5b6d-4ffd-9655-a766eecf2128-kube-api-access-n46v8\") pod \"auto-csr-approver-29564428-7k4b9\" (UID: \"d2548eb6-5b6d-4ffd-9655-a766eecf2128\") " pod="openshift-infra/auto-csr-approver-29564428-7k4b9" Mar 18 20:28:00 crc kubenswrapper[4950]: I0318 20:28:00.503063 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n46v8\" (UniqueName: \"kubernetes.io/projected/d2548eb6-5b6d-4ffd-9655-a766eecf2128-kube-api-access-n46v8\") pod \"auto-csr-approver-29564428-7k4b9\" (UID: \"d2548eb6-5b6d-4ffd-9655-a766eecf2128\") " pod="openshift-infra/auto-csr-approver-29564428-7k4b9" Mar 18 20:28:00 crc kubenswrapper[4950]: I0318 20:28:00.777193 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564428-7k4b9" Mar 18 20:28:02 crc kubenswrapper[4950]: I0318 20:28:02.484868 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564428-7k4b9"] Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.035193 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564428-7k4b9" event={"ID":"d2548eb6-5b6d-4ffd-9655-a766eecf2128","Type":"ContainerStarted","Data":"94d9aa917c9114534ab1014bc992f7f0c41ddcba65a5467a3d7389f448218209"} Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.037086 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0217d09c-db1e-45b6-bc64-c3eca777e7dc","Type":"ContainerStarted","Data":"cf4c56175b75b3f1ced1fd7d0c734b9230cb2c9a634abf085dc8bdb522f80551"} Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.037110 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0217d09c-db1e-45b6-bc64-c3eca777e7dc","Type":"ContainerStarted","Data":"e8912d9e397ece5941a51539bc31e219f238391233a8d61b16965058aa4cff40"} Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.038674 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65","Type":"ContainerStarted","Data":"3466f5503ac2a971a31d2cb12b54e4dbb107fb69f7f31b74b2511edd002cca64"} Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.040260 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"861b3002-ed64-4b45-b1fa-dc6eabc6a611","Type":"ContainerStarted","Data":"e3dcff1971656e6b2a038d556bf18d138c9cfff265d2ca9e52f845ed6aa55b08"} Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.040350 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="861b3002-ed64-4b45-b1fa-dc6eabc6a611" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://e3dcff1971656e6b2a038d556bf18d138c9cfff265d2ca9e52f845ed6aa55b08" gracePeriod=30 Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.044523 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc721dc9-f559-47b2-acca-3ca09cf54e27","Type":"ContainerStarted","Data":"d7cd45b46b5df8b3724260bd260dabac67af1f3dc2706f464ce48b0414ed0973"} Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.044563 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc721dc9-f559-47b2-acca-3ca09cf54e27","Type":"ContainerStarted","Data":"a697431cbb9845e40c6398476927fba53d410feda8242149b251f2f0190f320a"} Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.044754 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fc721dc9-f559-47b2-acca-3ca09cf54e27" containerName="nova-metadata-log" containerID="cri-o://a697431cbb9845e40c6398476927fba53d410feda8242149b251f2f0190f320a" gracePeriod=30 Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.044767 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fc721dc9-f559-47b2-acca-3ca09cf54e27" containerName="nova-metadata-metadata" containerID="cri-o://d7cd45b46b5df8b3724260bd260dabac67af1f3dc2706f464ce48b0414ed0973" gracePeriod=30 Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.069955 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.444268283 podStartE2EDuration="8.06993134s" podCreationTimestamp="2026-03-18 20:27:55 +0000 UTC" firstStartedPulling="2026-03-18 20:27:56.274633927 +0000 UTC m=+1289.515475795" lastFinishedPulling="2026-03-18 20:28:01.900296974 +0000 UTC m=+1295.141138852" observedRunningTime="2026-03-18 20:28:03.062088278 +0000 UTC m=+1296.302930146" watchObservedRunningTime="2026-03-18 20:28:03.06993134 +0000 UTC m=+1296.310773228" Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.081748 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.9243087020000003 podStartE2EDuration="8.081730544s" podCreationTimestamp="2026-03-18 20:27:55 +0000 UTC" firstStartedPulling="2026-03-18 20:27:56.750106878 +0000 UTC m=+1289.990948746" lastFinishedPulling="2026-03-18 20:28:01.9075287 +0000 UTC m=+1295.148370588" observedRunningTime="2026-03-18 20:28:03.081643621 +0000 UTC m=+1296.322485489" watchObservedRunningTime="2026-03-18 20:28:03.081730544 +0000 UTC m=+1296.322572412" Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.096141 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.926404105 podStartE2EDuration="8.096121504s" podCreationTimestamp="2026-03-18 20:27:55 +0000 UTC" firstStartedPulling="2026-03-18 20:27:56.732822543 +0000 UTC m=+1289.973664421" lastFinishedPulling="2026-03-18 20:28:01.902539942 +0000 UTC m=+1295.143381820" observedRunningTime="2026-03-18 20:28:03.09322553 +0000 UTC m=+1296.334067398" watchObservedRunningTime="2026-03-18 20:28:03.096121504 +0000 UTC m=+1296.336963372" Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.697798 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.716074 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.1664431459999998 podStartE2EDuration="8.716053756s" podCreationTimestamp="2026-03-18 20:27:55 +0000 UTC" firstStartedPulling="2026-03-18 20:27:56.352959243 +0000 UTC m=+1289.593801111" lastFinishedPulling="2026-03-18 20:28:01.902569813 +0000 UTC m=+1295.143411721" observedRunningTime="2026-03-18 20:28:03.123890189 +0000 UTC m=+1296.364732067" watchObservedRunningTime="2026-03-18 20:28:03.716053756 +0000 UTC m=+1296.956895624" Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.782190 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc721dc9-f559-47b2-acca-3ca09cf54e27-combined-ca-bundle\") pod \"fc721dc9-f559-47b2-acca-3ca09cf54e27\" (UID: \"fc721dc9-f559-47b2-acca-3ca09cf54e27\") " Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.782347 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc721dc9-f559-47b2-acca-3ca09cf54e27-config-data\") pod \"fc721dc9-f559-47b2-acca-3ca09cf54e27\" (UID: \"fc721dc9-f559-47b2-acca-3ca09cf54e27\") " Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.782487 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc721dc9-f559-47b2-acca-3ca09cf54e27-logs\") pod \"fc721dc9-f559-47b2-acca-3ca09cf54e27\" (UID: \"fc721dc9-f559-47b2-acca-3ca09cf54e27\") " Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.782529 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ts7nt\" (UniqueName: \"kubernetes.io/projected/fc721dc9-f559-47b2-acca-3ca09cf54e27-kube-api-access-ts7nt\") pod \"fc721dc9-f559-47b2-acca-3ca09cf54e27\" (UID: \"fc721dc9-f559-47b2-acca-3ca09cf54e27\") " Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.783029 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc721dc9-f559-47b2-acca-3ca09cf54e27-logs" (OuterVolumeSpecName: "logs") pod "fc721dc9-f559-47b2-acca-3ca09cf54e27" (UID: "fc721dc9-f559-47b2-acca-3ca09cf54e27"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.783281 4950 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc721dc9-f559-47b2-acca-3ca09cf54e27-logs\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.802602 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc721dc9-f559-47b2-acca-3ca09cf54e27-kube-api-access-ts7nt" (OuterVolumeSpecName: "kube-api-access-ts7nt") pod "fc721dc9-f559-47b2-acca-3ca09cf54e27" (UID: "fc721dc9-f559-47b2-acca-3ca09cf54e27"). InnerVolumeSpecName "kube-api-access-ts7nt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.823400 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc721dc9-f559-47b2-acca-3ca09cf54e27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fc721dc9-f559-47b2-acca-3ca09cf54e27" (UID: "fc721dc9-f559-47b2-acca-3ca09cf54e27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.828556 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc721dc9-f559-47b2-acca-3ca09cf54e27-config-data" (OuterVolumeSpecName: "config-data") pod "fc721dc9-f559-47b2-acca-3ca09cf54e27" (UID: "fc721dc9-f559-47b2-acca-3ca09cf54e27"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.836563 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.836611 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.884985 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc721dc9-f559-47b2-acca-3ca09cf54e27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.885015 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc721dc9-f559-47b2-acca-3ca09cf54e27-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:03 crc kubenswrapper[4950]: I0318 20:28:03.885026 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ts7nt\" (UniqueName: \"kubernetes.io/projected/fc721dc9-f559-47b2-acca-3ca09cf54e27-kube-api-access-ts7nt\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.055050 4950 generic.go:334] "Generic (PLEG): container finished" podID="fc721dc9-f559-47b2-acca-3ca09cf54e27" containerID="d7cd45b46b5df8b3724260bd260dabac67af1f3dc2706f464ce48b0414ed0973" exitCode=0 Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.055079 4950 generic.go:334] "Generic (PLEG): container finished" podID="fc721dc9-f559-47b2-acca-3ca09cf54e27" containerID="a697431cbb9845e40c6398476927fba53d410feda8242149b251f2f0190f320a" exitCode=143 Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.055918 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.061097 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc721dc9-f559-47b2-acca-3ca09cf54e27","Type":"ContainerDied","Data":"d7cd45b46b5df8b3724260bd260dabac67af1f3dc2706f464ce48b0414ed0973"} Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.061189 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc721dc9-f559-47b2-acca-3ca09cf54e27","Type":"ContainerDied","Data":"a697431cbb9845e40c6398476927fba53d410feda8242149b251f2f0190f320a"} Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.061221 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc721dc9-f559-47b2-acca-3ca09cf54e27","Type":"ContainerDied","Data":"62024613fd6026238d95b25e641b4b79f5f82559a95c71e8626c11e455bf3f03"} Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.061255 4950 scope.go:117] "RemoveContainer" containerID="d7cd45b46b5df8b3724260bd260dabac67af1f3dc2706f464ce48b0414ed0973" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.091306 4950 scope.go:117] "RemoveContainer" containerID="a697431cbb9845e40c6398476927fba53d410feda8242149b251f2f0190f320a" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.106674 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.121773 4950 scope.go:117] "RemoveContainer" containerID="d7cd45b46b5df8b3724260bd260dabac67af1f3dc2706f464ce48b0414ed0973" Mar 18 20:28:04 crc kubenswrapper[4950]: E0318 20:28:04.122835 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7cd45b46b5df8b3724260bd260dabac67af1f3dc2706f464ce48b0414ed0973\": container with ID starting with d7cd45b46b5df8b3724260bd260dabac67af1f3dc2706f464ce48b0414ed0973 not found: ID does not exist" containerID="d7cd45b46b5df8b3724260bd260dabac67af1f3dc2706f464ce48b0414ed0973" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.122929 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7cd45b46b5df8b3724260bd260dabac67af1f3dc2706f464ce48b0414ed0973"} err="failed to get container status \"d7cd45b46b5df8b3724260bd260dabac67af1f3dc2706f464ce48b0414ed0973\": rpc error: code = NotFound desc = could not find container \"d7cd45b46b5df8b3724260bd260dabac67af1f3dc2706f464ce48b0414ed0973\": container with ID starting with d7cd45b46b5df8b3724260bd260dabac67af1f3dc2706f464ce48b0414ed0973 not found: ID does not exist" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.123017 4950 scope.go:117] "RemoveContainer" containerID="a697431cbb9845e40c6398476927fba53d410feda8242149b251f2f0190f320a" Mar 18 20:28:04 crc kubenswrapper[4950]: E0318 20:28:04.123734 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a697431cbb9845e40c6398476927fba53d410feda8242149b251f2f0190f320a\": container with ID starting with a697431cbb9845e40c6398476927fba53d410feda8242149b251f2f0190f320a not found: ID does not exist" containerID="a697431cbb9845e40c6398476927fba53d410feda8242149b251f2f0190f320a" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.123840 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a697431cbb9845e40c6398476927fba53d410feda8242149b251f2f0190f320a"} err="failed to get container status \"a697431cbb9845e40c6398476927fba53d410feda8242149b251f2f0190f320a\": rpc error: code = NotFound desc = could not find container \"a697431cbb9845e40c6398476927fba53d410feda8242149b251f2f0190f320a\": container with ID starting with a697431cbb9845e40c6398476927fba53d410feda8242149b251f2f0190f320a not found: ID does not exist" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.123914 4950 scope.go:117] "RemoveContainer" containerID="d7cd45b46b5df8b3724260bd260dabac67af1f3dc2706f464ce48b0414ed0973" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.124784 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7cd45b46b5df8b3724260bd260dabac67af1f3dc2706f464ce48b0414ed0973"} err="failed to get container status \"d7cd45b46b5df8b3724260bd260dabac67af1f3dc2706f464ce48b0414ed0973\": rpc error: code = NotFound desc = could not find container \"d7cd45b46b5df8b3724260bd260dabac67af1f3dc2706f464ce48b0414ed0973\": container with ID starting with d7cd45b46b5df8b3724260bd260dabac67af1f3dc2706f464ce48b0414ed0973 not found: ID does not exist" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.124831 4950 scope.go:117] "RemoveContainer" containerID="a697431cbb9845e40c6398476927fba53d410feda8242149b251f2f0190f320a" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.125065 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a697431cbb9845e40c6398476927fba53d410feda8242149b251f2f0190f320a"} err="failed to get container status \"a697431cbb9845e40c6398476927fba53d410feda8242149b251f2f0190f320a\": rpc error: code = NotFound desc = could not find container \"a697431cbb9845e40c6398476927fba53d410feda8242149b251f2f0190f320a\": container with ID starting with a697431cbb9845e40c6398476927fba53d410feda8242149b251f2f0190f320a not found: ID does not exist" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.126736 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.141165 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 18 20:28:04 crc kubenswrapper[4950]: E0318 20:28:04.141499 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc721dc9-f559-47b2-acca-3ca09cf54e27" containerName="nova-metadata-log" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.141514 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc721dc9-f559-47b2-acca-3ca09cf54e27" containerName="nova-metadata-log" Mar 18 20:28:04 crc kubenswrapper[4950]: E0318 20:28:04.141535 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc721dc9-f559-47b2-acca-3ca09cf54e27" containerName="nova-metadata-metadata" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.141541 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc721dc9-f559-47b2-acca-3ca09cf54e27" containerName="nova-metadata-metadata" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.141716 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc721dc9-f559-47b2-acca-3ca09cf54e27" containerName="nova-metadata-metadata" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.141739 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc721dc9-f559-47b2-acca-3ca09cf54e27" containerName="nova-metadata-log" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.142538 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.144817 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.165556 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.167244 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.326301 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2028e28b-656a-493b-9803-f63a21e625d2-logs\") pod \"nova-metadata-0\" (UID: \"2028e28b-656a-493b-9803-f63a21e625d2\") " pod="openstack/nova-metadata-0" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.326438 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2028e28b-656a-493b-9803-f63a21e625d2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2028e28b-656a-493b-9803-f63a21e625d2\") " pod="openstack/nova-metadata-0" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.326478 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2028e28b-656a-493b-9803-f63a21e625d2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2028e28b-656a-493b-9803-f63a21e625d2\") " pod="openstack/nova-metadata-0" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.326517 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2kzb\" (UniqueName: \"kubernetes.io/projected/2028e28b-656a-493b-9803-f63a21e625d2-kube-api-access-s2kzb\") pod \"nova-metadata-0\" (UID: \"2028e28b-656a-493b-9803-f63a21e625d2\") " pod="openstack/nova-metadata-0" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.326546 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2028e28b-656a-493b-9803-f63a21e625d2-config-data\") pod \"nova-metadata-0\" (UID: \"2028e28b-656a-493b-9803-f63a21e625d2\") " pod="openstack/nova-metadata-0" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.427645 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2028e28b-656a-493b-9803-f63a21e625d2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2028e28b-656a-493b-9803-f63a21e625d2\") " pod="openstack/nova-metadata-0" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.427698 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2028e28b-656a-493b-9803-f63a21e625d2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2028e28b-656a-493b-9803-f63a21e625d2\") " pod="openstack/nova-metadata-0" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.427744 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kzb\" (UniqueName: \"kubernetes.io/projected/2028e28b-656a-493b-9803-f63a21e625d2-kube-api-access-s2kzb\") pod \"nova-metadata-0\" (UID: \"2028e28b-656a-493b-9803-f63a21e625d2\") " pod="openstack/nova-metadata-0" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.427770 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2028e28b-656a-493b-9803-f63a21e625d2-config-data\") pod \"nova-metadata-0\" (UID: \"2028e28b-656a-493b-9803-f63a21e625d2\") " pod="openstack/nova-metadata-0" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.427851 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2028e28b-656a-493b-9803-f63a21e625d2-logs\") pod \"nova-metadata-0\" (UID: \"2028e28b-656a-493b-9803-f63a21e625d2\") " pod="openstack/nova-metadata-0" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.428320 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2028e28b-656a-493b-9803-f63a21e625d2-logs\") pod \"nova-metadata-0\" (UID: \"2028e28b-656a-493b-9803-f63a21e625d2\") " pod="openstack/nova-metadata-0" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.434307 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2028e28b-656a-493b-9803-f63a21e625d2-config-data\") pod \"nova-metadata-0\" (UID: \"2028e28b-656a-493b-9803-f63a21e625d2\") " pod="openstack/nova-metadata-0" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.441775 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2028e28b-656a-493b-9803-f63a21e625d2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2028e28b-656a-493b-9803-f63a21e625d2\") " pod="openstack/nova-metadata-0" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.445497 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2028e28b-656a-493b-9803-f63a21e625d2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2028e28b-656a-493b-9803-f63a21e625d2\") " pod="openstack/nova-metadata-0" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.447927 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kzb\" (UniqueName: \"kubernetes.io/projected/2028e28b-656a-493b-9803-f63a21e625d2-kube-api-access-s2kzb\") pod \"nova-metadata-0\" (UID: \"2028e28b-656a-493b-9803-f63a21e625d2\") " pod="openstack/nova-metadata-0" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.470864 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 18 20:28:04 crc kubenswrapper[4950]: I0318 20:28:04.931875 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 18 20:28:04 crc kubenswrapper[4950]: W0318 20:28:04.957492 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2028e28b_656a_493b_9803_f63a21e625d2.slice/crio-6c29529ae8cb2600907f58225a88f7b0b30a7743b9190efe0de5f9424134a0f4 WatchSource:0}: Error finding container 6c29529ae8cb2600907f58225a88f7b0b30a7743b9190efe0de5f9424134a0f4: Status 404 returned error can't find the container with id 6c29529ae8cb2600907f58225a88f7b0b30a7743b9190efe0de5f9424134a0f4 Mar 18 20:28:05 crc kubenswrapper[4950]: I0318 20:28:05.069122 4950 generic.go:334] "Generic (PLEG): container finished" podID="9c71dbcb-5a73-4c85-8a1f-fa389a0e284f" containerID="0e584954e3a8b88f9cc604d88832e1d8a7009b70353dc58272712cf61225203c" exitCode=0 Mar 18 20:28:05 crc kubenswrapper[4950]: I0318 20:28:05.069397 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-n6btl" event={"ID":"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f","Type":"ContainerDied","Data":"0e584954e3a8b88f9cc604d88832e1d8a7009b70353dc58272712cf61225203c"} Mar 18 20:28:05 crc kubenswrapper[4950]: I0318 20:28:05.071239 4950 generic.go:334] "Generic (PLEG): container finished" podID="d2548eb6-5b6d-4ffd-9655-a766eecf2128" containerID="b646686c45c126d79b65dc5a0bf87aa8f95eac390bc0427a5b4e6f5ac1fc3993" exitCode=0 Mar 18 20:28:05 crc kubenswrapper[4950]: I0318 20:28:05.071275 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564428-7k4b9" event={"ID":"d2548eb6-5b6d-4ffd-9655-a766eecf2128","Type":"ContainerDied","Data":"b646686c45c126d79b65dc5a0bf87aa8f95eac390bc0427a5b4e6f5ac1fc3993"} Mar 18 20:28:05 crc kubenswrapper[4950]: I0318 20:28:05.072442 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2028e28b-656a-493b-9803-f63a21e625d2","Type":"ContainerStarted","Data":"6c29529ae8cb2600907f58225a88f7b0b30a7743b9190efe0de5f9424134a0f4"} Mar 18 20:28:05 crc kubenswrapper[4950]: I0318 20:28:05.464492 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 18 20:28:05 crc kubenswrapper[4950]: I0318 20:28:05.464541 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 18 20:28:05 crc kubenswrapper[4950]: I0318 20:28:05.500512 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc721dc9-f559-47b2-acca-3ca09cf54e27" path="/var/lib/kubelet/pods/fc721dc9-f559-47b2-acca-3ca09cf54e27/volumes" Mar 18 20:28:05 crc kubenswrapper[4950]: I0318 20:28:05.719901 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 18 20:28:05 crc kubenswrapper[4950]: I0318 20:28:05.720239 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 18 20:28:05 crc kubenswrapper[4950]: I0318 20:28:05.746461 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 18 20:28:05 crc kubenswrapper[4950]: I0318 20:28:05.798148 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:05 crc kubenswrapper[4950]: I0318 20:28:05.910579 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.007818 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78b7d887cf-gdzl7"] Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.008877 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" podUID="0013b445-540c-472b-9ead-44958805929b" containerName="dnsmasq-dns" containerID="cri-o://243cc4d14f94fd3cd99d4d03c8a4c6451725b42cad9f6e9d0c523e978a85fddd" gracePeriod=10 Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.091909 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.097724 4950 generic.go:334] "Generic (PLEG): container finished" podID="63712b81-a3d5-47d4-82bf-840c6e21d7a4" containerID="fc4e5553e36115b387543d76999d22ff7c0e8db181d2cebb7614644d2203da08" exitCode=137 Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.097814 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63712b81-a3d5-47d4-82bf-840c6e21d7a4","Type":"ContainerDied","Data":"fc4e5553e36115b387543d76999d22ff7c0e8db181d2cebb7614644d2203da08"} Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.097840 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"63712b81-a3d5-47d4-82bf-840c6e21d7a4","Type":"ContainerDied","Data":"a7e250b5af11d31a6b9d29c6a5c254deb22871636e8e7f0e3a1409b19361081d"} Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.097858 4950 scope.go:117] "RemoveContainer" containerID="fc4e5553e36115b387543d76999d22ff7c0e8db181d2cebb7614644d2203da08" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.100957 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2028e28b-656a-493b-9803-f63a21e625d2","Type":"ContainerStarted","Data":"56b047c7aefa3f6419c33bf1aee3c50ceef62f7344ad054854db568e4def31da"} Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.101011 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2028e28b-656a-493b-9803-f63a21e625d2","Type":"ContainerStarted","Data":"85d64af31861bce45b3aaea2c95bf24ccd68c37beb794dff2732a6085b7c2f1b"} Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.137037 4950 scope.go:117] "RemoveContainer" containerID="96f30879e79b3bf96f10dd7924b19e0da53f18f4d384f453b12e6ad9c6c84cfb" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.179182 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-combined-ca-bundle\") pod \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.179262 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-config-data\") pod \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.179357 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-sg-core-conf-yaml\") pod \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.179381 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63712b81-a3d5-47d4-82bf-840c6e21d7a4-log-httpd\") pod \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.179452 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-scripts\") pod \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.179479 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63712b81-a3d5-47d4-82bf-840c6e21d7a4-run-httpd\") pod \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.179510 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g58tv\" (UniqueName: \"kubernetes.io/projected/63712b81-a3d5-47d4-82bf-840c6e21d7a4-kube-api-access-g58tv\") pod \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\" (UID: \"63712b81-a3d5-47d4-82bf-840c6e21d7a4\") " Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.180161 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.181508 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63712b81-a3d5-47d4-82bf-840c6e21d7a4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "63712b81-a3d5-47d4-82bf-840c6e21d7a4" (UID: "63712b81-a3d5-47d4-82bf-840c6e21d7a4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.182067 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63712b81-a3d5-47d4-82bf-840c6e21d7a4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "63712b81-a3d5-47d4-82bf-840c6e21d7a4" (UID: "63712b81-a3d5-47d4-82bf-840c6e21d7a4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.193233 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-scripts" (OuterVolumeSpecName: "scripts") pod "63712b81-a3d5-47d4-82bf-840c6e21d7a4" (UID: "63712b81-a3d5-47d4-82bf-840c6e21d7a4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.230310 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63712b81-a3d5-47d4-82bf-840c6e21d7a4-kube-api-access-g58tv" (OuterVolumeSpecName: "kube-api-access-g58tv") pod "63712b81-a3d5-47d4-82bf-840c6e21d7a4" (UID: "63712b81-a3d5-47d4-82bf-840c6e21d7a4"). InnerVolumeSpecName "kube-api-access-g58tv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.235193 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.2351736779999998 podStartE2EDuration="2.235173678s" podCreationTimestamp="2026-03-18 20:28:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:28:06.153755231 +0000 UTC m=+1299.394597359" watchObservedRunningTime="2026-03-18 20:28:06.235173678 +0000 UTC m=+1299.476015546" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.243748 4950 scope.go:117] "RemoveContainer" containerID="3f056d86dba09f734bcfb4061f63d07b79afe4c7c92a4963af8cb90cb83346f4" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.261587 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "63712b81-a3d5-47d4-82bf-840c6e21d7a4" (UID: "63712b81-a3d5-47d4-82bf-840c6e21d7a4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.281182 4950 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.281214 4950 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63712b81-a3d5-47d4-82bf-840c6e21d7a4-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.281222 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.281230 4950 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63712b81-a3d5-47d4-82bf-840c6e21d7a4-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.281239 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g58tv\" (UniqueName: \"kubernetes.io/projected/63712b81-a3d5-47d4-82bf-840c6e21d7a4-kube-api-access-g58tv\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.324628 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" podUID="0013b445-540c-472b-9ead-44958805929b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.156:5353: connect: connection refused" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.344573 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "63712b81-a3d5-47d4-82bf-840c6e21d7a4" (UID: "63712b81-a3d5-47d4-82bf-840c6e21d7a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.359107 4950 scope.go:117] "RemoveContainer" containerID="7e1603ad7ab1a76e4e31de3ebc38f287b7e264b42c6d0a363eff00114cc96c45" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.368516 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-config-data" (OuterVolumeSpecName: "config-data") pod "63712b81-a3d5-47d4-82bf-840c6e21d7a4" (UID: "63712b81-a3d5-47d4-82bf-840c6e21d7a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.383179 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.383214 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63712b81-a3d5-47d4-82bf-840c6e21d7a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.394030 4950 scope.go:117] "RemoveContainer" containerID="fc4e5553e36115b387543d76999d22ff7c0e8db181d2cebb7614644d2203da08" Mar 18 20:28:06 crc kubenswrapper[4950]: E0318 20:28:06.394950 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc4e5553e36115b387543d76999d22ff7c0e8db181d2cebb7614644d2203da08\": container with ID starting with fc4e5553e36115b387543d76999d22ff7c0e8db181d2cebb7614644d2203da08 not found: ID does not exist" containerID="fc4e5553e36115b387543d76999d22ff7c0e8db181d2cebb7614644d2203da08" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.394995 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc4e5553e36115b387543d76999d22ff7c0e8db181d2cebb7614644d2203da08"} err="failed to get container status \"fc4e5553e36115b387543d76999d22ff7c0e8db181d2cebb7614644d2203da08\": rpc error: code = NotFound desc = could not find container \"fc4e5553e36115b387543d76999d22ff7c0e8db181d2cebb7614644d2203da08\": container with ID starting with fc4e5553e36115b387543d76999d22ff7c0e8db181d2cebb7614644d2203da08 not found: ID does not exist" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.395021 4950 scope.go:117] "RemoveContainer" containerID="96f30879e79b3bf96f10dd7924b19e0da53f18f4d384f453b12e6ad9c6c84cfb" Mar 18 20:28:06 crc kubenswrapper[4950]: E0318 20:28:06.398586 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96f30879e79b3bf96f10dd7924b19e0da53f18f4d384f453b12e6ad9c6c84cfb\": container with ID starting with 96f30879e79b3bf96f10dd7924b19e0da53f18f4d384f453b12e6ad9c6c84cfb not found: ID does not exist" containerID="96f30879e79b3bf96f10dd7924b19e0da53f18f4d384f453b12e6ad9c6c84cfb" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.398622 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96f30879e79b3bf96f10dd7924b19e0da53f18f4d384f453b12e6ad9c6c84cfb"} err="failed to get container status \"96f30879e79b3bf96f10dd7924b19e0da53f18f4d384f453b12e6ad9c6c84cfb\": rpc error: code = NotFound desc = could not find container \"96f30879e79b3bf96f10dd7924b19e0da53f18f4d384f453b12e6ad9c6c84cfb\": container with ID starting with 96f30879e79b3bf96f10dd7924b19e0da53f18f4d384f453b12e6ad9c6c84cfb not found: ID does not exist" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.398643 4950 scope.go:117] "RemoveContainer" containerID="3f056d86dba09f734bcfb4061f63d07b79afe4c7c92a4963af8cb90cb83346f4" Mar 18 20:28:06 crc kubenswrapper[4950]: E0318 20:28:06.398962 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f056d86dba09f734bcfb4061f63d07b79afe4c7c92a4963af8cb90cb83346f4\": container with ID starting with 3f056d86dba09f734bcfb4061f63d07b79afe4c7c92a4963af8cb90cb83346f4 not found: ID does not exist" containerID="3f056d86dba09f734bcfb4061f63d07b79afe4c7c92a4963af8cb90cb83346f4" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.398990 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f056d86dba09f734bcfb4061f63d07b79afe4c7c92a4963af8cb90cb83346f4"} err="failed to get container status \"3f056d86dba09f734bcfb4061f63d07b79afe4c7c92a4963af8cb90cb83346f4\": rpc error: code = NotFound desc = could not find container \"3f056d86dba09f734bcfb4061f63d07b79afe4c7c92a4963af8cb90cb83346f4\": container with ID starting with 3f056d86dba09f734bcfb4061f63d07b79afe4c7c92a4963af8cb90cb83346f4 not found: ID does not exist" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.399007 4950 scope.go:117] "RemoveContainer" containerID="7e1603ad7ab1a76e4e31de3ebc38f287b7e264b42c6d0a363eff00114cc96c45" Mar 18 20:28:06 crc kubenswrapper[4950]: E0318 20:28:06.400919 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e1603ad7ab1a76e4e31de3ebc38f287b7e264b42c6d0a363eff00114cc96c45\": container with ID starting with 7e1603ad7ab1a76e4e31de3ebc38f287b7e264b42c6d0a363eff00114cc96c45 not found: ID does not exist" containerID="7e1603ad7ab1a76e4e31de3ebc38f287b7e264b42c6d0a363eff00114cc96c45" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.400951 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e1603ad7ab1a76e4e31de3ebc38f287b7e264b42c6d0a363eff00114cc96c45"} err="failed to get container status \"7e1603ad7ab1a76e4e31de3ebc38f287b7e264b42c6d0a363eff00114cc96c45\": rpc error: code = NotFound desc = could not find container \"7e1603ad7ab1a76e4e31de3ebc38f287b7e264b42c6d0a363eff00114cc96c45\": container with ID starting with 7e1603ad7ab1a76e4e31de3ebc38f287b7e264b42c6d0a363eff00114cc96c45 not found: ID does not exist" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.549623 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0217d09c-db1e-45b6-bc64-c3eca777e7dc" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.173:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.549907 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0217d09c-db1e-45b6-bc64-c3eca777e7dc" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.173:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.567950 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-n6btl" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.686839 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txcrn\" (UniqueName: \"kubernetes.io/projected/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-kube-api-access-txcrn\") pod \"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f\" (UID: \"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f\") " Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.686881 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-combined-ca-bundle\") pod \"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f\" (UID: \"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f\") " Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.686963 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-config-data\") pod \"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f\" (UID: \"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f\") " Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.687018 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-scripts\") pod \"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f\" (UID: \"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f\") " Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.690449 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-scripts" (OuterVolumeSpecName: "scripts") pod "9c71dbcb-5a73-4c85-8a1f-fa389a0e284f" (UID: "9c71dbcb-5a73-4c85-8a1f-fa389a0e284f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.691783 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-kube-api-access-txcrn" (OuterVolumeSpecName: "kube-api-access-txcrn") pod "9c71dbcb-5a73-4c85-8a1f-fa389a0e284f" (UID: "9c71dbcb-5a73-4c85-8a1f-fa389a0e284f"). InnerVolumeSpecName "kube-api-access-txcrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.695177 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.735689 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c71dbcb-5a73-4c85-8a1f-fa389a0e284f" (UID: "9c71dbcb-5a73-4c85-8a1f-fa389a0e284f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.735778 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-config-data" (OuterVolumeSpecName: "config-data") pod "9c71dbcb-5a73-4c85-8a1f-fa389a0e284f" (UID: "9c71dbcb-5a73-4c85-8a1f-fa389a0e284f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.737366 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564428-7k4b9" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.788933 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-ovsdbserver-nb\") pod \"0013b445-540c-472b-9ead-44958805929b\" (UID: \"0013b445-540c-472b-9ead-44958805929b\") " Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.789065 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-dns-svc\") pod \"0013b445-540c-472b-9ead-44958805929b\" (UID: \"0013b445-540c-472b-9ead-44958805929b\") " Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.789551 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-869b7\" (UniqueName: \"kubernetes.io/projected/0013b445-540c-472b-9ead-44958805929b-kube-api-access-869b7\") pod \"0013b445-540c-472b-9ead-44958805929b\" (UID: \"0013b445-540c-472b-9ead-44958805929b\") " Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.789928 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-ovsdbserver-sb\") pod \"0013b445-540c-472b-9ead-44958805929b\" (UID: \"0013b445-540c-472b-9ead-44958805929b\") " Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.789963 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-config\") pod \"0013b445-540c-472b-9ead-44958805929b\" (UID: \"0013b445-540c-472b-9ead-44958805929b\") " Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.790397 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.790408 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txcrn\" (UniqueName: \"kubernetes.io/projected/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-kube-api-access-txcrn\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.790434 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.790443 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.793070 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0013b445-540c-472b-9ead-44958805929b-kube-api-access-869b7" (OuterVolumeSpecName: "kube-api-access-869b7") pod "0013b445-540c-472b-9ead-44958805929b" (UID: "0013b445-540c-472b-9ead-44958805929b"). InnerVolumeSpecName "kube-api-access-869b7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.862319 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0013b445-540c-472b-9ead-44958805929b" (UID: "0013b445-540c-472b-9ead-44958805929b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.869048 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0013b445-540c-472b-9ead-44958805929b" (UID: "0013b445-540c-472b-9ead-44958805929b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.873942 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-config" (OuterVolumeSpecName: "config") pod "0013b445-540c-472b-9ead-44958805929b" (UID: "0013b445-540c-472b-9ead-44958805929b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.887832 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0013b445-540c-472b-9ead-44958805929b" (UID: "0013b445-540c-472b-9ead-44958805929b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.892043 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n46v8\" (UniqueName: \"kubernetes.io/projected/d2548eb6-5b6d-4ffd-9655-a766eecf2128-kube-api-access-n46v8\") pod \"d2548eb6-5b6d-4ffd-9655-a766eecf2128\" (UID: \"d2548eb6-5b6d-4ffd-9655-a766eecf2128\") " Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.892463 4950 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.892483 4950 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.892494 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-869b7\" (UniqueName: \"kubernetes.io/projected/0013b445-540c-472b-9ead-44958805929b-kube-api-access-869b7\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.892504 4950 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.892514 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0013b445-540c-472b-9ead-44958805929b-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.898071 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2548eb6-5b6d-4ffd-9655-a766eecf2128-kube-api-access-n46v8" (OuterVolumeSpecName: "kube-api-access-n46v8") pod "d2548eb6-5b6d-4ffd-9655-a766eecf2128" (UID: "d2548eb6-5b6d-4ffd-9655-a766eecf2128"). InnerVolumeSpecName "kube-api-access-n46v8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:28:06 crc kubenswrapper[4950]: I0318 20:28:06.993559 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n46v8\" (UniqueName: \"kubernetes.io/projected/d2548eb6-5b6d-4ffd-9655-a766eecf2128-kube-api-access-n46v8\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.110253 4950 generic.go:334] "Generic (PLEG): container finished" podID="0013b445-540c-472b-9ead-44958805929b" containerID="243cc4d14f94fd3cd99d4d03c8a4c6451725b42cad9f6e9d0c523e978a85fddd" exitCode=0 Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.110313 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" event={"ID":"0013b445-540c-472b-9ead-44958805929b","Type":"ContainerDied","Data":"243cc4d14f94fd3cd99d4d03c8a4c6451725b42cad9f6e9d0c523e978a85fddd"} Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.110561 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" event={"ID":"0013b445-540c-472b-9ead-44958805929b","Type":"ContainerDied","Data":"c6fac8752917e54e8553c6e43ad7b4408d8b7d6f86f2a8443a53a0b24cacc620"} Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.110323 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78b7d887cf-gdzl7" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.110579 4950 scope.go:117] "RemoveContainer" containerID="243cc4d14f94fd3cd99d4d03c8a4c6451725b42cad9f6e9d0c523e978a85fddd" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.113991 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-n6btl" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.114279 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-n6btl" event={"ID":"9c71dbcb-5a73-4c85-8a1f-fa389a0e284f","Type":"ContainerDied","Data":"b63e54fb95815beecad31e64b2d7392d7bbbaecdc3036f209febbb4ab576686c"} Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.114371 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b63e54fb95815beecad31e64b2d7392d7bbbaecdc3036f209febbb4ab576686c" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.115904 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.118248 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564428-7k4b9" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.121496 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564428-7k4b9" event={"ID":"d2548eb6-5b6d-4ffd-9655-a766eecf2128","Type":"ContainerDied","Data":"94d9aa917c9114534ab1014bc992f7f0c41ddcba65a5467a3d7389f448218209"} Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.121610 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94d9aa917c9114534ab1014bc992f7f0c41ddcba65a5467a3d7389f448218209" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.139809 4950 scope.go:117] "RemoveContainer" containerID="77033a50291f56779c6983dc1cfa7d6d4a076aeb6a41ecb9042247ce5c227943" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.189799 4950 scope.go:117] "RemoveContainer" containerID="243cc4d14f94fd3cd99d4d03c8a4c6451725b42cad9f6e9d0c523e978a85fddd" Mar 18 20:28:07 crc kubenswrapper[4950]: E0318 20:28:07.190345 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"243cc4d14f94fd3cd99d4d03c8a4c6451725b42cad9f6e9d0c523e978a85fddd\": container with ID starting with 243cc4d14f94fd3cd99d4d03c8a4c6451725b42cad9f6e9d0c523e978a85fddd not found: ID does not exist" containerID="243cc4d14f94fd3cd99d4d03c8a4c6451725b42cad9f6e9d0c523e978a85fddd" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.190383 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"243cc4d14f94fd3cd99d4d03c8a4c6451725b42cad9f6e9d0c523e978a85fddd"} err="failed to get container status \"243cc4d14f94fd3cd99d4d03c8a4c6451725b42cad9f6e9d0c523e978a85fddd\": rpc error: code = NotFound desc = could not find container \"243cc4d14f94fd3cd99d4d03c8a4c6451725b42cad9f6e9d0c523e978a85fddd\": container with ID starting with 243cc4d14f94fd3cd99d4d03c8a4c6451725b42cad9f6e9d0c523e978a85fddd not found: ID does not exist" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.190419 4950 scope.go:117] "RemoveContainer" containerID="77033a50291f56779c6983dc1cfa7d6d4a076aeb6a41ecb9042247ce5c227943" Mar 18 20:28:07 crc kubenswrapper[4950]: E0318 20:28:07.191181 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77033a50291f56779c6983dc1cfa7d6d4a076aeb6a41ecb9042247ce5c227943\": container with ID starting with 77033a50291f56779c6983dc1cfa7d6d4a076aeb6a41ecb9042247ce5c227943 not found: ID does not exist" containerID="77033a50291f56779c6983dc1cfa7d6d4a076aeb6a41ecb9042247ce5c227943" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.191214 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77033a50291f56779c6983dc1cfa7d6d4a076aeb6a41ecb9042247ce5c227943"} err="failed to get container status \"77033a50291f56779c6983dc1cfa7d6d4a076aeb6a41ecb9042247ce5c227943\": rpc error: code = NotFound desc = could not find container \"77033a50291f56779c6983dc1cfa7d6d4a076aeb6a41ecb9042247ce5c227943\": container with ID starting with 77033a50291f56779c6983dc1cfa7d6d4a076aeb6a41ecb9042247ce5c227943 not found: ID does not exist" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.201968 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.223713 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.243072 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:28:07 crc kubenswrapper[4950]: E0318 20:28:07.243456 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63712b81-a3d5-47d4-82bf-840c6e21d7a4" containerName="ceilometer-notification-agent" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.243469 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="63712b81-a3d5-47d4-82bf-840c6e21d7a4" containerName="ceilometer-notification-agent" Mar 18 20:28:07 crc kubenswrapper[4950]: E0318 20:28:07.243479 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63712b81-a3d5-47d4-82bf-840c6e21d7a4" containerName="sg-core" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.243486 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="63712b81-a3d5-47d4-82bf-840c6e21d7a4" containerName="sg-core" Mar 18 20:28:07 crc kubenswrapper[4950]: E0318 20:28:07.243493 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0013b445-540c-472b-9ead-44958805929b" containerName="dnsmasq-dns" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.243499 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="0013b445-540c-472b-9ead-44958805929b" containerName="dnsmasq-dns" Mar 18 20:28:07 crc kubenswrapper[4950]: E0318 20:28:07.243508 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0013b445-540c-472b-9ead-44958805929b" containerName="init" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.243514 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="0013b445-540c-472b-9ead-44958805929b" containerName="init" Mar 18 20:28:07 crc kubenswrapper[4950]: E0318 20:28:07.243524 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63712b81-a3d5-47d4-82bf-840c6e21d7a4" containerName="ceilometer-central-agent" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.243530 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="63712b81-a3d5-47d4-82bf-840c6e21d7a4" containerName="ceilometer-central-agent" Mar 18 20:28:07 crc kubenswrapper[4950]: E0318 20:28:07.243542 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c71dbcb-5a73-4c85-8a1f-fa389a0e284f" containerName="nova-manage" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.243549 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c71dbcb-5a73-4c85-8a1f-fa389a0e284f" containerName="nova-manage" Mar 18 20:28:07 crc kubenswrapper[4950]: E0318 20:28:07.243564 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2548eb6-5b6d-4ffd-9655-a766eecf2128" containerName="oc" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.243569 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2548eb6-5b6d-4ffd-9655-a766eecf2128" containerName="oc" Mar 18 20:28:07 crc kubenswrapper[4950]: E0318 20:28:07.243595 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63712b81-a3d5-47d4-82bf-840c6e21d7a4" containerName="proxy-httpd" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.243602 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="63712b81-a3d5-47d4-82bf-840c6e21d7a4" containerName="proxy-httpd" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.243780 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="63712b81-a3d5-47d4-82bf-840c6e21d7a4" containerName="ceilometer-central-agent" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.243799 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="63712b81-a3d5-47d4-82bf-840c6e21d7a4" containerName="sg-core" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.243810 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2548eb6-5b6d-4ffd-9655-a766eecf2128" containerName="oc" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.243822 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="0013b445-540c-472b-9ead-44958805929b" containerName="dnsmasq-dns" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.243833 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="63712b81-a3d5-47d4-82bf-840c6e21d7a4" containerName="ceilometer-notification-agent" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.243847 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="63712b81-a3d5-47d4-82bf-840c6e21d7a4" containerName="proxy-httpd" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.243860 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c71dbcb-5a73-4c85-8a1f-fa389a0e284f" containerName="nova-manage" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.246096 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.249397 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.257739 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.266596 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.277038 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78b7d887cf-gdzl7"] Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.290456 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78b7d887cf-gdzl7"] Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.310181 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rw66\" (UniqueName: \"kubernetes.io/projected/e1ca2e06-431e-45b9-b492-20c491324945-kube-api-access-4rw66\") pod \"ceilometer-0\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.310281 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-scripts\") pod \"ceilometer-0\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.310340 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.310364 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-config-data\") pod \"ceilometer-0\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.310390 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.310427 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1ca2e06-431e-45b9-b492-20c491324945-log-httpd\") pod \"ceilometer-0\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.310447 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1ca2e06-431e-45b9-b492-20c491324945-run-httpd\") pod \"ceilometer-0\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.360708 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.360911 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0217d09c-db1e-45b6-bc64-c3eca777e7dc" containerName="nova-api-log" containerID="cri-o://e8912d9e397ece5941a51539bc31e219f238391233a8d61b16965058aa4cff40" gracePeriod=30 Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.361267 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0217d09c-db1e-45b6-bc64-c3eca777e7dc" containerName="nova-api-api" containerID="cri-o://cf4c56175b75b3f1ced1fd7d0c734b9230cb2c9a634abf085dc8bdb522f80551" gracePeriod=30 Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.375014 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.390753 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.412218 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rw66\" (UniqueName: \"kubernetes.io/projected/e1ca2e06-431e-45b9-b492-20c491324945-kube-api-access-4rw66\") pod \"ceilometer-0\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.412489 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-scripts\") pod \"ceilometer-0\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.412622 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.412703 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-config-data\") pod \"ceilometer-0\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.412786 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.412860 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1ca2e06-431e-45b9-b492-20c491324945-log-httpd\") pod \"ceilometer-0\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.412925 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1ca2e06-431e-45b9-b492-20c491324945-run-httpd\") pod \"ceilometer-0\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.413478 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1ca2e06-431e-45b9-b492-20c491324945-run-httpd\") pod \"ceilometer-0\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.413610 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1ca2e06-431e-45b9-b492-20c491324945-log-httpd\") pod \"ceilometer-0\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.417660 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-scripts\") pod \"ceilometer-0\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.419721 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-config-data\") pod \"ceilometer-0\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.422091 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.432083 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rw66\" (UniqueName: \"kubernetes.io/projected/e1ca2e06-431e-45b9-b492-20c491324945-kube-api-access-4rw66\") pod \"ceilometer-0\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.433707 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.488901 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0013b445-540c-472b-9ead-44958805929b" path="/var/lib/kubelet/pods/0013b445-540c-472b-9ead-44958805929b/volumes" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.489526 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63712b81-a3d5-47d4-82bf-840c6e21d7a4" path="/var/lib/kubelet/pods/63712b81-a3d5-47d4-82bf-840c6e21d7a4/volumes" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.568892 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.823204 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564422-glkgm"] Mar 18 20:28:07 crc kubenswrapper[4950]: I0318 20:28:07.834349 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564422-glkgm"] Mar 18 20:28:08 crc kubenswrapper[4950]: W0318 20:28:08.122630 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1ca2e06_431e_45b9_b492_20c491324945.slice/crio-e0e57a505f8fc4a0fcd2a90c2f236e3d7e404ba5e0e4cca9313bea52628ce377 WatchSource:0}: Error finding container e0e57a505f8fc4a0fcd2a90c2f236e3d7e404ba5e0e4cca9313bea52628ce377: Status 404 returned error can't find the container with id e0e57a505f8fc4a0fcd2a90c2f236e3d7e404ba5e0e4cca9313bea52628ce377 Mar 18 20:28:08 crc kubenswrapper[4950]: I0318 20:28:08.131572 4950 generic.go:334] "Generic (PLEG): container finished" podID="0217d09c-db1e-45b6-bc64-c3eca777e7dc" containerID="e8912d9e397ece5941a51539bc31e219f238391233a8d61b16965058aa4cff40" exitCode=143 Mar 18 20:28:08 crc kubenswrapper[4950]: I0318 20:28:08.131735 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0217d09c-db1e-45b6-bc64-c3eca777e7dc","Type":"ContainerDied","Data":"e8912d9e397ece5941a51539bc31e219f238391233a8d61b16965058aa4cff40"} Mar 18 20:28:08 crc kubenswrapper[4950]: I0318 20:28:08.132809 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65" containerName="nova-scheduler-scheduler" containerID="cri-o://3466f5503ac2a971a31d2cb12b54e4dbb107fb69f7f31b74b2511edd002cca64" gracePeriod=30 Mar 18 20:28:08 crc kubenswrapper[4950]: I0318 20:28:08.133078 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2028e28b-656a-493b-9803-f63a21e625d2" containerName="nova-metadata-log" containerID="cri-o://85d64af31861bce45b3aaea2c95bf24ccd68c37beb794dff2732a6085b7c2f1b" gracePeriod=30 Mar 18 20:28:08 crc kubenswrapper[4950]: I0318 20:28:08.133540 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2028e28b-656a-493b-9803-f63a21e625d2" containerName="nova-metadata-metadata" containerID="cri-o://56b047c7aefa3f6419c33bf1aee3c50ceef62f7344ad054854db568e4def31da" gracePeriod=30 Mar 18 20:28:08 crc kubenswrapper[4950]: I0318 20:28:08.141260 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:28:08 crc kubenswrapper[4950]: I0318 20:28:08.947423 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.140920 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1ca2e06-431e-45b9-b492-20c491324945","Type":"ContainerStarted","Data":"9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168"} Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.141196 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1ca2e06-431e-45b9-b492-20c491324945","Type":"ContainerStarted","Data":"e0e57a505f8fc4a0fcd2a90c2f236e3d7e404ba5e0e4cca9313bea52628ce377"} Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.142227 4950 generic.go:334] "Generic (PLEG): container finished" podID="ff62857a-136b-4b75-abf8-655bab5c8984" containerID="9b8cd39ebaa272935fa8e4cd424226e975751123cef5c91404d6c296fc9242b4" exitCode=0 Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.142267 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lzx4p" event={"ID":"ff62857a-136b-4b75-abf8-655bab5c8984","Type":"ContainerDied","Data":"9b8cd39ebaa272935fa8e4cd424226e975751123cef5c91404d6c296fc9242b4"} Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.144076 4950 generic.go:334] "Generic (PLEG): container finished" podID="2028e28b-656a-493b-9803-f63a21e625d2" containerID="56b047c7aefa3f6419c33bf1aee3c50ceef62f7344ad054854db568e4def31da" exitCode=0 Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.144095 4950 generic.go:334] "Generic (PLEG): container finished" podID="2028e28b-656a-493b-9803-f63a21e625d2" containerID="85d64af31861bce45b3aaea2c95bf24ccd68c37beb794dff2732a6085b7c2f1b" exitCode=143 Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.144110 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2028e28b-656a-493b-9803-f63a21e625d2","Type":"ContainerDied","Data":"56b047c7aefa3f6419c33bf1aee3c50ceef62f7344ad054854db568e4def31da"} Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.144126 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2028e28b-656a-493b-9803-f63a21e625d2","Type":"ContainerDied","Data":"85d64af31861bce45b3aaea2c95bf24ccd68c37beb794dff2732a6085b7c2f1b"} Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.144135 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2028e28b-656a-493b-9803-f63a21e625d2","Type":"ContainerDied","Data":"6c29529ae8cb2600907f58225a88f7b0b30a7743b9190efe0de5f9424134a0f4"} Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.144150 4950 scope.go:117] "RemoveContainer" containerID="56b047c7aefa3f6419c33bf1aee3c50ceef62f7344ad054854db568e4def31da" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.144265 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.148030 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2kzb\" (UniqueName: \"kubernetes.io/projected/2028e28b-656a-493b-9803-f63a21e625d2-kube-api-access-s2kzb\") pod \"2028e28b-656a-493b-9803-f63a21e625d2\" (UID: \"2028e28b-656a-493b-9803-f63a21e625d2\") " Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.148073 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2028e28b-656a-493b-9803-f63a21e625d2-combined-ca-bundle\") pod \"2028e28b-656a-493b-9803-f63a21e625d2\" (UID: \"2028e28b-656a-493b-9803-f63a21e625d2\") " Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.148199 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2028e28b-656a-493b-9803-f63a21e625d2-config-data\") pod \"2028e28b-656a-493b-9803-f63a21e625d2\" (UID: \"2028e28b-656a-493b-9803-f63a21e625d2\") " Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.148275 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2028e28b-656a-493b-9803-f63a21e625d2-nova-metadata-tls-certs\") pod \"2028e28b-656a-493b-9803-f63a21e625d2\" (UID: \"2028e28b-656a-493b-9803-f63a21e625d2\") " Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.148312 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2028e28b-656a-493b-9803-f63a21e625d2-logs\") pod \"2028e28b-656a-493b-9803-f63a21e625d2\" (UID: \"2028e28b-656a-493b-9803-f63a21e625d2\") " Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.149078 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2028e28b-656a-493b-9803-f63a21e625d2-logs" (OuterVolumeSpecName: "logs") pod "2028e28b-656a-493b-9803-f63a21e625d2" (UID: "2028e28b-656a-493b-9803-f63a21e625d2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.171300 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2028e28b-656a-493b-9803-f63a21e625d2-kube-api-access-s2kzb" (OuterVolumeSpecName: "kube-api-access-s2kzb") pod "2028e28b-656a-493b-9803-f63a21e625d2" (UID: "2028e28b-656a-493b-9803-f63a21e625d2"). InnerVolumeSpecName "kube-api-access-s2kzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.196588 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2028e28b-656a-493b-9803-f63a21e625d2-config-data" (OuterVolumeSpecName: "config-data") pod "2028e28b-656a-493b-9803-f63a21e625d2" (UID: "2028e28b-656a-493b-9803-f63a21e625d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.197950 4950 scope.go:117] "RemoveContainer" containerID="85d64af31861bce45b3aaea2c95bf24ccd68c37beb794dff2732a6085b7c2f1b" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.219592 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2028e28b-656a-493b-9803-f63a21e625d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2028e28b-656a-493b-9803-f63a21e625d2" (UID: "2028e28b-656a-493b-9803-f63a21e625d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.244826 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2028e28b-656a-493b-9803-f63a21e625d2-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "2028e28b-656a-493b-9803-f63a21e625d2" (UID: "2028e28b-656a-493b-9803-f63a21e625d2"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.250013 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2028e28b-656a-493b-9803-f63a21e625d2-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.250050 4950 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2028e28b-656a-493b-9803-f63a21e625d2-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.250061 4950 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2028e28b-656a-493b-9803-f63a21e625d2-logs\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.250068 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2kzb\" (UniqueName: \"kubernetes.io/projected/2028e28b-656a-493b-9803-f63a21e625d2-kube-api-access-s2kzb\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.250076 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2028e28b-656a-493b-9803-f63a21e625d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.309717 4950 scope.go:117] "RemoveContainer" containerID="56b047c7aefa3f6419c33bf1aee3c50ceef62f7344ad054854db568e4def31da" Mar 18 20:28:09 crc kubenswrapper[4950]: E0318 20:28:09.310026 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56b047c7aefa3f6419c33bf1aee3c50ceef62f7344ad054854db568e4def31da\": container with ID starting with 56b047c7aefa3f6419c33bf1aee3c50ceef62f7344ad054854db568e4def31da not found: ID does not exist" containerID="56b047c7aefa3f6419c33bf1aee3c50ceef62f7344ad054854db568e4def31da" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.310058 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56b047c7aefa3f6419c33bf1aee3c50ceef62f7344ad054854db568e4def31da"} err="failed to get container status \"56b047c7aefa3f6419c33bf1aee3c50ceef62f7344ad054854db568e4def31da\": rpc error: code = NotFound desc = could not find container \"56b047c7aefa3f6419c33bf1aee3c50ceef62f7344ad054854db568e4def31da\": container with ID starting with 56b047c7aefa3f6419c33bf1aee3c50ceef62f7344ad054854db568e4def31da not found: ID does not exist" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.310078 4950 scope.go:117] "RemoveContainer" containerID="85d64af31861bce45b3aaea2c95bf24ccd68c37beb794dff2732a6085b7c2f1b" Mar 18 20:28:09 crc kubenswrapper[4950]: E0318 20:28:09.310524 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85d64af31861bce45b3aaea2c95bf24ccd68c37beb794dff2732a6085b7c2f1b\": container with ID starting with 85d64af31861bce45b3aaea2c95bf24ccd68c37beb794dff2732a6085b7c2f1b not found: ID does not exist" containerID="85d64af31861bce45b3aaea2c95bf24ccd68c37beb794dff2732a6085b7c2f1b" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.310546 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85d64af31861bce45b3aaea2c95bf24ccd68c37beb794dff2732a6085b7c2f1b"} err="failed to get container status \"85d64af31861bce45b3aaea2c95bf24ccd68c37beb794dff2732a6085b7c2f1b\": rpc error: code = NotFound desc = could not find container \"85d64af31861bce45b3aaea2c95bf24ccd68c37beb794dff2732a6085b7c2f1b\": container with ID starting with 85d64af31861bce45b3aaea2c95bf24ccd68c37beb794dff2732a6085b7c2f1b not found: ID does not exist" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.310558 4950 scope.go:117] "RemoveContainer" containerID="56b047c7aefa3f6419c33bf1aee3c50ceef62f7344ad054854db568e4def31da" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.310819 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56b047c7aefa3f6419c33bf1aee3c50ceef62f7344ad054854db568e4def31da"} err="failed to get container status \"56b047c7aefa3f6419c33bf1aee3c50ceef62f7344ad054854db568e4def31da\": rpc error: code = NotFound desc = could not find container \"56b047c7aefa3f6419c33bf1aee3c50ceef62f7344ad054854db568e4def31da\": container with ID starting with 56b047c7aefa3f6419c33bf1aee3c50ceef62f7344ad054854db568e4def31da not found: ID does not exist" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.310840 4950 scope.go:117] "RemoveContainer" containerID="85d64af31861bce45b3aaea2c95bf24ccd68c37beb794dff2732a6085b7c2f1b" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.311050 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85d64af31861bce45b3aaea2c95bf24ccd68c37beb794dff2732a6085b7c2f1b"} err="failed to get container status \"85d64af31861bce45b3aaea2c95bf24ccd68c37beb794dff2732a6085b7c2f1b\": rpc error: code = NotFound desc = could not find container \"85d64af31861bce45b3aaea2c95bf24ccd68c37beb794dff2732a6085b7c2f1b\": container with ID starting with 85d64af31861bce45b3aaea2c95bf24ccd68c37beb794dff2732a6085b7c2f1b not found: ID does not exist" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.488490 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b60e86f9-3271-4263-bcaa-c4629cd1346d" path="/var/lib/kubelet/pods/b60e86f9-3271-4263-bcaa-c4629cd1346d/volumes" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.495789 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.507757 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.520039 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 18 20:28:09 crc kubenswrapper[4950]: E0318 20:28:09.520522 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2028e28b-656a-493b-9803-f63a21e625d2" containerName="nova-metadata-metadata" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.520543 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="2028e28b-656a-493b-9803-f63a21e625d2" containerName="nova-metadata-metadata" Mar 18 20:28:09 crc kubenswrapper[4950]: E0318 20:28:09.520571 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2028e28b-656a-493b-9803-f63a21e625d2" containerName="nova-metadata-log" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.520579 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="2028e28b-656a-493b-9803-f63a21e625d2" containerName="nova-metadata-log" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.520778 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="2028e28b-656a-493b-9803-f63a21e625d2" containerName="nova-metadata-log" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.520804 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="2028e28b-656a-493b-9803-f63a21e625d2" containerName="nova-metadata-metadata" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.521767 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.524233 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.525700 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.530403 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.657076 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ade0ed1d-2241-456b-b72f-3236fbc6ed45-logs\") pod \"nova-metadata-0\" (UID: \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\") " pod="openstack/nova-metadata-0" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.657383 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ade0ed1d-2241-456b-b72f-3236fbc6ed45-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\") " pod="openstack/nova-metadata-0" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.657425 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ade0ed1d-2241-456b-b72f-3236fbc6ed45-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\") " pod="openstack/nova-metadata-0" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.657552 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ade0ed1d-2241-456b-b72f-3236fbc6ed45-config-data\") pod \"nova-metadata-0\" (UID: \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\") " pod="openstack/nova-metadata-0" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.657879 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8trkg\" (UniqueName: \"kubernetes.io/projected/ade0ed1d-2241-456b-b72f-3236fbc6ed45-kube-api-access-8trkg\") pod \"nova-metadata-0\" (UID: \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\") " pod="openstack/nova-metadata-0" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.759447 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8trkg\" (UniqueName: \"kubernetes.io/projected/ade0ed1d-2241-456b-b72f-3236fbc6ed45-kube-api-access-8trkg\") pod \"nova-metadata-0\" (UID: \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\") " pod="openstack/nova-metadata-0" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.759509 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ade0ed1d-2241-456b-b72f-3236fbc6ed45-logs\") pod \"nova-metadata-0\" (UID: \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\") " pod="openstack/nova-metadata-0" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.759560 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ade0ed1d-2241-456b-b72f-3236fbc6ed45-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\") " pod="openstack/nova-metadata-0" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.759583 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ade0ed1d-2241-456b-b72f-3236fbc6ed45-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\") " pod="openstack/nova-metadata-0" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.759625 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ade0ed1d-2241-456b-b72f-3236fbc6ed45-config-data\") pod \"nova-metadata-0\" (UID: \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\") " pod="openstack/nova-metadata-0" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.759918 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ade0ed1d-2241-456b-b72f-3236fbc6ed45-logs\") pod \"nova-metadata-0\" (UID: \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\") " pod="openstack/nova-metadata-0" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.765385 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ade0ed1d-2241-456b-b72f-3236fbc6ed45-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\") " pod="openstack/nova-metadata-0" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.765546 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ade0ed1d-2241-456b-b72f-3236fbc6ed45-config-data\") pod \"nova-metadata-0\" (UID: \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\") " pod="openstack/nova-metadata-0" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.769122 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ade0ed1d-2241-456b-b72f-3236fbc6ed45-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\") " pod="openstack/nova-metadata-0" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.776865 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8trkg\" (UniqueName: \"kubernetes.io/projected/ade0ed1d-2241-456b-b72f-3236fbc6ed45-kube-api-access-8trkg\") pod \"nova-metadata-0\" (UID: \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\") " pod="openstack/nova-metadata-0" Mar 18 20:28:09 crc kubenswrapper[4950]: I0318 20:28:09.836655 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.138004 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.165940 4950 generic.go:334] "Generic (PLEG): container finished" podID="d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65" containerID="3466f5503ac2a971a31d2cb12b54e4dbb107fb69f7f31b74b2511edd002cca64" exitCode=0 Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.166027 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65","Type":"ContainerDied","Data":"3466f5503ac2a971a31d2cb12b54e4dbb107fb69f7f31b74b2511edd002cca64"} Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.166054 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65","Type":"ContainerDied","Data":"6f2822f65df5bffa4cabe7fee26ad23f67964917d15200cc4fd5f4b42350ea09"} Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.166070 4950 scope.go:117] "RemoveContainer" containerID="3466f5503ac2a971a31d2cb12b54e4dbb107fb69f7f31b74b2511edd002cca64" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.166185 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.167721 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pzxt\" (UniqueName: \"kubernetes.io/projected/d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65-kube-api-access-4pzxt\") pod \"d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65\" (UID: \"d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65\") " Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.167789 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65-combined-ca-bundle\") pod \"d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65\" (UID: \"d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65\") " Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.167811 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65-config-data\") pod \"d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65\" (UID: \"d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65\") " Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.184792 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1ca2e06-431e-45b9-b492-20c491324945","Type":"ContainerStarted","Data":"8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b"} Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.184896 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65-kube-api-access-4pzxt" (OuterVolumeSpecName: "kube-api-access-4pzxt") pod "d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65" (UID: "d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65"). InnerVolumeSpecName "kube-api-access-4pzxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.230493 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65-config-data" (OuterVolumeSpecName: "config-data") pod "d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65" (UID: "d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.267477 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65" (UID: "d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.269482 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pzxt\" (UniqueName: \"kubernetes.io/projected/d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65-kube-api-access-4pzxt\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.269503 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.269511 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.285764 4950 scope.go:117] "RemoveContainer" containerID="3466f5503ac2a971a31d2cb12b54e4dbb107fb69f7f31b74b2511edd002cca64" Mar 18 20:28:10 crc kubenswrapper[4950]: E0318 20:28:10.286131 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3466f5503ac2a971a31d2cb12b54e4dbb107fb69f7f31b74b2511edd002cca64\": container with ID starting with 3466f5503ac2a971a31d2cb12b54e4dbb107fb69f7f31b74b2511edd002cca64 not found: ID does not exist" containerID="3466f5503ac2a971a31d2cb12b54e4dbb107fb69f7f31b74b2511edd002cca64" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.286341 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3466f5503ac2a971a31d2cb12b54e4dbb107fb69f7f31b74b2511edd002cca64"} err="failed to get container status \"3466f5503ac2a971a31d2cb12b54e4dbb107fb69f7f31b74b2511edd002cca64\": rpc error: code = NotFound desc = could not find container \"3466f5503ac2a971a31d2cb12b54e4dbb107fb69f7f31b74b2511edd002cca64\": container with ID starting with 3466f5503ac2a971a31d2cb12b54e4dbb107fb69f7f31b74b2511edd002cca64 not found: ID does not exist" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.385739 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 18 20:28:10 crc kubenswrapper[4950]: W0318 20:28:10.392603 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podade0ed1d_2241_456b_b72f_3236fbc6ed45.slice/crio-5820e4b400c1e3437a059bb369b1ba4bd00d87dc642f281e167d90a04ede89a0 WatchSource:0}: Error finding container 5820e4b400c1e3437a059bb369b1ba4bd00d87dc642f281e167d90a04ede89a0: Status 404 returned error can't find the container with id 5820e4b400c1e3437a059bb369b1ba4bd00d87dc642f281e167d90a04ede89a0 Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.510404 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.518706 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.553178 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 18 20:28:10 crc kubenswrapper[4950]: E0318 20:28:10.553478 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65" containerName="nova-scheduler-scheduler" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.553496 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65" containerName="nova-scheduler-scheduler" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.553681 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65" containerName="nova-scheduler-scheduler" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.554198 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.555902 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.581588 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.586772 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lzx4p" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.677014 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvdtp\" (UniqueName: \"kubernetes.io/projected/cbbcbf5a-8667-4e11-ad3e-f72b30084b6d-kube-api-access-pvdtp\") pod \"nova-scheduler-0\" (UID: \"cbbcbf5a-8667-4e11-ad3e-f72b30084b6d\") " pod="openstack/nova-scheduler-0" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.679111 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbbcbf5a-8667-4e11-ad3e-f72b30084b6d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cbbcbf5a-8667-4e11-ad3e-f72b30084b6d\") " pod="openstack/nova-scheduler-0" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.679257 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbbcbf5a-8667-4e11-ad3e-f72b30084b6d-config-data\") pod \"nova-scheduler-0\" (UID: \"cbbcbf5a-8667-4e11-ad3e-f72b30084b6d\") " pod="openstack/nova-scheduler-0" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.780512 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rwsw\" (UniqueName: \"kubernetes.io/projected/ff62857a-136b-4b75-abf8-655bab5c8984-kube-api-access-7rwsw\") pod \"ff62857a-136b-4b75-abf8-655bab5c8984\" (UID: \"ff62857a-136b-4b75-abf8-655bab5c8984\") " Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.780708 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff62857a-136b-4b75-abf8-655bab5c8984-combined-ca-bundle\") pod \"ff62857a-136b-4b75-abf8-655bab5c8984\" (UID: \"ff62857a-136b-4b75-abf8-655bab5c8984\") " Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.780750 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff62857a-136b-4b75-abf8-655bab5c8984-scripts\") pod \"ff62857a-136b-4b75-abf8-655bab5c8984\" (UID: \"ff62857a-136b-4b75-abf8-655bab5c8984\") " Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.780804 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff62857a-136b-4b75-abf8-655bab5c8984-config-data\") pod \"ff62857a-136b-4b75-abf8-655bab5c8984\" (UID: \"ff62857a-136b-4b75-abf8-655bab5c8984\") " Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.781210 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvdtp\" (UniqueName: \"kubernetes.io/projected/cbbcbf5a-8667-4e11-ad3e-f72b30084b6d-kube-api-access-pvdtp\") pod \"nova-scheduler-0\" (UID: \"cbbcbf5a-8667-4e11-ad3e-f72b30084b6d\") " pod="openstack/nova-scheduler-0" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.781355 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbbcbf5a-8667-4e11-ad3e-f72b30084b6d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cbbcbf5a-8667-4e11-ad3e-f72b30084b6d\") " pod="openstack/nova-scheduler-0" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.781390 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbbcbf5a-8667-4e11-ad3e-f72b30084b6d-config-data\") pod \"nova-scheduler-0\" (UID: \"cbbcbf5a-8667-4e11-ad3e-f72b30084b6d\") " pod="openstack/nova-scheduler-0" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.788347 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff62857a-136b-4b75-abf8-655bab5c8984-scripts" (OuterVolumeSpecName: "scripts") pod "ff62857a-136b-4b75-abf8-655bab5c8984" (UID: "ff62857a-136b-4b75-abf8-655bab5c8984"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.788351 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff62857a-136b-4b75-abf8-655bab5c8984-kube-api-access-7rwsw" (OuterVolumeSpecName: "kube-api-access-7rwsw") pod "ff62857a-136b-4b75-abf8-655bab5c8984" (UID: "ff62857a-136b-4b75-abf8-655bab5c8984"). InnerVolumeSpecName "kube-api-access-7rwsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.788916 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbbcbf5a-8667-4e11-ad3e-f72b30084b6d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cbbcbf5a-8667-4e11-ad3e-f72b30084b6d\") " pod="openstack/nova-scheduler-0" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.789049 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbbcbf5a-8667-4e11-ad3e-f72b30084b6d-config-data\") pod \"nova-scheduler-0\" (UID: \"cbbcbf5a-8667-4e11-ad3e-f72b30084b6d\") " pod="openstack/nova-scheduler-0" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.800343 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvdtp\" (UniqueName: \"kubernetes.io/projected/cbbcbf5a-8667-4e11-ad3e-f72b30084b6d-kube-api-access-pvdtp\") pod \"nova-scheduler-0\" (UID: \"cbbcbf5a-8667-4e11-ad3e-f72b30084b6d\") " pod="openstack/nova-scheduler-0" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.819538 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff62857a-136b-4b75-abf8-655bab5c8984-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff62857a-136b-4b75-abf8-655bab5c8984" (UID: "ff62857a-136b-4b75-abf8-655bab5c8984"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.822541 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff62857a-136b-4b75-abf8-655bab5c8984-config-data" (OuterVolumeSpecName: "config-data") pod "ff62857a-136b-4b75-abf8-655bab5c8984" (UID: "ff62857a-136b-4b75-abf8-655bab5c8984"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.882585 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rwsw\" (UniqueName: \"kubernetes.io/projected/ff62857a-136b-4b75-abf8-655bab5c8984-kube-api-access-7rwsw\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.882617 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff62857a-136b-4b75-abf8-655bab5c8984-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.882626 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff62857a-136b-4b75-abf8-655bab5c8984-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.882634 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff62857a-136b-4b75-abf8-655bab5c8984-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:10 crc kubenswrapper[4950]: I0318 20:28:10.908477 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.209333 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1ca2e06-431e-45b9-b492-20c491324945","Type":"ContainerStarted","Data":"1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2"} Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.217675 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ade0ed1d-2241-456b-b72f-3236fbc6ed45","Type":"ContainerStarted","Data":"4b4d20fec20d82446b0a36dba6ec857459a48495cecba2ff3f21ce5a342e2289"} Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.217824 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ade0ed1d-2241-456b-b72f-3236fbc6ed45","Type":"ContainerStarted","Data":"fc3cba41084950eb31f480a81d4c3033b0a539a97624c7ff2e321821201872f9"} Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.217905 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ade0ed1d-2241-456b-b72f-3236fbc6ed45","Type":"ContainerStarted","Data":"5820e4b400c1e3437a059bb369b1ba4bd00d87dc642f281e167d90a04ede89a0"} Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.220549 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lzx4p" event={"ID":"ff62857a-136b-4b75-abf8-655bab5c8984","Type":"ContainerDied","Data":"38eb01a41c0da6a858cd5db49e4dc60f76864a5d13e1fbb92c5b25cb16129b18"} Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.220663 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38eb01a41c0da6a858cd5db49e4dc60f76864a5d13e1fbb92c5b25cb16129b18" Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.220775 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lzx4p" Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.249267 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.249246648 podStartE2EDuration="2.249246648s" podCreationTimestamp="2026-03-18 20:28:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:28:11.239585009 +0000 UTC m=+1304.480426877" watchObservedRunningTime="2026-03-18 20:28:11.249246648 +0000 UTC m=+1304.490088516" Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.274116 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 18 20:28:11 crc kubenswrapper[4950]: E0318 20:28:11.274581 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff62857a-136b-4b75-abf8-655bab5c8984" containerName="nova-cell1-conductor-db-sync" Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.274596 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff62857a-136b-4b75-abf8-655bab5c8984" containerName="nova-cell1-conductor-db-sync" Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.274753 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff62857a-136b-4b75-abf8-655bab5c8984" containerName="nova-cell1-conductor-db-sync" Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.275336 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.286094 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.290809 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6knm\" (UniqueName: \"kubernetes.io/projected/41637a9b-2c22-4627-85df-b902f8d35ceb-kube-api-access-p6knm\") pod \"nova-cell1-conductor-0\" (UID: \"41637a9b-2c22-4627-85df-b902f8d35ceb\") " pod="openstack/nova-cell1-conductor-0" Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.290872 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41637a9b-2c22-4627-85df-b902f8d35ceb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"41637a9b-2c22-4627-85df-b902f8d35ceb\") " pod="openstack/nova-cell1-conductor-0" Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.291012 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41637a9b-2c22-4627-85df-b902f8d35ceb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"41637a9b-2c22-4627-85df-b902f8d35ceb\") " pod="openstack/nova-cell1-conductor-0" Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.330095 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.359852 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 18 20:28:11 crc kubenswrapper[4950]: W0318 20:28:11.366349 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbbcbf5a_8667_4e11_ad3e_f72b30084b6d.slice/crio-a97374e6c98d15349a9157c23184b6dfb090ea79ea316c76e6876b9e85dc1200 WatchSource:0}: Error finding container a97374e6c98d15349a9157c23184b6dfb090ea79ea316c76e6876b9e85dc1200: Status 404 returned error can't find the container with id a97374e6c98d15349a9157c23184b6dfb090ea79ea316c76e6876b9e85dc1200 Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.392997 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41637a9b-2c22-4627-85df-b902f8d35ceb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"41637a9b-2c22-4627-85df-b902f8d35ceb\") " pod="openstack/nova-cell1-conductor-0" Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.393070 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6knm\" (UniqueName: \"kubernetes.io/projected/41637a9b-2c22-4627-85df-b902f8d35ceb-kube-api-access-p6knm\") pod \"nova-cell1-conductor-0\" (UID: \"41637a9b-2c22-4627-85df-b902f8d35ceb\") " pod="openstack/nova-cell1-conductor-0" Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.393096 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41637a9b-2c22-4627-85df-b902f8d35ceb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"41637a9b-2c22-4627-85df-b902f8d35ceb\") " pod="openstack/nova-cell1-conductor-0" Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.399306 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41637a9b-2c22-4627-85df-b902f8d35ceb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"41637a9b-2c22-4627-85df-b902f8d35ceb\") " pod="openstack/nova-cell1-conductor-0" Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.410535 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6knm\" (UniqueName: \"kubernetes.io/projected/41637a9b-2c22-4627-85df-b902f8d35ceb-kube-api-access-p6knm\") pod \"nova-cell1-conductor-0\" (UID: \"41637a9b-2c22-4627-85df-b902f8d35ceb\") " pod="openstack/nova-cell1-conductor-0" Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.417281 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41637a9b-2c22-4627-85df-b902f8d35ceb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"41637a9b-2c22-4627-85df-b902f8d35ceb\") " pod="openstack/nova-cell1-conductor-0" Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.491377 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2028e28b-656a-493b-9803-f63a21e625d2" path="/var/lib/kubelet/pods/2028e28b-656a-493b-9803-f63a21e625d2/volumes" Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.493735 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65" path="/var/lib/kubelet/pods/d2ad0b9f-caf6-4cc7-bd82-6145c4cbaf65/volumes" Mar 18 20:28:11 crc kubenswrapper[4950]: I0318 20:28:11.599767 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 18 20:28:12 crc kubenswrapper[4950]: I0318 20:28:12.121789 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 18 20:28:12 crc kubenswrapper[4950]: I0318 20:28:12.240554 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"41637a9b-2c22-4627-85df-b902f8d35ceb","Type":"ContainerStarted","Data":"ecd6db9218253bff810647f9842fb3caed5819c31c746e6a0f6394da8ff3c2a8"} Mar 18 20:28:12 crc kubenswrapper[4950]: I0318 20:28:12.244602 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cbbcbf5a-8667-4e11-ad3e-f72b30084b6d","Type":"ContainerStarted","Data":"65e45e86d329a18fae51cb507d8d87a43866d7ee499a73bc9a986762b2bde176"} Mar 18 20:28:12 crc kubenswrapper[4950]: I0318 20:28:12.244630 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cbbcbf5a-8667-4e11-ad3e-f72b30084b6d","Type":"ContainerStarted","Data":"a97374e6c98d15349a9157c23184b6dfb090ea79ea316c76e6876b9e85dc1200"} Mar 18 20:28:12 crc kubenswrapper[4950]: I0318 20:28:12.259279 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.259266434 podStartE2EDuration="2.259266434s" podCreationTimestamp="2026-03-18 20:28:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:28:12.257100398 +0000 UTC m=+1305.497942266" watchObservedRunningTime="2026-03-18 20:28:12.259266434 +0000 UTC m=+1305.500108302" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.246389 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.252879 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"41637a9b-2c22-4627-85df-b902f8d35ceb","Type":"ContainerStarted","Data":"1fd7631bb058020435d4b4234d2671be42c106ca67a734a7663e1131b2a266dd"} Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.253332 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.254768 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1ca2e06-431e-45b9-b492-20c491324945","Type":"ContainerStarted","Data":"f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f"} Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.255646 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.257321 4950 generic.go:334] "Generic (PLEG): container finished" podID="0217d09c-db1e-45b6-bc64-c3eca777e7dc" containerID="cf4c56175b75b3f1ced1fd7d0c734b9230cb2c9a634abf085dc8bdb522f80551" exitCode=0 Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.257372 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.257428 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0217d09c-db1e-45b6-bc64-c3eca777e7dc","Type":"ContainerDied","Data":"cf4c56175b75b3f1ced1fd7d0c734b9230cb2c9a634abf085dc8bdb522f80551"} Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.257446 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0217d09c-db1e-45b6-bc64-c3eca777e7dc","Type":"ContainerDied","Data":"98f18e55bb38a673453f674909d7ae1ec80c77ea73d96072a71b85c36d647ead"} Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.257463 4950 scope.go:117] "RemoveContainer" containerID="cf4c56175b75b3f1ced1fd7d0c734b9230cb2c9a634abf085dc8bdb522f80551" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.278441 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0217d09c-db1e-45b6-bc64-c3eca777e7dc-combined-ca-bundle\") pod \"0217d09c-db1e-45b6-bc64-c3eca777e7dc\" (UID: \"0217d09c-db1e-45b6-bc64-c3eca777e7dc\") " Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.278525 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0217d09c-db1e-45b6-bc64-c3eca777e7dc-logs\") pod \"0217d09c-db1e-45b6-bc64-c3eca777e7dc\" (UID: \"0217d09c-db1e-45b6-bc64-c3eca777e7dc\") " Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.278655 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgmhc\" (UniqueName: \"kubernetes.io/projected/0217d09c-db1e-45b6-bc64-c3eca777e7dc-kube-api-access-vgmhc\") pod \"0217d09c-db1e-45b6-bc64-c3eca777e7dc\" (UID: \"0217d09c-db1e-45b6-bc64-c3eca777e7dc\") " Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.278687 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0217d09c-db1e-45b6-bc64-c3eca777e7dc-config-data\") pod \"0217d09c-db1e-45b6-bc64-c3eca777e7dc\" (UID: \"0217d09c-db1e-45b6-bc64-c3eca777e7dc\") " Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.280307 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0217d09c-db1e-45b6-bc64-c3eca777e7dc-logs" (OuterVolumeSpecName: "logs") pod "0217d09c-db1e-45b6-bc64-c3eca777e7dc" (UID: "0217d09c-db1e-45b6-bc64-c3eca777e7dc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.285012 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0217d09c-db1e-45b6-bc64-c3eca777e7dc-kube-api-access-vgmhc" (OuterVolumeSpecName: "kube-api-access-vgmhc") pod "0217d09c-db1e-45b6-bc64-c3eca777e7dc" (UID: "0217d09c-db1e-45b6-bc64-c3eca777e7dc"). InnerVolumeSpecName "kube-api-access-vgmhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.311923 4950 scope.go:117] "RemoveContainer" containerID="e8912d9e397ece5941a51539bc31e219f238391233a8d61b16965058aa4cff40" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.315207 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0217d09c-db1e-45b6-bc64-c3eca777e7dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0217d09c-db1e-45b6-bc64-c3eca777e7dc" (UID: "0217d09c-db1e-45b6-bc64-c3eca777e7dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.319569 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0217d09c-db1e-45b6-bc64-c3eca777e7dc-config-data" (OuterVolumeSpecName: "config-data") pod "0217d09c-db1e-45b6-bc64-c3eca777e7dc" (UID: "0217d09c-db1e-45b6-bc64-c3eca777e7dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.326340 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.3263216079999998 podStartE2EDuration="2.326321608s" podCreationTimestamp="2026-03-18 20:28:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:28:13.323856175 +0000 UTC m=+1306.564698043" watchObservedRunningTime="2026-03-18 20:28:13.326321608 +0000 UTC m=+1306.567163476" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.350003 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.073018206 podStartE2EDuration="6.349986148s" podCreationTimestamp="2026-03-18 20:28:07 +0000 UTC" firstStartedPulling="2026-03-18 20:28:08.130314343 +0000 UTC m=+1301.371156211" lastFinishedPulling="2026-03-18 20:28:12.407282285 +0000 UTC m=+1305.648124153" observedRunningTime="2026-03-18 20:28:13.349355981 +0000 UTC m=+1306.590197849" watchObservedRunningTime="2026-03-18 20:28:13.349986148 +0000 UTC m=+1306.590828026" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.382901 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0217d09c-db1e-45b6-bc64-c3eca777e7dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.382929 4950 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0217d09c-db1e-45b6-bc64-c3eca777e7dc-logs\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.382938 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgmhc\" (UniqueName: \"kubernetes.io/projected/0217d09c-db1e-45b6-bc64-c3eca777e7dc-kube-api-access-vgmhc\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.387317 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0217d09c-db1e-45b6-bc64-c3eca777e7dc-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.387399 4950 scope.go:117] "RemoveContainer" containerID="cf4c56175b75b3f1ced1fd7d0c734b9230cb2c9a634abf085dc8bdb522f80551" Mar 18 20:28:13 crc kubenswrapper[4950]: E0318 20:28:13.388898 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf4c56175b75b3f1ced1fd7d0c734b9230cb2c9a634abf085dc8bdb522f80551\": container with ID starting with cf4c56175b75b3f1ced1fd7d0c734b9230cb2c9a634abf085dc8bdb522f80551 not found: ID does not exist" containerID="cf4c56175b75b3f1ced1fd7d0c734b9230cb2c9a634abf085dc8bdb522f80551" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.388939 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf4c56175b75b3f1ced1fd7d0c734b9230cb2c9a634abf085dc8bdb522f80551"} err="failed to get container status \"cf4c56175b75b3f1ced1fd7d0c734b9230cb2c9a634abf085dc8bdb522f80551\": rpc error: code = NotFound desc = could not find container \"cf4c56175b75b3f1ced1fd7d0c734b9230cb2c9a634abf085dc8bdb522f80551\": container with ID starting with cf4c56175b75b3f1ced1fd7d0c734b9230cb2c9a634abf085dc8bdb522f80551 not found: ID does not exist" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.388971 4950 scope.go:117] "RemoveContainer" containerID="e8912d9e397ece5941a51539bc31e219f238391233a8d61b16965058aa4cff40" Mar 18 20:28:13 crc kubenswrapper[4950]: E0318 20:28:13.392696 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8912d9e397ece5941a51539bc31e219f238391233a8d61b16965058aa4cff40\": container with ID starting with e8912d9e397ece5941a51539bc31e219f238391233a8d61b16965058aa4cff40 not found: ID does not exist" containerID="e8912d9e397ece5941a51539bc31e219f238391233a8d61b16965058aa4cff40" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.392843 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8912d9e397ece5941a51539bc31e219f238391233a8d61b16965058aa4cff40"} err="failed to get container status \"e8912d9e397ece5941a51539bc31e219f238391233a8d61b16965058aa4cff40\": rpc error: code = NotFound desc = could not find container \"e8912d9e397ece5941a51539bc31e219f238391233a8d61b16965058aa4cff40\": container with ID starting with e8912d9e397ece5941a51539bc31e219f238391233a8d61b16965058aa4cff40 not found: ID does not exist" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.588712 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.601674 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.615735 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 18 20:28:13 crc kubenswrapper[4950]: E0318 20:28:13.616083 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0217d09c-db1e-45b6-bc64-c3eca777e7dc" containerName="nova-api-api" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.616098 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="0217d09c-db1e-45b6-bc64-c3eca777e7dc" containerName="nova-api-api" Mar 18 20:28:13 crc kubenswrapper[4950]: E0318 20:28:13.616132 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0217d09c-db1e-45b6-bc64-c3eca777e7dc" containerName="nova-api-log" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.616138 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="0217d09c-db1e-45b6-bc64-c3eca777e7dc" containerName="nova-api-log" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.616296 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="0217d09c-db1e-45b6-bc64-c3eca777e7dc" containerName="nova-api-api" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.616314 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="0217d09c-db1e-45b6-bc64-c3eca777e7dc" containerName="nova-api-log" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.624823 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.624920 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.627157 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.695613 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e016c0d-1c05-48ee-8f29-17d94a91edbc-logs\") pod \"nova-api-0\" (UID: \"0e016c0d-1c05-48ee-8f29-17d94a91edbc\") " pod="openstack/nova-api-0" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.695949 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4ksr\" (UniqueName: \"kubernetes.io/projected/0e016c0d-1c05-48ee-8f29-17d94a91edbc-kube-api-access-t4ksr\") pod \"nova-api-0\" (UID: \"0e016c0d-1c05-48ee-8f29-17d94a91edbc\") " pod="openstack/nova-api-0" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.696090 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e016c0d-1c05-48ee-8f29-17d94a91edbc-config-data\") pod \"nova-api-0\" (UID: \"0e016c0d-1c05-48ee-8f29-17d94a91edbc\") " pod="openstack/nova-api-0" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.696276 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e016c0d-1c05-48ee-8f29-17d94a91edbc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0e016c0d-1c05-48ee-8f29-17d94a91edbc\") " pod="openstack/nova-api-0" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.798588 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e016c0d-1c05-48ee-8f29-17d94a91edbc-logs\") pod \"nova-api-0\" (UID: \"0e016c0d-1c05-48ee-8f29-17d94a91edbc\") " pod="openstack/nova-api-0" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.798826 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4ksr\" (UniqueName: \"kubernetes.io/projected/0e016c0d-1c05-48ee-8f29-17d94a91edbc-kube-api-access-t4ksr\") pod \"nova-api-0\" (UID: \"0e016c0d-1c05-48ee-8f29-17d94a91edbc\") " pod="openstack/nova-api-0" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.798926 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e016c0d-1c05-48ee-8f29-17d94a91edbc-config-data\") pod \"nova-api-0\" (UID: \"0e016c0d-1c05-48ee-8f29-17d94a91edbc\") " pod="openstack/nova-api-0" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.799044 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e016c0d-1c05-48ee-8f29-17d94a91edbc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0e016c0d-1c05-48ee-8f29-17d94a91edbc\") " pod="openstack/nova-api-0" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.799326 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e016c0d-1c05-48ee-8f29-17d94a91edbc-logs\") pod \"nova-api-0\" (UID: \"0e016c0d-1c05-48ee-8f29-17d94a91edbc\") " pod="openstack/nova-api-0" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.802287 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e016c0d-1c05-48ee-8f29-17d94a91edbc-config-data\") pod \"nova-api-0\" (UID: \"0e016c0d-1c05-48ee-8f29-17d94a91edbc\") " pod="openstack/nova-api-0" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.802833 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e016c0d-1c05-48ee-8f29-17d94a91edbc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0e016c0d-1c05-48ee-8f29-17d94a91edbc\") " pod="openstack/nova-api-0" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.813263 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4ksr\" (UniqueName: \"kubernetes.io/projected/0e016c0d-1c05-48ee-8f29-17d94a91edbc-kube-api-access-t4ksr\") pod \"nova-api-0\" (UID: \"0e016c0d-1c05-48ee-8f29-17d94a91edbc\") " pod="openstack/nova-api-0" Mar 18 20:28:13 crc kubenswrapper[4950]: I0318 20:28:13.953635 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 18 20:28:14 crc kubenswrapper[4950]: I0318 20:28:14.423969 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 18 20:28:14 crc kubenswrapper[4950]: W0318 20:28:14.424659 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e016c0d_1c05_48ee_8f29_17d94a91edbc.slice/crio-92e4478cbfd3380ccb1f2565a74f5876934412cd3e70fb52104a61106ed33cb4 WatchSource:0}: Error finding container 92e4478cbfd3380ccb1f2565a74f5876934412cd3e70fb52104a61106ed33cb4: Status 404 returned error can't find the container with id 92e4478cbfd3380ccb1f2565a74f5876934412cd3e70fb52104a61106ed33cb4 Mar 18 20:28:15 crc kubenswrapper[4950]: I0318 20:28:15.278556 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0e016c0d-1c05-48ee-8f29-17d94a91edbc","Type":"ContainerStarted","Data":"0602d592b25338753eb5b4f815d3a5dd411a9ea2df2b92bca1139b1c3757e650"} Mar 18 20:28:15 crc kubenswrapper[4950]: I0318 20:28:15.279267 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0e016c0d-1c05-48ee-8f29-17d94a91edbc","Type":"ContainerStarted","Data":"cc0d53162ba3f45a5b04553df12d6904f100b9b9a962ca5c925248bda6ca150e"} Mar 18 20:28:15 crc kubenswrapper[4950]: I0318 20:28:15.279369 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0e016c0d-1c05-48ee-8f29-17d94a91edbc","Type":"ContainerStarted","Data":"92e4478cbfd3380ccb1f2565a74f5876934412cd3e70fb52104a61106ed33cb4"} Mar 18 20:28:15 crc kubenswrapper[4950]: I0318 20:28:15.299405 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.29938941 podStartE2EDuration="2.29938941s" podCreationTimestamp="2026-03-18 20:28:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:28:15.297879921 +0000 UTC m=+1308.538721789" watchObservedRunningTime="2026-03-18 20:28:15.29938941 +0000 UTC m=+1308.540231278" Mar 18 20:28:15 crc kubenswrapper[4950]: I0318 20:28:15.492933 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0217d09c-db1e-45b6-bc64-c3eca777e7dc" path="/var/lib/kubelet/pods/0217d09c-db1e-45b6-bc64-c3eca777e7dc/volumes" Mar 18 20:28:15 crc kubenswrapper[4950]: I0318 20:28:15.908939 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 18 20:28:19 crc kubenswrapper[4950]: I0318 20:28:19.840108 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 18 20:28:19 crc kubenswrapper[4950]: I0318 20:28:19.840958 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 18 20:28:20 crc kubenswrapper[4950]: I0318 20:28:20.855623 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ade0ed1d-2241-456b-b72f-3236fbc6ed45" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.182:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 18 20:28:20 crc kubenswrapper[4950]: I0318 20:28:20.855649 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ade0ed1d-2241-456b-b72f-3236fbc6ed45" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.182:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 18 20:28:20 crc kubenswrapper[4950]: I0318 20:28:20.909312 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 18 20:28:20 crc kubenswrapper[4950]: I0318 20:28:20.948734 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 18 20:28:21 crc kubenswrapper[4950]: I0318 20:28:21.377947 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 18 20:28:21 crc kubenswrapper[4950]: I0318 20:28:21.654470 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Mar 18 20:28:23 crc kubenswrapper[4950]: I0318 20:28:23.954656 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 18 20:28:23 crc kubenswrapper[4950]: I0318 20:28:23.955016 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 18 20:28:25 crc kubenswrapper[4950]: I0318 20:28:25.036613 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0e016c0d-1c05-48ee-8f29-17d94a91edbc" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 18 20:28:25 crc kubenswrapper[4950]: I0318 20:28:25.037105 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0e016c0d-1c05-48ee-8f29-17d94a91edbc" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 18 20:28:27 crc kubenswrapper[4950]: I0318 20:28:27.839759 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 18 20:28:27 crc kubenswrapper[4950]: I0318 20:28:27.840187 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 18 20:28:29 crc kubenswrapper[4950]: I0318 20:28:29.849858 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 18 20:28:29 crc kubenswrapper[4950]: I0318 20:28:29.850581 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 18 20:28:29 crc kubenswrapper[4950]: I0318 20:28:29.859576 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 18 20:28:30 crc kubenswrapper[4950]: I0318 20:28:30.442571 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 18 20:28:31 crc kubenswrapper[4950]: I0318 20:28:31.954533 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 18 20:28:31 crc kubenswrapper[4950]: I0318 20:28:31.954920 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 18 20:28:33 crc kubenswrapper[4950]: I0318 20:28:33.462880 4950 generic.go:334] "Generic (PLEG): container finished" podID="861b3002-ed64-4b45-b1fa-dc6eabc6a611" containerID="e3dcff1971656e6b2a038d556bf18d138c9cfff265d2ca9e52f845ed6aa55b08" exitCode=137 Mar 18 20:28:33 crc kubenswrapper[4950]: I0318 20:28:33.463082 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"861b3002-ed64-4b45-b1fa-dc6eabc6a611","Type":"ContainerDied","Data":"e3dcff1971656e6b2a038d556bf18d138c9cfff265d2ca9e52f845ed6aa55b08"} Mar 18 20:28:33 crc kubenswrapper[4950]: I0318 20:28:33.463243 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"861b3002-ed64-4b45-b1fa-dc6eabc6a611","Type":"ContainerDied","Data":"af556040a5d747b0db1decfb629b321a2b82da507ca31f08943b0e8af2929376"} Mar 18 20:28:33 crc kubenswrapper[4950]: I0318 20:28:33.463262 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af556040a5d747b0db1decfb629b321a2b82da507ca31f08943b0e8af2929376" Mar 18 20:28:33 crc kubenswrapper[4950]: I0318 20:28:33.491767 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:33 crc kubenswrapper[4950]: I0318 20:28:33.623615 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/861b3002-ed64-4b45-b1fa-dc6eabc6a611-combined-ca-bundle\") pod \"861b3002-ed64-4b45-b1fa-dc6eabc6a611\" (UID: \"861b3002-ed64-4b45-b1fa-dc6eabc6a611\") " Mar 18 20:28:33 crc kubenswrapper[4950]: I0318 20:28:33.623813 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/861b3002-ed64-4b45-b1fa-dc6eabc6a611-config-data\") pod \"861b3002-ed64-4b45-b1fa-dc6eabc6a611\" (UID: \"861b3002-ed64-4b45-b1fa-dc6eabc6a611\") " Mar 18 20:28:33 crc kubenswrapper[4950]: I0318 20:28:33.623840 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpl8j\" (UniqueName: \"kubernetes.io/projected/861b3002-ed64-4b45-b1fa-dc6eabc6a611-kube-api-access-rpl8j\") pod \"861b3002-ed64-4b45-b1fa-dc6eabc6a611\" (UID: \"861b3002-ed64-4b45-b1fa-dc6eabc6a611\") " Mar 18 20:28:33 crc kubenswrapper[4950]: I0318 20:28:33.630503 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/861b3002-ed64-4b45-b1fa-dc6eabc6a611-kube-api-access-rpl8j" (OuterVolumeSpecName: "kube-api-access-rpl8j") pod "861b3002-ed64-4b45-b1fa-dc6eabc6a611" (UID: "861b3002-ed64-4b45-b1fa-dc6eabc6a611"). InnerVolumeSpecName "kube-api-access-rpl8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:28:33 crc kubenswrapper[4950]: I0318 20:28:33.649480 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/861b3002-ed64-4b45-b1fa-dc6eabc6a611-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "861b3002-ed64-4b45-b1fa-dc6eabc6a611" (UID: "861b3002-ed64-4b45-b1fa-dc6eabc6a611"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:33 crc kubenswrapper[4950]: I0318 20:28:33.652029 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/861b3002-ed64-4b45-b1fa-dc6eabc6a611-config-data" (OuterVolumeSpecName: "config-data") pod "861b3002-ed64-4b45-b1fa-dc6eabc6a611" (UID: "861b3002-ed64-4b45-b1fa-dc6eabc6a611"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:33 crc kubenswrapper[4950]: I0318 20:28:33.725482 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/861b3002-ed64-4b45-b1fa-dc6eabc6a611-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:33 crc kubenswrapper[4950]: I0318 20:28:33.725536 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/861b3002-ed64-4b45-b1fa-dc6eabc6a611-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:33 crc kubenswrapper[4950]: I0318 20:28:33.725547 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpl8j\" (UniqueName: \"kubernetes.io/projected/861b3002-ed64-4b45-b1fa-dc6eabc6a611-kube-api-access-rpl8j\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:33 crc kubenswrapper[4950]: I0318 20:28:33.835716 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:28:33 crc kubenswrapper[4950]: I0318 20:28:33.835772 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:28:33 crc kubenswrapper[4950]: I0318 20:28:33.960185 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 18 20:28:33 crc kubenswrapper[4950]: I0318 20:28:33.961528 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 18 20:28:33 crc kubenswrapper[4950]: I0318 20:28:33.966764 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.474906 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.480703 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.540611 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.557738 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.568048 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 18 20:28:34 crc kubenswrapper[4950]: E0318 20:28:34.568540 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="861b3002-ed64-4b45-b1fa-dc6eabc6a611" containerName="nova-cell1-novncproxy-novncproxy" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.568563 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="861b3002-ed64-4b45-b1fa-dc6eabc6a611" containerName="nova-cell1-novncproxy-novncproxy" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.568816 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="861b3002-ed64-4b45-b1fa-dc6eabc6a611" containerName="nova-cell1-novncproxy-novncproxy" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.569552 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.602285 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.602643 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.611294 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.631144 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.710712 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-776c49c95f-d9656"] Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.714010 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-776c49c95f-d9656" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.733175 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-776c49c95f-d9656"] Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.758002 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49765550-f019-4bdd-8a9c-ffadf57e59c4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"49765550-f019-4bdd-8a9c-ffadf57e59c4\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.758115 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49765550-f019-4bdd-8a9c-ffadf57e59c4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"49765550-f019-4bdd-8a9c-ffadf57e59c4\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.758142 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bq56\" (UniqueName: \"kubernetes.io/projected/49765550-f019-4bdd-8a9c-ffadf57e59c4-kube-api-access-5bq56\") pod \"nova-cell1-novncproxy-0\" (UID: \"49765550-f019-4bdd-8a9c-ffadf57e59c4\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.758161 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/49765550-f019-4bdd-8a9c-ffadf57e59c4-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"49765550-f019-4bdd-8a9c-ffadf57e59c4\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.758188 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/49765550-f019-4bdd-8a9c-ffadf57e59c4-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"49765550-f019-4bdd-8a9c-ffadf57e59c4\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.859317 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bq56\" (UniqueName: \"kubernetes.io/projected/49765550-f019-4bdd-8a9c-ffadf57e59c4-kube-api-access-5bq56\") pod \"nova-cell1-novncproxy-0\" (UID: \"49765550-f019-4bdd-8a9c-ffadf57e59c4\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.859384 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-dns-svc\") pod \"dnsmasq-dns-776c49c95f-d9656\" (UID: \"e7feba91-0f11-4825-8b4c-a24bc2807566\") " pod="openstack/dnsmasq-dns-776c49c95f-d9656" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.859433 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/49765550-f019-4bdd-8a9c-ffadf57e59c4-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"49765550-f019-4bdd-8a9c-ffadf57e59c4\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.859485 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/49765550-f019-4bdd-8a9c-ffadf57e59c4-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"49765550-f019-4bdd-8a9c-ffadf57e59c4\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.859511 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sppj5\" (UniqueName: \"kubernetes.io/projected/e7feba91-0f11-4825-8b4c-a24bc2807566-kube-api-access-sppj5\") pod \"dnsmasq-dns-776c49c95f-d9656\" (UID: \"e7feba91-0f11-4825-8b4c-a24bc2807566\") " pod="openstack/dnsmasq-dns-776c49c95f-d9656" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.859567 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-ovsdbserver-sb\") pod \"dnsmasq-dns-776c49c95f-d9656\" (UID: \"e7feba91-0f11-4825-8b4c-a24bc2807566\") " pod="openstack/dnsmasq-dns-776c49c95f-d9656" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.859640 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49765550-f019-4bdd-8a9c-ffadf57e59c4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"49765550-f019-4bdd-8a9c-ffadf57e59c4\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.859680 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-ovsdbserver-nb\") pod \"dnsmasq-dns-776c49c95f-d9656\" (UID: \"e7feba91-0f11-4825-8b4c-a24bc2807566\") " pod="openstack/dnsmasq-dns-776c49c95f-d9656" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.859735 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-config\") pod \"dnsmasq-dns-776c49c95f-d9656\" (UID: \"e7feba91-0f11-4825-8b4c-a24bc2807566\") " pod="openstack/dnsmasq-dns-776c49c95f-d9656" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.859796 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49765550-f019-4bdd-8a9c-ffadf57e59c4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"49765550-f019-4bdd-8a9c-ffadf57e59c4\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.866576 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/49765550-f019-4bdd-8a9c-ffadf57e59c4-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"49765550-f019-4bdd-8a9c-ffadf57e59c4\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.870126 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/49765550-f019-4bdd-8a9c-ffadf57e59c4-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"49765550-f019-4bdd-8a9c-ffadf57e59c4\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.875988 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bq56\" (UniqueName: \"kubernetes.io/projected/49765550-f019-4bdd-8a9c-ffadf57e59c4-kube-api-access-5bq56\") pod \"nova-cell1-novncproxy-0\" (UID: \"49765550-f019-4bdd-8a9c-ffadf57e59c4\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.879862 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49765550-f019-4bdd-8a9c-ffadf57e59c4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"49765550-f019-4bdd-8a9c-ffadf57e59c4\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.882282 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49765550-f019-4bdd-8a9c-ffadf57e59c4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"49765550-f019-4bdd-8a9c-ffadf57e59c4\") " pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.908019 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.965582 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-dns-svc\") pod \"dnsmasq-dns-776c49c95f-d9656\" (UID: \"e7feba91-0f11-4825-8b4c-a24bc2807566\") " pod="openstack/dnsmasq-dns-776c49c95f-d9656" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.965652 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sppj5\" (UniqueName: \"kubernetes.io/projected/e7feba91-0f11-4825-8b4c-a24bc2807566-kube-api-access-sppj5\") pod \"dnsmasq-dns-776c49c95f-d9656\" (UID: \"e7feba91-0f11-4825-8b4c-a24bc2807566\") " pod="openstack/dnsmasq-dns-776c49c95f-d9656" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.965695 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-ovsdbserver-sb\") pod \"dnsmasq-dns-776c49c95f-d9656\" (UID: \"e7feba91-0f11-4825-8b4c-a24bc2807566\") " pod="openstack/dnsmasq-dns-776c49c95f-d9656" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.965757 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-ovsdbserver-nb\") pod \"dnsmasq-dns-776c49c95f-d9656\" (UID: \"e7feba91-0f11-4825-8b4c-a24bc2807566\") " pod="openstack/dnsmasq-dns-776c49c95f-d9656" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.965795 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-config\") pod \"dnsmasq-dns-776c49c95f-d9656\" (UID: \"e7feba91-0f11-4825-8b4c-a24bc2807566\") " pod="openstack/dnsmasq-dns-776c49c95f-d9656" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.966896 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-config\") pod \"dnsmasq-dns-776c49c95f-d9656\" (UID: \"e7feba91-0f11-4825-8b4c-a24bc2807566\") " pod="openstack/dnsmasq-dns-776c49c95f-d9656" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.967487 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-dns-svc\") pod \"dnsmasq-dns-776c49c95f-d9656\" (UID: \"e7feba91-0f11-4825-8b4c-a24bc2807566\") " pod="openstack/dnsmasq-dns-776c49c95f-d9656" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.968148 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-ovsdbserver-nb\") pod \"dnsmasq-dns-776c49c95f-d9656\" (UID: \"e7feba91-0f11-4825-8b4c-a24bc2807566\") " pod="openstack/dnsmasq-dns-776c49c95f-d9656" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.968213 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-ovsdbserver-sb\") pod \"dnsmasq-dns-776c49c95f-d9656\" (UID: \"e7feba91-0f11-4825-8b4c-a24bc2807566\") " pod="openstack/dnsmasq-dns-776c49c95f-d9656" Mar 18 20:28:34 crc kubenswrapper[4950]: I0318 20:28:34.987407 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sppj5\" (UniqueName: \"kubernetes.io/projected/e7feba91-0f11-4825-8b4c-a24bc2807566-kube-api-access-sppj5\") pod \"dnsmasq-dns-776c49c95f-d9656\" (UID: \"e7feba91-0f11-4825-8b4c-a24bc2807566\") " pod="openstack/dnsmasq-dns-776c49c95f-d9656" Mar 18 20:28:35 crc kubenswrapper[4950]: I0318 20:28:35.046013 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-776c49c95f-d9656" Mar 18 20:28:35 crc kubenswrapper[4950]: I0318 20:28:35.495860 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="861b3002-ed64-4b45-b1fa-dc6eabc6a611" path="/var/lib/kubelet/pods/861b3002-ed64-4b45-b1fa-dc6eabc6a611/volumes" Mar 18 20:28:35 crc kubenswrapper[4950]: I0318 20:28:35.549529 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-776c49c95f-d9656"] Mar 18 20:28:35 crc kubenswrapper[4950]: I0318 20:28:35.586500 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 18 20:28:36 crc kubenswrapper[4950]: I0318 20:28:36.496107 4950 generic.go:334] "Generic (PLEG): container finished" podID="e7feba91-0f11-4825-8b4c-a24bc2807566" containerID="6d08377bb6463505044677f3066c51b22fb5f47c8d7c25b14a1faa87637e6f89" exitCode=0 Mar 18 20:28:36 crc kubenswrapper[4950]: I0318 20:28:36.497051 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-776c49c95f-d9656" event={"ID":"e7feba91-0f11-4825-8b4c-a24bc2807566","Type":"ContainerDied","Data":"6d08377bb6463505044677f3066c51b22fb5f47c8d7c25b14a1faa87637e6f89"} Mar 18 20:28:36 crc kubenswrapper[4950]: I0318 20:28:36.497111 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-776c49c95f-d9656" event={"ID":"e7feba91-0f11-4825-8b4c-a24bc2807566","Type":"ContainerStarted","Data":"9ff88862cbebfcf5f39958385155b309205c42a27c9067d471683a08c158315e"} Mar 18 20:28:36 crc kubenswrapper[4950]: I0318 20:28:36.503584 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"49765550-f019-4bdd-8a9c-ffadf57e59c4","Type":"ContainerStarted","Data":"3bfe9dd48411bbb68e01b9a45b23099d5a14bd3c32f2c5affee2933d24a3b57b"} Mar 18 20:28:36 crc kubenswrapper[4950]: I0318 20:28:36.503618 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"49765550-f019-4bdd-8a9c-ffadf57e59c4","Type":"ContainerStarted","Data":"3365d786f6440bd5e2649f1c0a953232e9918622245150d0b84582db7923ec92"} Mar 18 20:28:36 crc kubenswrapper[4950]: I0318 20:28:36.567946 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.567929786 podStartE2EDuration="2.567929786s" podCreationTimestamp="2026-03-18 20:28:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:28:36.542924902 +0000 UTC m=+1329.783766780" watchObservedRunningTime="2026-03-18 20:28:36.567929786 +0000 UTC m=+1329.808771654" Mar 18 20:28:37 crc kubenswrapper[4950]: I0318 20:28:37.181079 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 18 20:28:37 crc kubenswrapper[4950]: I0318 20:28:37.516886 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:28:37 crc kubenswrapper[4950]: I0318 20:28:37.518150 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e1ca2e06-431e-45b9-b492-20c491324945" containerName="ceilometer-central-agent" containerID="cri-o://9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168" gracePeriod=30 Mar 18 20:28:37 crc kubenswrapper[4950]: I0318 20:28:37.518470 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e1ca2e06-431e-45b9-b492-20c491324945" containerName="sg-core" containerID="cri-o://1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2" gracePeriod=30 Mar 18 20:28:37 crc kubenswrapper[4950]: I0318 20:28:37.518568 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e1ca2e06-431e-45b9-b492-20c491324945" containerName="proxy-httpd" containerID="cri-o://f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f" gracePeriod=30 Mar 18 20:28:37 crc kubenswrapper[4950]: I0318 20:28:37.518605 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e1ca2e06-431e-45b9-b492-20c491324945" containerName="ceilometer-notification-agent" containerID="cri-o://8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b" gracePeriod=30 Mar 18 20:28:37 crc kubenswrapper[4950]: I0318 20:28:37.529147 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-776c49c95f-d9656" event={"ID":"e7feba91-0f11-4825-8b4c-a24bc2807566","Type":"ContainerStarted","Data":"a2e15ed2c693569d6521c74966e17c04ee48d9d64e059b59d8a9d355ac2cbd14"} Mar 18 20:28:37 crc kubenswrapper[4950]: I0318 20:28:37.529821 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0e016c0d-1c05-48ee-8f29-17d94a91edbc" containerName="nova-api-log" containerID="cri-o://cc0d53162ba3f45a5b04553df12d6904f100b9b9a962ca5c925248bda6ca150e" gracePeriod=30 Mar 18 20:28:37 crc kubenswrapper[4950]: I0318 20:28:37.529981 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0e016c0d-1c05-48ee-8f29-17d94a91edbc" containerName="nova-api-api" containerID="cri-o://0602d592b25338753eb5b4f815d3a5dd411a9ea2df2b92bca1139b1c3757e650" gracePeriod=30 Mar 18 20:28:37 crc kubenswrapper[4950]: I0318 20:28:37.592968 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-776c49c95f-d9656" podStartSLOduration=3.592946198 podStartE2EDuration="3.592946198s" podCreationTimestamp="2026-03-18 20:28:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:28:37.567739299 +0000 UTC m=+1330.808581167" watchObservedRunningTime="2026-03-18 20:28:37.592946198 +0000 UTC m=+1330.833788066" Mar 18 20:28:37 crc kubenswrapper[4950]: I0318 20:28:37.645924 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="e1ca2e06-431e-45b9-b492-20c491324945" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.181:3000/\": read tcp 10.217.0.2:35054->10.217.0.181:3000: read: connection reset by peer" Mar 18 20:28:37 crc kubenswrapper[4950]: I0318 20:28:37.646385 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="e1ca2e06-431e-45b9-b492-20c491324945" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.181:3000/\": dial tcp 10.217.0.181:3000: connect: connection refused" Mar 18 20:28:37 crc kubenswrapper[4950]: E0318 20:28:37.694350 4950 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e016c0d_1c05_48ee_8f29_17d94a91edbc.slice/crio-conmon-cc0d53162ba3f45a5b04553df12d6904f100b9b9a962ca5c925248bda6ca150e.scope\": RecentStats: unable to find data in memory cache]" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.462479 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.533047 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-sg-core-conf-yaml\") pod \"e1ca2e06-431e-45b9-b492-20c491324945\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.533106 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-scripts\") pod \"e1ca2e06-431e-45b9-b492-20c491324945\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.533162 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-config-data\") pod \"e1ca2e06-431e-45b9-b492-20c491324945\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.533331 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1ca2e06-431e-45b9-b492-20c491324945-log-httpd\") pod \"e1ca2e06-431e-45b9-b492-20c491324945\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.533512 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1ca2e06-431e-45b9-b492-20c491324945-run-httpd\") pod \"e1ca2e06-431e-45b9-b492-20c491324945\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.533590 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rw66\" (UniqueName: \"kubernetes.io/projected/e1ca2e06-431e-45b9-b492-20c491324945-kube-api-access-4rw66\") pod \"e1ca2e06-431e-45b9-b492-20c491324945\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.533624 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-combined-ca-bundle\") pod \"e1ca2e06-431e-45b9-b492-20c491324945\" (UID: \"e1ca2e06-431e-45b9-b492-20c491324945\") " Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.533920 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1ca2e06-431e-45b9-b492-20c491324945-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e1ca2e06-431e-45b9-b492-20c491324945" (UID: "e1ca2e06-431e-45b9-b492-20c491324945"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.534212 4950 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1ca2e06-431e-45b9-b492-20c491324945-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.535203 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1ca2e06-431e-45b9-b492-20c491324945-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e1ca2e06-431e-45b9-b492-20c491324945" (UID: "e1ca2e06-431e-45b9-b492-20c491324945"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.539389 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1ca2e06-431e-45b9-b492-20c491324945-kube-api-access-4rw66" (OuterVolumeSpecName: "kube-api-access-4rw66") pod "e1ca2e06-431e-45b9-b492-20c491324945" (UID: "e1ca2e06-431e-45b9-b492-20c491324945"). InnerVolumeSpecName "kube-api-access-4rw66". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.546328 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-scripts" (OuterVolumeSpecName: "scripts") pod "e1ca2e06-431e-45b9-b492-20c491324945" (UID: "e1ca2e06-431e-45b9-b492-20c491324945"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.547239 4950 generic.go:334] "Generic (PLEG): container finished" podID="0e016c0d-1c05-48ee-8f29-17d94a91edbc" containerID="cc0d53162ba3f45a5b04553df12d6904f100b9b9a962ca5c925248bda6ca150e" exitCode=143 Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.547304 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0e016c0d-1c05-48ee-8f29-17d94a91edbc","Type":"ContainerDied","Data":"cc0d53162ba3f45a5b04553df12d6904f100b9b9a962ca5c925248bda6ca150e"} Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.551019 4950 generic.go:334] "Generic (PLEG): container finished" podID="e1ca2e06-431e-45b9-b492-20c491324945" containerID="f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f" exitCode=0 Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.551045 4950 generic.go:334] "Generic (PLEG): container finished" podID="e1ca2e06-431e-45b9-b492-20c491324945" containerID="1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2" exitCode=2 Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.551052 4950 generic.go:334] "Generic (PLEG): container finished" podID="e1ca2e06-431e-45b9-b492-20c491324945" containerID="8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b" exitCode=0 Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.551059 4950 generic.go:334] "Generic (PLEG): container finished" podID="e1ca2e06-431e-45b9-b492-20c491324945" containerID="9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168" exitCode=0 Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.551993 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.552470 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1ca2e06-431e-45b9-b492-20c491324945","Type":"ContainerDied","Data":"f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f"} Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.552503 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-776c49c95f-d9656" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.552514 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1ca2e06-431e-45b9-b492-20c491324945","Type":"ContainerDied","Data":"1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2"} Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.552523 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1ca2e06-431e-45b9-b492-20c491324945","Type":"ContainerDied","Data":"8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b"} Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.552534 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1ca2e06-431e-45b9-b492-20c491324945","Type":"ContainerDied","Data":"9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168"} Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.552542 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1ca2e06-431e-45b9-b492-20c491324945","Type":"ContainerDied","Data":"e0e57a505f8fc4a0fcd2a90c2f236e3d7e404ba5e0e4cca9313bea52628ce377"} Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.552560 4950 scope.go:117] "RemoveContainer" containerID="f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.575673 4950 scope.go:117] "RemoveContainer" containerID="1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.586711 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e1ca2e06-431e-45b9-b492-20c491324945" (UID: "e1ca2e06-431e-45b9-b492-20c491324945"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.599287 4950 scope.go:117] "RemoveContainer" containerID="8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.613799 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e1ca2e06-431e-45b9-b492-20c491324945" (UID: "e1ca2e06-431e-45b9-b492-20c491324945"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.625943 4950 scope.go:117] "RemoveContainer" containerID="9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.635827 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.635879 4950 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.635896 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.635906 4950 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1ca2e06-431e-45b9-b492-20c491324945-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.635916 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rw66\" (UniqueName: \"kubernetes.io/projected/e1ca2e06-431e-45b9-b492-20c491324945-kube-api-access-4rw66\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.668691 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-config-data" (OuterVolumeSpecName: "config-data") pod "e1ca2e06-431e-45b9-b492-20c491324945" (UID: "e1ca2e06-431e-45b9-b492-20c491324945"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.737318 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1ca2e06-431e-45b9-b492-20c491324945-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.747011 4950 scope.go:117] "RemoveContainer" containerID="f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f" Mar 18 20:28:38 crc kubenswrapper[4950]: E0318 20:28:38.750956 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f\": container with ID starting with f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f not found: ID does not exist" containerID="f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.751015 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f"} err="failed to get container status \"f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f\": rpc error: code = NotFound desc = could not find container \"f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f\": container with ID starting with f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f not found: ID does not exist" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.751039 4950 scope.go:117] "RemoveContainer" containerID="1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2" Mar 18 20:28:38 crc kubenswrapper[4950]: E0318 20:28:38.751458 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2\": container with ID starting with 1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2 not found: ID does not exist" containerID="1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.751482 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2"} err="failed to get container status \"1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2\": rpc error: code = NotFound desc = could not find container \"1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2\": container with ID starting with 1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2 not found: ID does not exist" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.751513 4950 scope.go:117] "RemoveContainer" containerID="8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b" Mar 18 20:28:38 crc kubenswrapper[4950]: E0318 20:28:38.752166 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b\": container with ID starting with 8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b not found: ID does not exist" containerID="8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.752266 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b"} err="failed to get container status \"8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b\": rpc error: code = NotFound desc = could not find container \"8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b\": container with ID starting with 8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b not found: ID does not exist" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.752341 4950 scope.go:117] "RemoveContainer" containerID="9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168" Mar 18 20:28:38 crc kubenswrapper[4950]: E0318 20:28:38.752713 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168\": container with ID starting with 9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168 not found: ID does not exist" containerID="9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.752807 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168"} err="failed to get container status \"9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168\": rpc error: code = NotFound desc = could not find container \"9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168\": container with ID starting with 9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168 not found: ID does not exist" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.752875 4950 scope.go:117] "RemoveContainer" containerID="f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.753209 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f"} err="failed to get container status \"f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f\": rpc error: code = NotFound desc = could not find container \"f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f\": container with ID starting with f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f not found: ID does not exist" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.753292 4950 scope.go:117] "RemoveContainer" containerID="1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.753640 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2"} err="failed to get container status \"1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2\": rpc error: code = NotFound desc = could not find container \"1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2\": container with ID starting with 1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2 not found: ID does not exist" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.753724 4950 scope.go:117] "RemoveContainer" containerID="8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.754059 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b"} err="failed to get container status \"8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b\": rpc error: code = NotFound desc = could not find container \"8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b\": container with ID starting with 8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b not found: ID does not exist" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.754143 4950 scope.go:117] "RemoveContainer" containerID="9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.754481 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168"} err="failed to get container status \"9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168\": rpc error: code = NotFound desc = could not find container \"9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168\": container with ID starting with 9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168 not found: ID does not exist" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.754565 4950 scope.go:117] "RemoveContainer" containerID="f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.754898 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f"} err="failed to get container status \"f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f\": rpc error: code = NotFound desc = could not find container \"f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f\": container with ID starting with f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f not found: ID does not exist" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.754922 4950 scope.go:117] "RemoveContainer" containerID="1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.755310 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2"} err="failed to get container status \"1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2\": rpc error: code = NotFound desc = could not find container \"1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2\": container with ID starting with 1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2 not found: ID does not exist" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.755329 4950 scope.go:117] "RemoveContainer" containerID="8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.756571 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b"} err="failed to get container status \"8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b\": rpc error: code = NotFound desc = could not find container \"8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b\": container with ID starting with 8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b not found: ID does not exist" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.756596 4950 scope.go:117] "RemoveContainer" containerID="9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.756971 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168"} err="failed to get container status \"9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168\": rpc error: code = NotFound desc = could not find container \"9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168\": container with ID starting with 9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168 not found: ID does not exist" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.756989 4950 scope.go:117] "RemoveContainer" containerID="f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.757298 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f"} err="failed to get container status \"f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f\": rpc error: code = NotFound desc = could not find container \"f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f\": container with ID starting with f81263b30f6559de41fb4b60ac3a7cd0c6e7772517a2570d19392e12a42b871f not found: ID does not exist" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.757342 4950 scope.go:117] "RemoveContainer" containerID="1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.757754 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2"} err="failed to get container status \"1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2\": rpc error: code = NotFound desc = could not find container \"1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2\": container with ID starting with 1975db1668d30e27071411d8a5b28101634caeda035d4bdc3e0d64f3dd65c5d2 not found: ID does not exist" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.757774 4950 scope.go:117] "RemoveContainer" containerID="8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.758199 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b"} err="failed to get container status \"8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b\": rpc error: code = NotFound desc = could not find container \"8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b\": container with ID starting with 8eaf292acfefafeab7ba0242261fa31c7f14e8f7b717e591dbb5326e379bea0b not found: ID does not exist" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.758226 4950 scope.go:117] "RemoveContainer" containerID="9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.758603 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168"} err="failed to get container status \"9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168\": rpc error: code = NotFound desc = could not find container \"9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168\": container with ID starting with 9c27da34aa4011aafd6d5ecc9662543e52697856dc0f77f4f024a6197cc53168 not found: ID does not exist" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.888171 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.895525 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.906028 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:28:38 crc kubenswrapper[4950]: E0318 20:28:38.906525 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1ca2e06-431e-45b9-b492-20c491324945" containerName="proxy-httpd" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.906588 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1ca2e06-431e-45b9-b492-20c491324945" containerName="proxy-httpd" Mar 18 20:28:38 crc kubenswrapper[4950]: E0318 20:28:38.906681 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1ca2e06-431e-45b9-b492-20c491324945" containerName="ceilometer-central-agent" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.906731 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1ca2e06-431e-45b9-b492-20c491324945" containerName="ceilometer-central-agent" Mar 18 20:28:38 crc kubenswrapper[4950]: E0318 20:28:38.906789 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1ca2e06-431e-45b9-b492-20c491324945" containerName="ceilometer-notification-agent" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.906845 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1ca2e06-431e-45b9-b492-20c491324945" containerName="ceilometer-notification-agent" Mar 18 20:28:38 crc kubenswrapper[4950]: E0318 20:28:38.906904 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1ca2e06-431e-45b9-b492-20c491324945" containerName="sg-core" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.906957 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1ca2e06-431e-45b9-b492-20c491324945" containerName="sg-core" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.907145 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1ca2e06-431e-45b9-b492-20c491324945" containerName="ceilometer-central-agent" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.907211 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1ca2e06-431e-45b9-b492-20c491324945" containerName="proxy-httpd" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.907269 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1ca2e06-431e-45b9-b492-20c491324945" containerName="sg-core" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.907341 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1ca2e06-431e-45b9-b492-20c491324945" containerName="ceilometer-notification-agent" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.908858 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.911063 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.912697 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 18 20:28:38 crc kubenswrapper[4950]: I0318 20:28:38.922857 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.042105 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-log-httpd\") pod \"ceilometer-0\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " pod="openstack/ceilometer-0" Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.042156 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-scripts\") pod \"ceilometer-0\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " pod="openstack/ceilometer-0" Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.042177 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-config-data\") pod \"ceilometer-0\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " pod="openstack/ceilometer-0" Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.042214 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " pod="openstack/ceilometer-0" Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.042259 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " pod="openstack/ceilometer-0" Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.042294 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn6vh\" (UniqueName: \"kubernetes.io/projected/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-kube-api-access-cn6vh\") pod \"ceilometer-0\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " pod="openstack/ceilometer-0" Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.042308 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-run-httpd\") pod \"ceilometer-0\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " pod="openstack/ceilometer-0" Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.143482 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-scripts\") pod \"ceilometer-0\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " pod="openstack/ceilometer-0" Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.143525 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-config-data\") pod \"ceilometer-0\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " pod="openstack/ceilometer-0" Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.143581 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " pod="openstack/ceilometer-0" Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.143635 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " pod="openstack/ceilometer-0" Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.143685 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn6vh\" (UniqueName: \"kubernetes.io/projected/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-kube-api-access-cn6vh\") pod \"ceilometer-0\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " pod="openstack/ceilometer-0" Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.143701 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-run-httpd\") pod \"ceilometer-0\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " pod="openstack/ceilometer-0" Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.144509 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-run-httpd\") pod \"ceilometer-0\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " pod="openstack/ceilometer-0" Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.144576 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-log-httpd\") pod \"ceilometer-0\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " pod="openstack/ceilometer-0" Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.144807 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-log-httpd\") pod \"ceilometer-0\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " pod="openstack/ceilometer-0" Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.149055 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-config-data\") pod \"ceilometer-0\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " pod="openstack/ceilometer-0" Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.150025 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " pod="openstack/ceilometer-0" Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.154140 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-scripts\") pod \"ceilometer-0\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " pod="openstack/ceilometer-0" Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.156307 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " pod="openstack/ceilometer-0" Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.163141 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn6vh\" (UniqueName: \"kubernetes.io/projected/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-kube-api-access-cn6vh\") pod \"ceilometer-0\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " pod="openstack/ceilometer-0" Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.222814 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.489384 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1ca2e06-431e-45b9-b492-20c491324945" path="/var/lib/kubelet/pods/e1ca2e06-431e-45b9-b492-20c491324945/volumes" Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.755618 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.765819 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:28:39 crc kubenswrapper[4950]: I0318 20:28:39.908677 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:40 crc kubenswrapper[4950]: I0318 20:28:40.579283 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4","Type":"ContainerStarted","Data":"26694b172dfc476cd9f43d5c941bbc27e2c7c1848917d1ef425f19f9be959ab1"} Mar 18 20:28:40 crc kubenswrapper[4950]: I0318 20:28:40.579696 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4","Type":"ContainerStarted","Data":"801c50e703f86faabde87a79d4144276de5de95b7f88ad3a287ecbaffa5a7b95"} Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.149115 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.293982 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e016c0d-1c05-48ee-8f29-17d94a91edbc-combined-ca-bundle\") pod \"0e016c0d-1c05-48ee-8f29-17d94a91edbc\" (UID: \"0e016c0d-1c05-48ee-8f29-17d94a91edbc\") " Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.294226 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4ksr\" (UniqueName: \"kubernetes.io/projected/0e016c0d-1c05-48ee-8f29-17d94a91edbc-kube-api-access-t4ksr\") pod \"0e016c0d-1c05-48ee-8f29-17d94a91edbc\" (UID: \"0e016c0d-1c05-48ee-8f29-17d94a91edbc\") " Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.294296 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e016c0d-1c05-48ee-8f29-17d94a91edbc-config-data\") pod \"0e016c0d-1c05-48ee-8f29-17d94a91edbc\" (UID: \"0e016c0d-1c05-48ee-8f29-17d94a91edbc\") " Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.294333 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e016c0d-1c05-48ee-8f29-17d94a91edbc-logs\") pod \"0e016c0d-1c05-48ee-8f29-17d94a91edbc\" (UID: \"0e016c0d-1c05-48ee-8f29-17d94a91edbc\") " Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.294844 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e016c0d-1c05-48ee-8f29-17d94a91edbc-logs" (OuterVolumeSpecName: "logs") pod "0e016c0d-1c05-48ee-8f29-17d94a91edbc" (UID: "0e016c0d-1c05-48ee-8f29-17d94a91edbc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.303897 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e016c0d-1c05-48ee-8f29-17d94a91edbc-kube-api-access-t4ksr" (OuterVolumeSpecName: "kube-api-access-t4ksr") pod "0e016c0d-1c05-48ee-8f29-17d94a91edbc" (UID: "0e016c0d-1c05-48ee-8f29-17d94a91edbc"). InnerVolumeSpecName "kube-api-access-t4ksr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.319579 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e016c0d-1c05-48ee-8f29-17d94a91edbc-config-data" (OuterVolumeSpecName: "config-data") pod "0e016c0d-1c05-48ee-8f29-17d94a91edbc" (UID: "0e016c0d-1c05-48ee-8f29-17d94a91edbc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.332568 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e016c0d-1c05-48ee-8f29-17d94a91edbc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0e016c0d-1c05-48ee-8f29-17d94a91edbc" (UID: "0e016c0d-1c05-48ee-8f29-17d94a91edbc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.396172 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e016c0d-1c05-48ee-8f29-17d94a91edbc-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.396215 4950 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e016c0d-1c05-48ee-8f29-17d94a91edbc-logs\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.396224 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e016c0d-1c05-48ee-8f29-17d94a91edbc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.396234 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4ksr\" (UniqueName: \"kubernetes.io/projected/0e016c0d-1c05-48ee-8f29-17d94a91edbc-kube-api-access-t4ksr\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.592919 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4","Type":"ContainerStarted","Data":"f68d25ef222995d3bf8cd1815e1d419c78c61a550b24b4ef918dce44f60542ff"} Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.597117 4950 generic.go:334] "Generic (PLEG): container finished" podID="0e016c0d-1c05-48ee-8f29-17d94a91edbc" containerID="0602d592b25338753eb5b4f815d3a5dd411a9ea2df2b92bca1139b1c3757e650" exitCode=0 Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.597170 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.597176 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0e016c0d-1c05-48ee-8f29-17d94a91edbc","Type":"ContainerDied","Data":"0602d592b25338753eb5b4f815d3a5dd411a9ea2df2b92bca1139b1c3757e650"} Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.597211 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0e016c0d-1c05-48ee-8f29-17d94a91edbc","Type":"ContainerDied","Data":"92e4478cbfd3380ccb1f2565a74f5876934412cd3e70fb52104a61106ed33cb4"} Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.597230 4950 scope.go:117] "RemoveContainer" containerID="0602d592b25338753eb5b4f815d3a5dd411a9ea2df2b92bca1139b1c3757e650" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.629038 4950 scope.go:117] "RemoveContainer" containerID="cc0d53162ba3f45a5b04553df12d6904f100b9b9a962ca5c925248bda6ca150e" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.641206 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.657648 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.670079 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 18 20:28:41 crc kubenswrapper[4950]: E0318 20:28:41.671368 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e016c0d-1c05-48ee-8f29-17d94a91edbc" containerName="nova-api-api" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.671389 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e016c0d-1c05-48ee-8f29-17d94a91edbc" containerName="nova-api-api" Mar 18 20:28:41 crc kubenswrapper[4950]: E0318 20:28:41.671440 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e016c0d-1c05-48ee-8f29-17d94a91edbc" containerName="nova-api-log" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.671448 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e016c0d-1c05-48ee-8f29-17d94a91edbc" containerName="nova-api-log" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.671607 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e016c0d-1c05-48ee-8f29-17d94a91edbc" containerName="nova-api-api" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.671625 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e016c0d-1c05-48ee-8f29-17d94a91edbc" containerName="nova-api-log" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.672539 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.677948 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.678142 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.678301 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.702666 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.705901 4950 scope.go:117] "RemoveContainer" containerID="0602d592b25338753eb5b4f815d3a5dd411a9ea2df2b92bca1139b1c3757e650" Mar 18 20:28:41 crc kubenswrapper[4950]: E0318 20:28:41.711600 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0602d592b25338753eb5b4f815d3a5dd411a9ea2df2b92bca1139b1c3757e650\": container with ID starting with 0602d592b25338753eb5b4f815d3a5dd411a9ea2df2b92bca1139b1c3757e650 not found: ID does not exist" containerID="0602d592b25338753eb5b4f815d3a5dd411a9ea2df2b92bca1139b1c3757e650" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.711646 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0602d592b25338753eb5b4f815d3a5dd411a9ea2df2b92bca1139b1c3757e650"} err="failed to get container status \"0602d592b25338753eb5b4f815d3a5dd411a9ea2df2b92bca1139b1c3757e650\": rpc error: code = NotFound desc = could not find container \"0602d592b25338753eb5b4f815d3a5dd411a9ea2df2b92bca1139b1c3757e650\": container with ID starting with 0602d592b25338753eb5b4f815d3a5dd411a9ea2df2b92bca1139b1c3757e650 not found: ID does not exist" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.711674 4950 scope.go:117] "RemoveContainer" containerID="cc0d53162ba3f45a5b04553df12d6904f100b9b9a962ca5c925248bda6ca150e" Mar 18 20:28:41 crc kubenswrapper[4950]: E0318 20:28:41.715539 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc0d53162ba3f45a5b04553df12d6904f100b9b9a962ca5c925248bda6ca150e\": container with ID starting with cc0d53162ba3f45a5b04553df12d6904f100b9b9a962ca5c925248bda6ca150e not found: ID does not exist" containerID="cc0d53162ba3f45a5b04553df12d6904f100b9b9a962ca5c925248bda6ca150e" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.715587 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc0d53162ba3f45a5b04553df12d6904f100b9b9a962ca5c925248bda6ca150e"} err="failed to get container status \"cc0d53162ba3f45a5b04553df12d6904f100b9b9a962ca5c925248bda6ca150e\": rpc error: code = NotFound desc = could not find container \"cc0d53162ba3f45a5b04553df12d6904f100b9b9a962ca5c925248bda6ca150e\": container with ID starting with cc0d53162ba3f45a5b04553df12d6904f100b9b9a962ca5c925248bda6ca150e not found: ID does not exist" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.805700 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5vqj\" (UniqueName: \"kubernetes.io/projected/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-kube-api-access-t5vqj\") pod \"nova-api-0\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " pod="openstack/nova-api-0" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.805797 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-public-tls-certs\") pod \"nova-api-0\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " pod="openstack/nova-api-0" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.805819 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-config-data\") pod \"nova-api-0\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " pod="openstack/nova-api-0" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.805850 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " pod="openstack/nova-api-0" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.806055 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " pod="openstack/nova-api-0" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.806188 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-logs\") pod \"nova-api-0\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " pod="openstack/nova-api-0" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.908044 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5vqj\" (UniqueName: \"kubernetes.io/projected/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-kube-api-access-t5vqj\") pod \"nova-api-0\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " pod="openstack/nova-api-0" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.908096 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-public-tls-certs\") pod \"nova-api-0\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " pod="openstack/nova-api-0" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.908116 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-config-data\") pod \"nova-api-0\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " pod="openstack/nova-api-0" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.908157 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " pod="openstack/nova-api-0" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.908229 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " pod="openstack/nova-api-0" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.908267 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-logs\") pod \"nova-api-0\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " pod="openstack/nova-api-0" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.908717 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-logs\") pod \"nova-api-0\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " pod="openstack/nova-api-0" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.912928 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-config-data\") pod \"nova-api-0\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " pod="openstack/nova-api-0" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.913069 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " pod="openstack/nova-api-0" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.915174 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " pod="openstack/nova-api-0" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.932021 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5vqj\" (UniqueName: \"kubernetes.io/projected/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-kube-api-access-t5vqj\") pod \"nova-api-0\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " pod="openstack/nova-api-0" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.932052 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-public-tls-certs\") pod \"nova-api-0\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " pod="openstack/nova-api-0" Mar 18 20:28:41 crc kubenswrapper[4950]: I0318 20:28:41.997531 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 18 20:28:42 crc kubenswrapper[4950]: W0318 20:28:42.444401 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fd4b49e_aca5_4160_9bb9_2cf5b4925022.slice/crio-0399f469bfa4a2b6ba7e9a03d720fac5c36881d1a9b56c3aa70bbf8f31cc0e1f WatchSource:0}: Error finding container 0399f469bfa4a2b6ba7e9a03d720fac5c36881d1a9b56c3aa70bbf8f31cc0e1f: Status 404 returned error can't find the container with id 0399f469bfa4a2b6ba7e9a03d720fac5c36881d1a9b56c3aa70bbf8f31cc0e1f Mar 18 20:28:42 crc kubenswrapper[4950]: I0318 20:28:42.448004 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 18 20:28:42 crc kubenswrapper[4950]: I0318 20:28:42.605400 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2fd4b49e-aca5-4160-9bb9-2cf5b4925022","Type":"ContainerStarted","Data":"5660ff2159c127514406d6fa0a0e843212c22bba90a9f07c6e263a6e4d71c7c0"} Mar 18 20:28:42 crc kubenswrapper[4950]: I0318 20:28:42.605689 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2fd4b49e-aca5-4160-9bb9-2cf5b4925022","Type":"ContainerStarted","Data":"0399f469bfa4a2b6ba7e9a03d720fac5c36881d1a9b56c3aa70bbf8f31cc0e1f"} Mar 18 20:28:42 crc kubenswrapper[4950]: I0318 20:28:42.607801 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4","Type":"ContainerStarted","Data":"ed2b95458b4f8c61e19a214e892982d2ffffe72cb7ccec286bc5639939717059"} Mar 18 20:28:43 crc kubenswrapper[4950]: I0318 20:28:43.490664 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e016c0d-1c05-48ee-8f29-17d94a91edbc" path="/var/lib/kubelet/pods/0e016c0d-1c05-48ee-8f29-17d94a91edbc/volumes" Mar 18 20:28:43 crc kubenswrapper[4950]: I0318 20:28:43.621112 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2fd4b49e-aca5-4160-9bb9-2cf5b4925022","Type":"ContainerStarted","Data":"6c714f29806df1dfc267dcf912a6fa3c8048dd9245e022236aa98faf390b498c"} Mar 18 20:28:44 crc kubenswrapper[4950]: I0318 20:28:44.649770 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4","Type":"ContainerStarted","Data":"fe97b0c895351b7dbcef45f4ef47fb1416a35b2d85d96b295ebe69a9202c7022"} Mar 18 20:28:44 crc kubenswrapper[4950]: I0318 20:28:44.649898 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" containerName="ceilometer-central-agent" containerID="cri-o://26694b172dfc476cd9f43d5c941bbc27e2c7c1848917d1ef425f19f9be959ab1" gracePeriod=30 Mar 18 20:28:44 crc kubenswrapper[4950]: I0318 20:28:44.650241 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 18 20:28:44 crc kubenswrapper[4950]: I0318 20:28:44.650340 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" containerName="proxy-httpd" containerID="cri-o://fe97b0c895351b7dbcef45f4ef47fb1416a35b2d85d96b295ebe69a9202c7022" gracePeriod=30 Mar 18 20:28:44 crc kubenswrapper[4950]: I0318 20:28:44.650540 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" containerName="ceilometer-notification-agent" containerID="cri-o://f68d25ef222995d3bf8cd1815e1d419c78c61a550b24b4ef918dce44f60542ff" gracePeriod=30 Mar 18 20:28:44 crc kubenswrapper[4950]: I0318 20:28:44.650638 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" containerName="sg-core" containerID="cri-o://ed2b95458b4f8c61e19a214e892982d2ffffe72cb7ccec286bc5639939717059" gracePeriod=30 Mar 18 20:28:44 crc kubenswrapper[4950]: I0318 20:28:44.687883 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.384201105 podStartE2EDuration="6.687854205s" podCreationTimestamp="2026-03-18 20:28:38 +0000 UTC" firstStartedPulling="2026-03-18 20:28:39.755663232 +0000 UTC m=+1332.996505100" lastFinishedPulling="2026-03-18 20:28:44.059316322 +0000 UTC m=+1337.300158200" observedRunningTime="2026-03-18 20:28:44.68569565 +0000 UTC m=+1337.926537528" watchObservedRunningTime="2026-03-18 20:28:44.687854205 +0000 UTC m=+1337.928696113" Mar 18 20:28:44 crc kubenswrapper[4950]: I0318 20:28:44.704977 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.704959616 podStartE2EDuration="3.704959616s" podCreationTimestamp="2026-03-18 20:28:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:28:43.642255563 +0000 UTC m=+1336.883097431" watchObservedRunningTime="2026-03-18 20:28:44.704959616 +0000 UTC m=+1337.945801484" Mar 18 20:28:44 crc kubenswrapper[4950]: I0318 20:28:44.909442 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:44 crc kubenswrapper[4950]: I0318 20:28:44.925892 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.048460 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-776c49c95f-d9656" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.115248 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c6bb4bf9-h85pn"] Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.115769 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" podUID="e56878e7-f76d-48b6-b760-7f045d214b31" containerName="dnsmasq-dns" containerID="cri-o://a3962b4da8a143dd1b9742fb29395bfd12a2d9a771e9097dee95a2586f9c1f6e" gracePeriod=10 Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.611392 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.631431 4950 scope.go:117] "RemoveContainer" containerID="7244a7cd7d6e9bf822acc6b28539c9730114bd6610cbc203b10c7c13f309f94a" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.676751 4950 generic.go:334] "Generic (PLEG): container finished" podID="5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" containerID="fe97b0c895351b7dbcef45f4ef47fb1416a35b2d85d96b295ebe69a9202c7022" exitCode=0 Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.676788 4950 generic.go:334] "Generic (PLEG): container finished" podID="5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" containerID="ed2b95458b4f8c61e19a214e892982d2ffffe72cb7ccec286bc5639939717059" exitCode=2 Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.676797 4950 generic.go:334] "Generic (PLEG): container finished" podID="5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" containerID="f68d25ef222995d3bf8cd1815e1d419c78c61a550b24b4ef918dce44f60542ff" exitCode=0 Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.676858 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4","Type":"ContainerDied","Data":"fe97b0c895351b7dbcef45f4ef47fb1416a35b2d85d96b295ebe69a9202c7022"} Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.676922 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4","Type":"ContainerDied","Data":"ed2b95458b4f8c61e19a214e892982d2ffffe72cb7ccec286bc5639939717059"} Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.676936 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4","Type":"ContainerDied","Data":"f68d25ef222995d3bf8cd1815e1d419c78c61a550b24b4ef918dce44f60542ff"} Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.687544 4950 generic.go:334] "Generic (PLEG): container finished" podID="e56878e7-f76d-48b6-b760-7f045d214b31" containerID="a3962b4da8a143dd1b9742fb29395bfd12a2d9a771e9097dee95a2586f9c1f6e" exitCode=0 Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.688827 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" event={"ID":"e56878e7-f76d-48b6-b760-7f045d214b31","Type":"ContainerDied","Data":"a3962b4da8a143dd1b9742fb29395bfd12a2d9a771e9097dee95a2586f9c1f6e"} Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.688857 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" event={"ID":"e56878e7-f76d-48b6-b760-7f045d214b31","Type":"ContainerDied","Data":"3744ecbd0c76c69e44bee2529488df2bd0b13ac0e5659344ec43f4bd14aace9f"} Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.688873 4950 scope.go:117] "RemoveContainer" containerID="a3962b4da8a143dd1b9742fb29395bfd12a2d9a771e9097dee95a2586f9c1f6e" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.689521 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-dns-svc\") pod \"e56878e7-f76d-48b6-b760-7f045d214b31\" (UID: \"e56878e7-f76d-48b6-b760-7f045d214b31\") " Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.689684 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-ovsdbserver-nb\") pod \"e56878e7-f76d-48b6-b760-7f045d214b31\" (UID: \"e56878e7-f76d-48b6-b760-7f045d214b31\") " Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.689773 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-ovsdbserver-sb\") pod \"e56878e7-f76d-48b6-b760-7f045d214b31\" (UID: \"e56878e7-f76d-48b6-b760-7f045d214b31\") " Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.689819 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zn4rt\" (UniqueName: \"kubernetes.io/projected/e56878e7-f76d-48b6-b760-7f045d214b31-kube-api-access-zn4rt\") pod \"e56878e7-f76d-48b6-b760-7f045d214b31\" (UID: \"e56878e7-f76d-48b6-b760-7f045d214b31\") " Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.689925 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-config\") pod \"e56878e7-f76d-48b6-b760-7f045d214b31\" (UID: \"e56878e7-f76d-48b6-b760-7f045d214b31\") " Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.690542 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c6bb4bf9-h85pn" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.705921 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e56878e7-f76d-48b6-b760-7f045d214b31-kube-api-access-zn4rt" (OuterVolumeSpecName: "kube-api-access-zn4rt") pod "e56878e7-f76d-48b6-b760-7f045d214b31" (UID: "e56878e7-f76d-48b6-b760-7f045d214b31"). InnerVolumeSpecName "kube-api-access-zn4rt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.715922 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.729013 4950 scope.go:117] "RemoveContainer" containerID="e3e6920bc9551fc817f39f8bce6232ff1cfebfebda7b624eed583416cbe4e955" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.749327 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e56878e7-f76d-48b6-b760-7f045d214b31" (UID: "e56878e7-f76d-48b6-b760-7f045d214b31"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.779011 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-config" (OuterVolumeSpecName: "config") pod "e56878e7-f76d-48b6-b760-7f045d214b31" (UID: "e56878e7-f76d-48b6-b760-7f045d214b31"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.788839 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e56878e7-f76d-48b6-b760-7f045d214b31" (UID: "e56878e7-f76d-48b6-b760-7f045d214b31"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.790836 4950 scope.go:117] "RemoveContainer" containerID="a3962b4da8a143dd1b9742fb29395bfd12a2d9a771e9097dee95a2586f9c1f6e" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.792722 4950 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.792763 4950 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.792787 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zn4rt\" (UniqueName: \"kubernetes.io/projected/e56878e7-f76d-48b6-b760-7f045d214b31-kube-api-access-zn4rt\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.792801 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:45 crc kubenswrapper[4950]: E0318 20:28:45.793600 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3962b4da8a143dd1b9742fb29395bfd12a2d9a771e9097dee95a2586f9c1f6e\": container with ID starting with a3962b4da8a143dd1b9742fb29395bfd12a2d9a771e9097dee95a2586f9c1f6e not found: ID does not exist" containerID="a3962b4da8a143dd1b9742fb29395bfd12a2d9a771e9097dee95a2586f9c1f6e" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.793636 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3962b4da8a143dd1b9742fb29395bfd12a2d9a771e9097dee95a2586f9c1f6e"} err="failed to get container status \"a3962b4da8a143dd1b9742fb29395bfd12a2d9a771e9097dee95a2586f9c1f6e\": rpc error: code = NotFound desc = could not find container \"a3962b4da8a143dd1b9742fb29395bfd12a2d9a771e9097dee95a2586f9c1f6e\": container with ID starting with a3962b4da8a143dd1b9742fb29395bfd12a2d9a771e9097dee95a2586f9c1f6e not found: ID does not exist" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.793664 4950 scope.go:117] "RemoveContainer" containerID="e3e6920bc9551fc817f39f8bce6232ff1cfebfebda7b624eed583416cbe4e955" Mar 18 20:28:45 crc kubenswrapper[4950]: E0318 20:28:45.799638 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3e6920bc9551fc817f39f8bce6232ff1cfebfebda7b624eed583416cbe4e955\": container with ID starting with e3e6920bc9551fc817f39f8bce6232ff1cfebfebda7b624eed583416cbe4e955 not found: ID does not exist" containerID="e3e6920bc9551fc817f39f8bce6232ff1cfebfebda7b624eed583416cbe4e955" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.799696 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3e6920bc9551fc817f39f8bce6232ff1cfebfebda7b624eed583416cbe4e955"} err="failed to get container status \"e3e6920bc9551fc817f39f8bce6232ff1cfebfebda7b624eed583416cbe4e955\": rpc error: code = NotFound desc = could not find container \"e3e6920bc9551fc817f39f8bce6232ff1cfebfebda7b624eed583416cbe4e955\": container with ID starting with e3e6920bc9551fc817f39f8bce6232ff1cfebfebda7b624eed583416cbe4e955 not found: ID does not exist" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.816346 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e56878e7-f76d-48b6-b760-7f045d214b31" (UID: "e56878e7-f76d-48b6-b760-7f045d214b31"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.894935 4950 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e56878e7-f76d-48b6-b760-7f045d214b31-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.921932 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-lj42n"] Mar 18 20:28:45 crc kubenswrapper[4950]: E0318 20:28:45.922302 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e56878e7-f76d-48b6-b760-7f045d214b31" containerName="dnsmasq-dns" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.922320 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="e56878e7-f76d-48b6-b760-7f045d214b31" containerName="dnsmasq-dns" Mar 18 20:28:45 crc kubenswrapper[4950]: E0318 20:28:45.922343 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e56878e7-f76d-48b6-b760-7f045d214b31" containerName="init" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.922350 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="e56878e7-f76d-48b6-b760-7f045d214b31" containerName="init" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.922530 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="e56878e7-f76d-48b6-b760-7f045d214b31" containerName="dnsmasq-dns" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.923060 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lj42n" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.925549 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.925735 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.960071 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-lj42n"] Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.995981 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f58dabee-7195-40e9-952e-9dfeefc1ee08-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-lj42n\" (UID: \"f58dabee-7195-40e9-952e-9dfeefc1ee08\") " pod="openstack/nova-cell1-cell-mapping-lj42n" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.996028 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f58dabee-7195-40e9-952e-9dfeefc1ee08-scripts\") pod \"nova-cell1-cell-mapping-lj42n\" (UID: \"f58dabee-7195-40e9-952e-9dfeefc1ee08\") " pod="openstack/nova-cell1-cell-mapping-lj42n" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.996151 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f58dabee-7195-40e9-952e-9dfeefc1ee08-config-data\") pod \"nova-cell1-cell-mapping-lj42n\" (UID: \"f58dabee-7195-40e9-952e-9dfeefc1ee08\") " pod="openstack/nova-cell1-cell-mapping-lj42n" Mar 18 20:28:45 crc kubenswrapper[4950]: I0318 20:28:45.996187 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grsk9\" (UniqueName: \"kubernetes.io/projected/f58dabee-7195-40e9-952e-9dfeefc1ee08-kube-api-access-grsk9\") pod \"nova-cell1-cell-mapping-lj42n\" (UID: \"f58dabee-7195-40e9-952e-9dfeefc1ee08\") " pod="openstack/nova-cell1-cell-mapping-lj42n" Mar 18 20:28:46 crc kubenswrapper[4950]: I0318 20:28:46.027658 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c6bb4bf9-h85pn"] Mar 18 20:28:46 crc kubenswrapper[4950]: I0318 20:28:46.035951 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c6bb4bf9-h85pn"] Mar 18 20:28:46 crc kubenswrapper[4950]: I0318 20:28:46.098019 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grsk9\" (UniqueName: \"kubernetes.io/projected/f58dabee-7195-40e9-952e-9dfeefc1ee08-kube-api-access-grsk9\") pod \"nova-cell1-cell-mapping-lj42n\" (UID: \"f58dabee-7195-40e9-952e-9dfeefc1ee08\") " pod="openstack/nova-cell1-cell-mapping-lj42n" Mar 18 20:28:46 crc kubenswrapper[4950]: I0318 20:28:46.098114 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f58dabee-7195-40e9-952e-9dfeefc1ee08-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-lj42n\" (UID: \"f58dabee-7195-40e9-952e-9dfeefc1ee08\") " pod="openstack/nova-cell1-cell-mapping-lj42n" Mar 18 20:28:46 crc kubenswrapper[4950]: I0318 20:28:46.098142 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f58dabee-7195-40e9-952e-9dfeefc1ee08-scripts\") pod \"nova-cell1-cell-mapping-lj42n\" (UID: \"f58dabee-7195-40e9-952e-9dfeefc1ee08\") " pod="openstack/nova-cell1-cell-mapping-lj42n" Mar 18 20:28:46 crc kubenswrapper[4950]: I0318 20:28:46.098222 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f58dabee-7195-40e9-952e-9dfeefc1ee08-config-data\") pod \"nova-cell1-cell-mapping-lj42n\" (UID: \"f58dabee-7195-40e9-952e-9dfeefc1ee08\") " pod="openstack/nova-cell1-cell-mapping-lj42n" Mar 18 20:28:46 crc kubenswrapper[4950]: I0318 20:28:46.104522 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f58dabee-7195-40e9-952e-9dfeefc1ee08-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-lj42n\" (UID: \"f58dabee-7195-40e9-952e-9dfeefc1ee08\") " pod="openstack/nova-cell1-cell-mapping-lj42n" Mar 18 20:28:46 crc kubenswrapper[4950]: I0318 20:28:46.104960 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f58dabee-7195-40e9-952e-9dfeefc1ee08-config-data\") pod \"nova-cell1-cell-mapping-lj42n\" (UID: \"f58dabee-7195-40e9-952e-9dfeefc1ee08\") " pod="openstack/nova-cell1-cell-mapping-lj42n" Mar 18 20:28:46 crc kubenswrapper[4950]: I0318 20:28:46.111843 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f58dabee-7195-40e9-952e-9dfeefc1ee08-scripts\") pod \"nova-cell1-cell-mapping-lj42n\" (UID: \"f58dabee-7195-40e9-952e-9dfeefc1ee08\") " pod="openstack/nova-cell1-cell-mapping-lj42n" Mar 18 20:28:46 crc kubenswrapper[4950]: I0318 20:28:46.115511 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grsk9\" (UniqueName: \"kubernetes.io/projected/f58dabee-7195-40e9-952e-9dfeefc1ee08-kube-api-access-grsk9\") pod \"nova-cell1-cell-mapping-lj42n\" (UID: \"f58dabee-7195-40e9-952e-9dfeefc1ee08\") " pod="openstack/nova-cell1-cell-mapping-lj42n" Mar 18 20:28:46 crc kubenswrapper[4950]: I0318 20:28:46.246391 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lj42n" Mar 18 20:28:46 crc kubenswrapper[4950]: I0318 20:28:46.701353 4950 generic.go:334] "Generic (PLEG): container finished" podID="5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" containerID="26694b172dfc476cd9f43d5c941bbc27e2c7c1848917d1ef425f19f9be959ab1" exitCode=0 Mar 18 20:28:46 crc kubenswrapper[4950]: I0318 20:28:46.701694 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4","Type":"ContainerDied","Data":"26694b172dfc476cd9f43d5c941bbc27e2c7c1848917d1ef425f19f9be959ab1"} Mar 18 20:28:46 crc kubenswrapper[4950]: I0318 20:28:46.809581 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-lj42n"] Mar 18 20:28:46 crc kubenswrapper[4950]: W0318 20:28:46.814011 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf58dabee_7195_40e9_952e_9dfeefc1ee08.slice/crio-14448fcd21bdc40920382f8cee4a3af4a7e70f58e2e7e48885f443ec423346e9 WatchSource:0}: Error finding container 14448fcd21bdc40920382f8cee4a3af4a7e70f58e2e7e48885f443ec423346e9: Status 404 returned error can't find the container with id 14448fcd21bdc40920382f8cee4a3af4a7e70f58e2e7e48885f443ec423346e9 Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.439650 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.496169 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e56878e7-f76d-48b6-b760-7f045d214b31" path="/var/lib/kubelet/pods/e56878e7-f76d-48b6-b760-7f045d214b31/volumes" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.522000 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cn6vh\" (UniqueName: \"kubernetes.io/projected/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-kube-api-access-cn6vh\") pod \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.522173 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-config-data\") pod \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.522247 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-run-httpd\") pod \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.522293 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-scripts\") pod \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.522332 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-sg-core-conf-yaml\") pod \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.522506 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-combined-ca-bundle\") pod \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.522546 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-log-httpd\") pod \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\" (UID: \"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4\") " Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.525374 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" (UID: "5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.525887 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" (UID: "5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.541983 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-scripts" (OuterVolumeSpecName: "scripts") pod "5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" (UID: "5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.547248 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-kube-api-access-cn6vh" (OuterVolumeSpecName: "kube-api-access-cn6vh") pod "5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" (UID: "5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4"). InnerVolumeSpecName "kube-api-access-cn6vh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.555923 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" (UID: "5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.601080 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" (UID: "5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.628029 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.628173 4950 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.628231 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cn6vh\" (UniqueName: \"kubernetes.io/projected/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-kube-api-access-cn6vh\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.628308 4950 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.628363 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.628432 4950 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.633644 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-config-data" (OuterVolumeSpecName: "config-data") pod "5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" (UID: "5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.722879 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4","Type":"ContainerDied","Data":"801c50e703f86faabde87a79d4144276de5de95b7f88ad3a287ecbaffa5a7b95"} Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.723249 4950 scope.go:117] "RemoveContainer" containerID="fe97b0c895351b7dbcef45f4ef47fb1416a35b2d85d96b295ebe69a9202c7022" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.722903 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.729647 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.733102 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lj42n" event={"ID":"f58dabee-7195-40e9-952e-9dfeefc1ee08","Type":"ContainerStarted","Data":"656f6974b9820f48c5b34d3c8007059ae8bf8ad5ac44bf59d04c6ad0ee0e58f1"} Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.733156 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lj42n" event={"ID":"f58dabee-7195-40e9-952e-9dfeefc1ee08","Type":"ContainerStarted","Data":"14448fcd21bdc40920382f8cee4a3af4a7e70f58e2e7e48885f443ec423346e9"} Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.754986 4950 scope.go:117] "RemoveContainer" containerID="ed2b95458b4f8c61e19a214e892982d2ffffe72cb7ccec286bc5639939717059" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.763851 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-lj42n" podStartSLOduration=2.763830124 podStartE2EDuration="2.763830124s" podCreationTimestamp="2026-03-18 20:28:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:28:47.753526519 +0000 UTC m=+1340.994368387" watchObservedRunningTime="2026-03-18 20:28:47.763830124 +0000 UTC m=+1341.004671992" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.785767 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.788739 4950 scope.go:117] "RemoveContainer" containerID="f68d25ef222995d3bf8cd1815e1d419c78c61a550b24b4ef918dce44f60542ff" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.809592 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.824655 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:28:47 crc kubenswrapper[4950]: E0318 20:28:47.824996 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" containerName="ceilometer-central-agent" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.825013 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" containerName="ceilometer-central-agent" Mar 18 20:28:47 crc kubenswrapper[4950]: E0318 20:28:47.825032 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" containerName="sg-core" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.825038 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" containerName="sg-core" Mar 18 20:28:47 crc kubenswrapper[4950]: E0318 20:28:47.825052 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" containerName="proxy-httpd" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.825059 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" containerName="proxy-httpd" Mar 18 20:28:47 crc kubenswrapper[4950]: E0318 20:28:47.825067 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" containerName="ceilometer-notification-agent" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.825073 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" containerName="ceilometer-notification-agent" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.825223 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" containerName="proxy-httpd" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.825240 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" containerName="ceilometer-central-agent" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.825251 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" containerName="ceilometer-notification-agent" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.825263 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" containerName="sg-core" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.826819 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.830971 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.832898 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.833382 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.863008 4950 scope.go:117] "RemoveContainer" containerID="26694b172dfc476cd9f43d5c941bbc27e2c7c1848917d1ef425f19f9be959ab1" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.931991 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " pod="openstack/ceilometer-0" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.932066 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00969ec6-eca3-4790-b69f-ba825f809823-log-httpd\") pod \"ceilometer-0\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " pod="openstack/ceilometer-0" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.932091 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00969ec6-eca3-4790-b69f-ba825f809823-run-httpd\") pod \"ceilometer-0\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " pod="openstack/ceilometer-0" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.932120 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-config-data\") pod \"ceilometer-0\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " pod="openstack/ceilometer-0" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.932153 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " pod="openstack/ceilometer-0" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.932464 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fg7pr\" (UniqueName: \"kubernetes.io/projected/00969ec6-eca3-4790-b69f-ba825f809823-kube-api-access-fg7pr\") pod \"ceilometer-0\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " pod="openstack/ceilometer-0" Mar 18 20:28:47 crc kubenswrapper[4950]: I0318 20:28:47.932634 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-scripts\") pod \"ceilometer-0\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " pod="openstack/ceilometer-0" Mar 18 20:28:47 crc kubenswrapper[4950]: E0318 20:28:47.951690 4950 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b2e8fce_f0e3_4411_94fe_9e9bac3e22e4.slice/crio-801c50e703f86faabde87a79d4144276de5de95b7f88ad3a287ecbaffa5a7b95\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b2e8fce_f0e3_4411_94fe_9e9bac3e22e4.slice\": RecentStats: unable to find data in memory cache]" Mar 18 20:28:48 crc kubenswrapper[4950]: I0318 20:28:48.034774 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fg7pr\" (UniqueName: \"kubernetes.io/projected/00969ec6-eca3-4790-b69f-ba825f809823-kube-api-access-fg7pr\") pod \"ceilometer-0\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " pod="openstack/ceilometer-0" Mar 18 20:28:48 crc kubenswrapper[4950]: I0318 20:28:48.034875 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-scripts\") pod \"ceilometer-0\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " pod="openstack/ceilometer-0" Mar 18 20:28:48 crc kubenswrapper[4950]: I0318 20:28:48.034932 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " pod="openstack/ceilometer-0" Mar 18 20:28:48 crc kubenswrapper[4950]: I0318 20:28:48.034973 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00969ec6-eca3-4790-b69f-ba825f809823-log-httpd\") pod \"ceilometer-0\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " pod="openstack/ceilometer-0" Mar 18 20:28:48 crc kubenswrapper[4950]: I0318 20:28:48.035004 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00969ec6-eca3-4790-b69f-ba825f809823-run-httpd\") pod \"ceilometer-0\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " pod="openstack/ceilometer-0" Mar 18 20:28:48 crc kubenswrapper[4950]: I0318 20:28:48.035036 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-config-data\") pod \"ceilometer-0\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " pod="openstack/ceilometer-0" Mar 18 20:28:48 crc kubenswrapper[4950]: I0318 20:28:48.035074 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " pod="openstack/ceilometer-0" Mar 18 20:28:48 crc kubenswrapper[4950]: I0318 20:28:48.035453 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00969ec6-eca3-4790-b69f-ba825f809823-log-httpd\") pod \"ceilometer-0\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " pod="openstack/ceilometer-0" Mar 18 20:28:48 crc kubenswrapper[4950]: I0318 20:28:48.035490 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00969ec6-eca3-4790-b69f-ba825f809823-run-httpd\") pod \"ceilometer-0\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " pod="openstack/ceilometer-0" Mar 18 20:28:48 crc kubenswrapper[4950]: I0318 20:28:48.039331 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " pod="openstack/ceilometer-0" Mar 18 20:28:48 crc kubenswrapper[4950]: I0318 20:28:48.041756 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-scripts\") pod \"ceilometer-0\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " pod="openstack/ceilometer-0" Mar 18 20:28:48 crc kubenswrapper[4950]: I0318 20:28:48.042598 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " pod="openstack/ceilometer-0" Mar 18 20:28:48 crc kubenswrapper[4950]: I0318 20:28:48.044740 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-config-data\") pod \"ceilometer-0\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " pod="openstack/ceilometer-0" Mar 18 20:28:48 crc kubenswrapper[4950]: I0318 20:28:48.052007 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fg7pr\" (UniqueName: \"kubernetes.io/projected/00969ec6-eca3-4790-b69f-ba825f809823-kube-api-access-fg7pr\") pod \"ceilometer-0\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " pod="openstack/ceilometer-0" Mar 18 20:28:48 crc kubenswrapper[4950]: I0318 20:28:48.173208 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:28:48 crc kubenswrapper[4950]: I0318 20:28:48.637072 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:28:48 crc kubenswrapper[4950]: I0318 20:28:48.757062 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00969ec6-eca3-4790-b69f-ba825f809823","Type":"ContainerStarted","Data":"6291fa3b694a2d82bf767d88d1e5493818cb5eed30d9ccb6e9c0571df9e1aef2"} Mar 18 20:28:49 crc kubenswrapper[4950]: I0318 20:28:49.495186 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4" path="/var/lib/kubelet/pods/5b2e8fce-f0e3-4411-94fe-9e9bac3e22e4/volumes" Mar 18 20:28:49 crc kubenswrapper[4950]: I0318 20:28:49.769905 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00969ec6-eca3-4790-b69f-ba825f809823","Type":"ContainerStarted","Data":"8e823354f8410f2242ab0dd6d890021770466371f17c913ba99728c5a0073f32"} Mar 18 20:28:50 crc kubenswrapper[4950]: I0318 20:28:50.781112 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00969ec6-eca3-4790-b69f-ba825f809823","Type":"ContainerStarted","Data":"2ecc905e791972c42f485ef6d640aa017d161ec0f620d27f21a7090e6dbb0665"} Mar 18 20:28:50 crc kubenswrapper[4950]: I0318 20:28:50.781380 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00969ec6-eca3-4790-b69f-ba825f809823","Type":"ContainerStarted","Data":"50d2f66b6b7b82c882ff2011cb5fc6653db27f8ede3e1c9d09e985c9cab29fb9"} Mar 18 20:28:51 crc kubenswrapper[4950]: I0318 20:28:51.790130 4950 generic.go:334] "Generic (PLEG): container finished" podID="f58dabee-7195-40e9-952e-9dfeefc1ee08" containerID="656f6974b9820f48c5b34d3c8007059ae8bf8ad5ac44bf59d04c6ad0ee0e58f1" exitCode=0 Mar 18 20:28:51 crc kubenswrapper[4950]: I0318 20:28:51.790273 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lj42n" event={"ID":"f58dabee-7195-40e9-952e-9dfeefc1ee08","Type":"ContainerDied","Data":"656f6974b9820f48c5b34d3c8007059ae8bf8ad5ac44bf59d04c6ad0ee0e58f1"} Mar 18 20:28:51 crc kubenswrapper[4950]: I0318 20:28:51.999190 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 18 20:28:51 crc kubenswrapper[4950]: I0318 20:28:51.999250 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 18 20:28:52 crc kubenswrapper[4950]: I0318 20:28:52.802981 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00969ec6-eca3-4790-b69f-ba825f809823","Type":"ContainerStarted","Data":"45906ffaab429e682fdd942d7a9b2dd005d7cd0890f1a86ac3b2ce8cd586f984"} Mar 18 20:28:52 crc kubenswrapper[4950]: I0318 20:28:52.803388 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 18 20:28:52 crc kubenswrapper[4950]: I0318 20:28:52.835689 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.118742351 podStartE2EDuration="5.835668613s" podCreationTimestamp="2026-03-18 20:28:47 +0000 UTC" firstStartedPulling="2026-03-18 20:28:48.645901056 +0000 UTC m=+1341.886742934" lastFinishedPulling="2026-03-18 20:28:52.362827328 +0000 UTC m=+1345.603669196" observedRunningTime="2026-03-18 20:28:52.827535284 +0000 UTC m=+1346.068377172" watchObservedRunningTime="2026-03-18 20:28:52.835668613 +0000 UTC m=+1346.076510501" Mar 18 20:28:53 crc kubenswrapper[4950]: I0318 20:28:53.014528 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2fd4b49e-aca5-4160-9bb9-2cf5b4925022" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.189:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 18 20:28:53 crc kubenswrapper[4950]: I0318 20:28:53.014784 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2fd4b49e-aca5-4160-9bb9-2cf5b4925022" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.189:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 18 20:28:53 crc kubenswrapper[4950]: I0318 20:28:53.179913 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lj42n" Mar 18 20:28:53 crc kubenswrapper[4950]: I0318 20:28:53.257266 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f58dabee-7195-40e9-952e-9dfeefc1ee08-config-data\") pod \"f58dabee-7195-40e9-952e-9dfeefc1ee08\" (UID: \"f58dabee-7195-40e9-952e-9dfeefc1ee08\") " Mar 18 20:28:53 crc kubenswrapper[4950]: I0318 20:28:53.257590 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f58dabee-7195-40e9-952e-9dfeefc1ee08-combined-ca-bundle\") pod \"f58dabee-7195-40e9-952e-9dfeefc1ee08\" (UID: \"f58dabee-7195-40e9-952e-9dfeefc1ee08\") " Mar 18 20:28:53 crc kubenswrapper[4950]: I0318 20:28:53.257624 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grsk9\" (UniqueName: \"kubernetes.io/projected/f58dabee-7195-40e9-952e-9dfeefc1ee08-kube-api-access-grsk9\") pod \"f58dabee-7195-40e9-952e-9dfeefc1ee08\" (UID: \"f58dabee-7195-40e9-952e-9dfeefc1ee08\") " Mar 18 20:28:53 crc kubenswrapper[4950]: I0318 20:28:53.257660 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f58dabee-7195-40e9-952e-9dfeefc1ee08-scripts\") pod \"f58dabee-7195-40e9-952e-9dfeefc1ee08\" (UID: \"f58dabee-7195-40e9-952e-9dfeefc1ee08\") " Mar 18 20:28:53 crc kubenswrapper[4950]: I0318 20:28:53.262629 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f58dabee-7195-40e9-952e-9dfeefc1ee08-scripts" (OuterVolumeSpecName: "scripts") pod "f58dabee-7195-40e9-952e-9dfeefc1ee08" (UID: "f58dabee-7195-40e9-952e-9dfeefc1ee08"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:53 crc kubenswrapper[4950]: I0318 20:28:53.262997 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f58dabee-7195-40e9-952e-9dfeefc1ee08-kube-api-access-grsk9" (OuterVolumeSpecName: "kube-api-access-grsk9") pod "f58dabee-7195-40e9-952e-9dfeefc1ee08" (UID: "f58dabee-7195-40e9-952e-9dfeefc1ee08"). InnerVolumeSpecName "kube-api-access-grsk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:28:53 crc kubenswrapper[4950]: I0318 20:28:53.283306 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f58dabee-7195-40e9-952e-9dfeefc1ee08-config-data" (OuterVolumeSpecName: "config-data") pod "f58dabee-7195-40e9-952e-9dfeefc1ee08" (UID: "f58dabee-7195-40e9-952e-9dfeefc1ee08"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:53 crc kubenswrapper[4950]: I0318 20:28:53.285487 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f58dabee-7195-40e9-952e-9dfeefc1ee08-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f58dabee-7195-40e9-952e-9dfeefc1ee08" (UID: "f58dabee-7195-40e9-952e-9dfeefc1ee08"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:53 crc kubenswrapper[4950]: I0318 20:28:53.360219 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f58dabee-7195-40e9-952e-9dfeefc1ee08-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:53 crc kubenswrapper[4950]: I0318 20:28:53.360260 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f58dabee-7195-40e9-952e-9dfeefc1ee08-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:53 crc kubenswrapper[4950]: I0318 20:28:53.360273 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grsk9\" (UniqueName: \"kubernetes.io/projected/f58dabee-7195-40e9-952e-9dfeefc1ee08-kube-api-access-grsk9\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:53 crc kubenswrapper[4950]: I0318 20:28:53.360282 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f58dabee-7195-40e9-952e-9dfeefc1ee08-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:53 crc kubenswrapper[4950]: I0318 20:28:53.819702 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lj42n" event={"ID":"f58dabee-7195-40e9-952e-9dfeefc1ee08","Type":"ContainerDied","Data":"14448fcd21bdc40920382f8cee4a3af4a7e70f58e2e7e48885f443ec423346e9"} Mar 18 20:28:53 crc kubenswrapper[4950]: I0318 20:28:53.819923 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14448fcd21bdc40920382f8cee4a3af4a7e70f58e2e7e48885f443ec423346e9" Mar 18 20:28:53 crc kubenswrapper[4950]: I0318 20:28:53.820947 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lj42n" Mar 18 20:28:53 crc kubenswrapper[4950]: I0318 20:28:53.997796 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 18 20:28:53 crc kubenswrapper[4950]: I0318 20:28:53.998049 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2fd4b49e-aca5-4160-9bb9-2cf5b4925022" containerName="nova-api-log" containerID="cri-o://5660ff2159c127514406d6fa0a0e843212c22bba90a9f07c6e263a6e4d71c7c0" gracePeriod=30 Mar 18 20:28:53 crc kubenswrapper[4950]: I0318 20:28:53.998170 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2fd4b49e-aca5-4160-9bb9-2cf5b4925022" containerName="nova-api-api" containerID="cri-o://6c714f29806df1dfc267dcf912a6fa3c8048dd9245e022236aa98faf390b498c" gracePeriod=30 Mar 18 20:28:54 crc kubenswrapper[4950]: I0318 20:28:54.079797 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 18 20:28:54 crc kubenswrapper[4950]: I0318 20:28:54.080342 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ade0ed1d-2241-456b-b72f-3236fbc6ed45" containerName="nova-metadata-log" containerID="cri-o://fc3cba41084950eb31f480a81d4c3033b0a539a97624c7ff2e321821201872f9" gracePeriod=30 Mar 18 20:28:54 crc kubenswrapper[4950]: I0318 20:28:54.080417 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ade0ed1d-2241-456b-b72f-3236fbc6ed45" containerName="nova-metadata-metadata" containerID="cri-o://4b4d20fec20d82446b0a36dba6ec857459a48495cecba2ff3f21ce5a342e2289" gracePeriod=30 Mar 18 20:28:54 crc kubenswrapper[4950]: I0318 20:28:54.090273 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 18 20:28:54 crc kubenswrapper[4950]: I0318 20:28:54.090487 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="cbbcbf5a-8667-4e11-ad3e-f72b30084b6d" containerName="nova-scheduler-scheduler" containerID="cri-o://65e45e86d329a18fae51cb507d8d87a43866d7ee499a73bc9a986762b2bde176" gracePeriod=30 Mar 18 20:28:54 crc kubenswrapper[4950]: I0318 20:28:54.827897 4950 generic.go:334] "Generic (PLEG): container finished" podID="ade0ed1d-2241-456b-b72f-3236fbc6ed45" containerID="fc3cba41084950eb31f480a81d4c3033b0a539a97624c7ff2e321821201872f9" exitCode=143 Mar 18 20:28:54 crc kubenswrapper[4950]: I0318 20:28:54.827984 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ade0ed1d-2241-456b-b72f-3236fbc6ed45","Type":"ContainerDied","Data":"fc3cba41084950eb31f480a81d4c3033b0a539a97624c7ff2e321821201872f9"} Mar 18 20:28:54 crc kubenswrapper[4950]: I0318 20:28:54.830590 4950 generic.go:334] "Generic (PLEG): container finished" podID="2fd4b49e-aca5-4160-9bb9-2cf5b4925022" containerID="5660ff2159c127514406d6fa0a0e843212c22bba90a9f07c6e263a6e4d71c7c0" exitCode=143 Mar 18 20:28:54 crc kubenswrapper[4950]: I0318 20:28:54.830715 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2fd4b49e-aca5-4160-9bb9-2cf5b4925022","Type":"ContainerDied","Data":"5660ff2159c127514406d6fa0a0e843212c22bba90a9f07c6e263a6e4d71c7c0"} Mar 18 20:28:55 crc kubenswrapper[4950]: I0318 20:28:55.723020 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 18 20:28:55 crc kubenswrapper[4950]: I0318 20:28:55.802585 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbbcbf5a-8667-4e11-ad3e-f72b30084b6d-config-data\") pod \"cbbcbf5a-8667-4e11-ad3e-f72b30084b6d\" (UID: \"cbbcbf5a-8667-4e11-ad3e-f72b30084b6d\") " Mar 18 20:28:55 crc kubenswrapper[4950]: I0318 20:28:55.802635 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbbcbf5a-8667-4e11-ad3e-f72b30084b6d-combined-ca-bundle\") pod \"cbbcbf5a-8667-4e11-ad3e-f72b30084b6d\" (UID: \"cbbcbf5a-8667-4e11-ad3e-f72b30084b6d\") " Mar 18 20:28:55 crc kubenswrapper[4950]: I0318 20:28:55.802761 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvdtp\" (UniqueName: \"kubernetes.io/projected/cbbcbf5a-8667-4e11-ad3e-f72b30084b6d-kube-api-access-pvdtp\") pod \"cbbcbf5a-8667-4e11-ad3e-f72b30084b6d\" (UID: \"cbbcbf5a-8667-4e11-ad3e-f72b30084b6d\") " Mar 18 20:28:55 crc kubenswrapper[4950]: I0318 20:28:55.827650 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbbcbf5a-8667-4e11-ad3e-f72b30084b6d-kube-api-access-pvdtp" (OuterVolumeSpecName: "kube-api-access-pvdtp") pod "cbbcbf5a-8667-4e11-ad3e-f72b30084b6d" (UID: "cbbcbf5a-8667-4e11-ad3e-f72b30084b6d"). InnerVolumeSpecName "kube-api-access-pvdtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:28:55 crc kubenswrapper[4950]: I0318 20:28:55.831490 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbbcbf5a-8667-4e11-ad3e-f72b30084b6d-config-data" (OuterVolumeSpecName: "config-data") pod "cbbcbf5a-8667-4e11-ad3e-f72b30084b6d" (UID: "cbbcbf5a-8667-4e11-ad3e-f72b30084b6d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:55 crc kubenswrapper[4950]: I0318 20:28:55.840686 4950 generic.go:334] "Generic (PLEG): container finished" podID="cbbcbf5a-8667-4e11-ad3e-f72b30084b6d" containerID="65e45e86d329a18fae51cb507d8d87a43866d7ee499a73bc9a986762b2bde176" exitCode=0 Mar 18 20:28:55 crc kubenswrapper[4950]: I0318 20:28:55.840737 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cbbcbf5a-8667-4e11-ad3e-f72b30084b6d","Type":"ContainerDied","Data":"65e45e86d329a18fae51cb507d8d87a43866d7ee499a73bc9a986762b2bde176"} Mar 18 20:28:55 crc kubenswrapper[4950]: I0318 20:28:55.840761 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cbbcbf5a-8667-4e11-ad3e-f72b30084b6d","Type":"ContainerDied","Data":"a97374e6c98d15349a9157c23184b6dfb090ea79ea316c76e6876b9e85dc1200"} Mar 18 20:28:55 crc kubenswrapper[4950]: I0318 20:28:55.840779 4950 scope.go:117] "RemoveContainer" containerID="65e45e86d329a18fae51cb507d8d87a43866d7ee499a73bc9a986762b2bde176" Mar 18 20:28:55 crc kubenswrapper[4950]: I0318 20:28:55.840889 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 18 20:28:55 crc kubenswrapper[4950]: I0318 20:28:55.847854 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbbcbf5a-8667-4e11-ad3e-f72b30084b6d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cbbcbf5a-8667-4e11-ad3e-f72b30084b6d" (UID: "cbbcbf5a-8667-4e11-ad3e-f72b30084b6d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:55 crc kubenswrapper[4950]: I0318 20:28:55.896396 4950 scope.go:117] "RemoveContainer" containerID="65e45e86d329a18fae51cb507d8d87a43866d7ee499a73bc9a986762b2bde176" Mar 18 20:28:55 crc kubenswrapper[4950]: E0318 20:28:55.900816 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65e45e86d329a18fae51cb507d8d87a43866d7ee499a73bc9a986762b2bde176\": container with ID starting with 65e45e86d329a18fae51cb507d8d87a43866d7ee499a73bc9a986762b2bde176 not found: ID does not exist" containerID="65e45e86d329a18fae51cb507d8d87a43866d7ee499a73bc9a986762b2bde176" Mar 18 20:28:55 crc kubenswrapper[4950]: I0318 20:28:55.900868 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65e45e86d329a18fae51cb507d8d87a43866d7ee499a73bc9a986762b2bde176"} err="failed to get container status \"65e45e86d329a18fae51cb507d8d87a43866d7ee499a73bc9a986762b2bde176\": rpc error: code = NotFound desc = could not find container \"65e45e86d329a18fae51cb507d8d87a43866d7ee499a73bc9a986762b2bde176\": container with ID starting with 65e45e86d329a18fae51cb507d8d87a43866d7ee499a73bc9a986762b2bde176 not found: ID does not exist" Mar 18 20:28:55 crc kubenswrapper[4950]: I0318 20:28:55.904495 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbbcbf5a-8667-4e11-ad3e-f72b30084b6d-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:55 crc kubenswrapper[4950]: I0318 20:28:55.904524 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbbcbf5a-8667-4e11-ad3e-f72b30084b6d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:55 crc kubenswrapper[4950]: I0318 20:28:55.904540 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvdtp\" (UniqueName: \"kubernetes.io/projected/cbbcbf5a-8667-4e11-ad3e-f72b30084b6d-kube-api-access-pvdtp\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:56 crc kubenswrapper[4950]: I0318 20:28:56.193854 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 18 20:28:56 crc kubenswrapper[4950]: I0318 20:28:56.205805 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 18 20:28:56 crc kubenswrapper[4950]: I0318 20:28:56.224853 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 18 20:28:56 crc kubenswrapper[4950]: E0318 20:28:56.225700 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbbcbf5a-8667-4e11-ad3e-f72b30084b6d" containerName="nova-scheduler-scheduler" Mar 18 20:28:56 crc kubenswrapper[4950]: I0318 20:28:56.226872 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbbcbf5a-8667-4e11-ad3e-f72b30084b6d" containerName="nova-scheduler-scheduler" Mar 18 20:28:56 crc kubenswrapper[4950]: E0318 20:28:56.226987 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f58dabee-7195-40e9-952e-9dfeefc1ee08" containerName="nova-manage" Mar 18 20:28:56 crc kubenswrapper[4950]: I0318 20:28:56.227049 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="f58dabee-7195-40e9-952e-9dfeefc1ee08" containerName="nova-manage" Mar 18 20:28:56 crc kubenswrapper[4950]: I0318 20:28:56.227276 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="f58dabee-7195-40e9-952e-9dfeefc1ee08" containerName="nova-manage" Mar 18 20:28:56 crc kubenswrapper[4950]: I0318 20:28:56.227351 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbbcbf5a-8667-4e11-ad3e-f72b30084b6d" containerName="nova-scheduler-scheduler" Mar 18 20:28:56 crc kubenswrapper[4950]: I0318 20:28:56.228105 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 18 20:28:56 crc kubenswrapper[4950]: I0318 20:28:56.236004 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 18 20:28:56 crc kubenswrapper[4950]: I0318 20:28:56.250813 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 18 20:28:56 crc kubenswrapper[4950]: I0318 20:28:56.312991 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrfd2\" (UniqueName: \"kubernetes.io/projected/1e49132d-6e98-4270-8ba0-7500e8a84bfd-kube-api-access-vrfd2\") pod \"nova-scheduler-0\" (UID: \"1e49132d-6e98-4270-8ba0-7500e8a84bfd\") " pod="openstack/nova-scheduler-0" Mar 18 20:28:56 crc kubenswrapper[4950]: I0318 20:28:56.313156 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e49132d-6e98-4270-8ba0-7500e8a84bfd-config-data\") pod \"nova-scheduler-0\" (UID: \"1e49132d-6e98-4270-8ba0-7500e8a84bfd\") " pod="openstack/nova-scheduler-0" Mar 18 20:28:56 crc kubenswrapper[4950]: I0318 20:28:56.313251 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e49132d-6e98-4270-8ba0-7500e8a84bfd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1e49132d-6e98-4270-8ba0-7500e8a84bfd\") " pod="openstack/nova-scheduler-0" Mar 18 20:28:56 crc kubenswrapper[4950]: I0318 20:28:56.414988 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e49132d-6e98-4270-8ba0-7500e8a84bfd-config-data\") pod \"nova-scheduler-0\" (UID: \"1e49132d-6e98-4270-8ba0-7500e8a84bfd\") " pod="openstack/nova-scheduler-0" Mar 18 20:28:56 crc kubenswrapper[4950]: I0318 20:28:56.415071 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e49132d-6e98-4270-8ba0-7500e8a84bfd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1e49132d-6e98-4270-8ba0-7500e8a84bfd\") " pod="openstack/nova-scheduler-0" Mar 18 20:28:56 crc kubenswrapper[4950]: I0318 20:28:56.415230 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrfd2\" (UniqueName: \"kubernetes.io/projected/1e49132d-6e98-4270-8ba0-7500e8a84bfd-kube-api-access-vrfd2\") pod \"nova-scheduler-0\" (UID: \"1e49132d-6e98-4270-8ba0-7500e8a84bfd\") " pod="openstack/nova-scheduler-0" Mar 18 20:28:56 crc kubenswrapper[4950]: I0318 20:28:56.429705 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e49132d-6e98-4270-8ba0-7500e8a84bfd-config-data\") pod \"nova-scheduler-0\" (UID: \"1e49132d-6e98-4270-8ba0-7500e8a84bfd\") " pod="openstack/nova-scheduler-0" Mar 18 20:28:56 crc kubenswrapper[4950]: I0318 20:28:56.429946 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e49132d-6e98-4270-8ba0-7500e8a84bfd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1e49132d-6e98-4270-8ba0-7500e8a84bfd\") " pod="openstack/nova-scheduler-0" Mar 18 20:28:56 crc kubenswrapper[4950]: I0318 20:28:56.445286 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrfd2\" (UniqueName: \"kubernetes.io/projected/1e49132d-6e98-4270-8ba0-7500e8a84bfd-kube-api-access-vrfd2\") pod \"nova-scheduler-0\" (UID: \"1e49132d-6e98-4270-8ba0-7500e8a84bfd\") " pod="openstack/nova-scheduler-0" Mar 18 20:28:56 crc kubenswrapper[4950]: I0318 20:28:56.550912 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 18 20:28:56 crc kubenswrapper[4950]: I0318 20:28:56.965510 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.504898 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbbcbf5a-8667-4e11-ad3e-f72b30084b6d" path="/var/lib/kubelet/pods/cbbcbf5a-8667-4e11-ad3e-f72b30084b6d/volumes" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.653587 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.738854 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ade0ed1d-2241-456b-b72f-3236fbc6ed45-combined-ca-bundle\") pod \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\" (UID: \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\") " Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.739249 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ade0ed1d-2241-456b-b72f-3236fbc6ed45-config-data\") pod \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\" (UID: \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\") " Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.739337 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ade0ed1d-2241-456b-b72f-3236fbc6ed45-nova-metadata-tls-certs\") pod \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\" (UID: \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\") " Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.739371 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8trkg\" (UniqueName: \"kubernetes.io/projected/ade0ed1d-2241-456b-b72f-3236fbc6ed45-kube-api-access-8trkg\") pod \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\" (UID: \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\") " Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.740242 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ade0ed1d-2241-456b-b72f-3236fbc6ed45-logs\") pod \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\" (UID: \"ade0ed1d-2241-456b-b72f-3236fbc6ed45\") " Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.740670 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ade0ed1d-2241-456b-b72f-3236fbc6ed45-logs" (OuterVolumeSpecName: "logs") pod "ade0ed1d-2241-456b-b72f-3236fbc6ed45" (UID: "ade0ed1d-2241-456b-b72f-3236fbc6ed45"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.740928 4950 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ade0ed1d-2241-456b-b72f-3236fbc6ed45-logs\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.750927 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ade0ed1d-2241-456b-b72f-3236fbc6ed45-kube-api-access-8trkg" (OuterVolumeSpecName: "kube-api-access-8trkg") pod "ade0ed1d-2241-456b-b72f-3236fbc6ed45" (UID: "ade0ed1d-2241-456b-b72f-3236fbc6ed45"). InnerVolumeSpecName "kube-api-access-8trkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.806117 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ade0ed1d-2241-456b-b72f-3236fbc6ed45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ade0ed1d-2241-456b-b72f-3236fbc6ed45" (UID: "ade0ed1d-2241-456b-b72f-3236fbc6ed45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.807670 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ade0ed1d-2241-456b-b72f-3236fbc6ed45-config-data" (OuterVolumeSpecName: "config-data") pod "ade0ed1d-2241-456b-b72f-3236fbc6ed45" (UID: "ade0ed1d-2241-456b-b72f-3236fbc6ed45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.808732 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ade0ed1d-2241-456b-b72f-3236fbc6ed45-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "ade0ed1d-2241-456b-b72f-3236fbc6ed45" (UID: "ade0ed1d-2241-456b-b72f-3236fbc6ed45"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.842302 4950 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ade0ed1d-2241-456b-b72f-3236fbc6ed45-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.842329 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8trkg\" (UniqueName: \"kubernetes.io/projected/ade0ed1d-2241-456b-b72f-3236fbc6ed45-kube-api-access-8trkg\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.842340 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ade0ed1d-2241-456b-b72f-3236fbc6ed45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.842349 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ade0ed1d-2241-456b-b72f-3236fbc6ed45-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.869301 4950 generic.go:334] "Generic (PLEG): container finished" podID="ade0ed1d-2241-456b-b72f-3236fbc6ed45" containerID="4b4d20fec20d82446b0a36dba6ec857459a48495cecba2ff3f21ce5a342e2289" exitCode=0 Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.869365 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ade0ed1d-2241-456b-b72f-3236fbc6ed45","Type":"ContainerDied","Data":"4b4d20fec20d82446b0a36dba6ec857459a48495cecba2ff3f21ce5a342e2289"} Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.869392 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ade0ed1d-2241-456b-b72f-3236fbc6ed45","Type":"ContainerDied","Data":"5820e4b400c1e3437a059bb369b1ba4bd00d87dc642f281e167d90a04ede89a0"} Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.869425 4950 scope.go:117] "RemoveContainer" containerID="4b4d20fec20d82446b0a36dba6ec857459a48495cecba2ff3f21ce5a342e2289" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.869536 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.880829 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1e49132d-6e98-4270-8ba0-7500e8a84bfd","Type":"ContainerStarted","Data":"686aefc98fe5e995236c9e9081d35d2327bfed7197609227e579d482f1434a4a"} Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.880877 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1e49132d-6e98-4270-8ba0-7500e8a84bfd","Type":"ContainerStarted","Data":"f8b845152993dd3965f236dee7fafa4bef517d0b839cf2413ccfcd406364e747"} Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.902038 4950 scope.go:117] "RemoveContainer" containerID="fc3cba41084950eb31f480a81d4c3033b0a539a97624c7ff2e321821201872f9" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.902192 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.902169504 podStartE2EDuration="1.902169504s" podCreationTimestamp="2026-03-18 20:28:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:28:57.896086725 +0000 UTC m=+1351.136928603" watchObservedRunningTime="2026-03-18 20:28:57.902169504 +0000 UTC m=+1351.143011372" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.926827 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.927223 4950 scope.go:117] "RemoveContainer" containerID="4b4d20fec20d82446b0a36dba6ec857459a48495cecba2ff3f21ce5a342e2289" Mar 18 20:28:57 crc kubenswrapper[4950]: E0318 20:28:57.928025 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b4d20fec20d82446b0a36dba6ec857459a48495cecba2ff3f21ce5a342e2289\": container with ID starting with 4b4d20fec20d82446b0a36dba6ec857459a48495cecba2ff3f21ce5a342e2289 not found: ID does not exist" containerID="4b4d20fec20d82446b0a36dba6ec857459a48495cecba2ff3f21ce5a342e2289" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.928067 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b4d20fec20d82446b0a36dba6ec857459a48495cecba2ff3f21ce5a342e2289"} err="failed to get container status \"4b4d20fec20d82446b0a36dba6ec857459a48495cecba2ff3f21ce5a342e2289\": rpc error: code = NotFound desc = could not find container \"4b4d20fec20d82446b0a36dba6ec857459a48495cecba2ff3f21ce5a342e2289\": container with ID starting with 4b4d20fec20d82446b0a36dba6ec857459a48495cecba2ff3f21ce5a342e2289 not found: ID does not exist" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.928101 4950 scope.go:117] "RemoveContainer" containerID="fc3cba41084950eb31f480a81d4c3033b0a539a97624c7ff2e321821201872f9" Mar 18 20:28:57 crc kubenswrapper[4950]: E0318 20:28:57.929944 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc3cba41084950eb31f480a81d4c3033b0a539a97624c7ff2e321821201872f9\": container with ID starting with fc3cba41084950eb31f480a81d4c3033b0a539a97624c7ff2e321821201872f9 not found: ID does not exist" containerID="fc3cba41084950eb31f480a81d4c3033b0a539a97624c7ff2e321821201872f9" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.929981 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc3cba41084950eb31f480a81d4c3033b0a539a97624c7ff2e321821201872f9"} err="failed to get container status \"fc3cba41084950eb31f480a81d4c3033b0a539a97624c7ff2e321821201872f9\": rpc error: code = NotFound desc = could not find container \"fc3cba41084950eb31f480a81d4c3033b0a539a97624c7ff2e321821201872f9\": container with ID starting with fc3cba41084950eb31f480a81d4c3033b0a539a97624c7ff2e321821201872f9 not found: ID does not exist" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.940447 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.955549 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 18 20:28:57 crc kubenswrapper[4950]: E0318 20:28:57.956049 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ade0ed1d-2241-456b-b72f-3236fbc6ed45" containerName="nova-metadata-metadata" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.956060 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="ade0ed1d-2241-456b-b72f-3236fbc6ed45" containerName="nova-metadata-metadata" Mar 18 20:28:57 crc kubenswrapper[4950]: E0318 20:28:57.956082 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ade0ed1d-2241-456b-b72f-3236fbc6ed45" containerName="nova-metadata-log" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.956090 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="ade0ed1d-2241-456b-b72f-3236fbc6ed45" containerName="nova-metadata-log" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.956250 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="ade0ed1d-2241-456b-b72f-3236fbc6ed45" containerName="nova-metadata-metadata" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.956265 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="ade0ed1d-2241-456b-b72f-3236fbc6ed45" containerName="nova-metadata-log" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.957018 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.957088 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.981156 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 18 20:28:57 crc kubenswrapper[4950]: I0318 20:28:57.981358 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.045459 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f05a94b0-defc-4c6c-af5c-3ce8346261a3-config-data\") pod \"nova-metadata-0\" (UID: \"f05a94b0-defc-4c6c-af5c-3ce8346261a3\") " pod="openstack/nova-metadata-0" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.045505 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czxbb\" (UniqueName: \"kubernetes.io/projected/f05a94b0-defc-4c6c-af5c-3ce8346261a3-kube-api-access-czxbb\") pod \"nova-metadata-0\" (UID: \"f05a94b0-defc-4c6c-af5c-3ce8346261a3\") " pod="openstack/nova-metadata-0" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.045800 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f05a94b0-defc-4c6c-af5c-3ce8346261a3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f05a94b0-defc-4c6c-af5c-3ce8346261a3\") " pod="openstack/nova-metadata-0" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.045907 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f05a94b0-defc-4c6c-af5c-3ce8346261a3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f05a94b0-defc-4c6c-af5c-3ce8346261a3\") " pod="openstack/nova-metadata-0" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.046023 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f05a94b0-defc-4c6c-af5c-3ce8346261a3-logs\") pod \"nova-metadata-0\" (UID: \"f05a94b0-defc-4c6c-af5c-3ce8346261a3\") " pod="openstack/nova-metadata-0" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.148030 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f05a94b0-defc-4c6c-af5c-3ce8346261a3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f05a94b0-defc-4c6c-af5c-3ce8346261a3\") " pod="openstack/nova-metadata-0" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.148773 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f05a94b0-defc-4c6c-af5c-3ce8346261a3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f05a94b0-defc-4c6c-af5c-3ce8346261a3\") " pod="openstack/nova-metadata-0" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.148949 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f05a94b0-defc-4c6c-af5c-3ce8346261a3-logs\") pod \"nova-metadata-0\" (UID: \"f05a94b0-defc-4c6c-af5c-3ce8346261a3\") " pod="openstack/nova-metadata-0" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.149103 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f05a94b0-defc-4c6c-af5c-3ce8346261a3-config-data\") pod \"nova-metadata-0\" (UID: \"f05a94b0-defc-4c6c-af5c-3ce8346261a3\") " pod="openstack/nova-metadata-0" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.149493 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czxbb\" (UniqueName: \"kubernetes.io/projected/f05a94b0-defc-4c6c-af5c-3ce8346261a3-kube-api-access-czxbb\") pod \"nova-metadata-0\" (UID: \"f05a94b0-defc-4c6c-af5c-3ce8346261a3\") " pod="openstack/nova-metadata-0" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.149574 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f05a94b0-defc-4c6c-af5c-3ce8346261a3-logs\") pod \"nova-metadata-0\" (UID: \"f05a94b0-defc-4c6c-af5c-3ce8346261a3\") " pod="openstack/nova-metadata-0" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.151858 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f05a94b0-defc-4c6c-af5c-3ce8346261a3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f05a94b0-defc-4c6c-af5c-3ce8346261a3\") " pod="openstack/nova-metadata-0" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.152349 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f05a94b0-defc-4c6c-af5c-3ce8346261a3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f05a94b0-defc-4c6c-af5c-3ce8346261a3\") " pod="openstack/nova-metadata-0" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.160130 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f05a94b0-defc-4c6c-af5c-3ce8346261a3-config-data\") pod \"nova-metadata-0\" (UID: \"f05a94b0-defc-4c6c-af5c-3ce8346261a3\") " pod="openstack/nova-metadata-0" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.164696 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czxbb\" (UniqueName: \"kubernetes.io/projected/f05a94b0-defc-4c6c-af5c-3ce8346261a3-kube-api-access-czxbb\") pod \"nova-metadata-0\" (UID: \"f05a94b0-defc-4c6c-af5c-3ce8346261a3\") " pod="openstack/nova-metadata-0" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.299062 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.794029 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.888613 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.892341 4950 generic.go:334] "Generic (PLEG): container finished" podID="2fd4b49e-aca5-4160-9bb9-2cf5b4925022" containerID="6c714f29806df1dfc267dcf912a6fa3c8048dd9245e022236aa98faf390b498c" exitCode=0 Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.892372 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.892437 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2fd4b49e-aca5-4160-9bb9-2cf5b4925022","Type":"ContainerDied","Data":"6c714f29806df1dfc267dcf912a6fa3c8048dd9245e022236aa98faf390b498c"} Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.892463 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2fd4b49e-aca5-4160-9bb9-2cf5b4925022","Type":"ContainerDied","Data":"0399f469bfa4a2b6ba7e9a03d720fac5c36881d1a9b56c3aa70bbf8f31cc0e1f"} Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.892501 4950 scope.go:117] "RemoveContainer" containerID="6c714f29806df1dfc267dcf912a6fa3c8048dd9245e022236aa98faf390b498c" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.896459 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f05a94b0-defc-4c6c-af5c-3ce8346261a3","Type":"ContainerStarted","Data":"004f7445044baca7d3d1c0b1c7115cb4ba4f3ce43efdcabfa8b4000dc0e3b26f"} Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.934701 4950 scope.go:117] "RemoveContainer" containerID="5660ff2159c127514406d6fa0a0e843212c22bba90a9f07c6e263a6e4d71c7c0" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.975240 4950 scope.go:117] "RemoveContainer" containerID="6c714f29806df1dfc267dcf912a6fa3c8048dd9245e022236aa98faf390b498c" Mar 18 20:28:58 crc kubenswrapper[4950]: E0318 20:28:58.975839 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c714f29806df1dfc267dcf912a6fa3c8048dd9245e022236aa98faf390b498c\": container with ID starting with 6c714f29806df1dfc267dcf912a6fa3c8048dd9245e022236aa98faf390b498c not found: ID does not exist" containerID="6c714f29806df1dfc267dcf912a6fa3c8048dd9245e022236aa98faf390b498c" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.975878 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c714f29806df1dfc267dcf912a6fa3c8048dd9245e022236aa98faf390b498c"} err="failed to get container status \"6c714f29806df1dfc267dcf912a6fa3c8048dd9245e022236aa98faf390b498c\": rpc error: code = NotFound desc = could not find container \"6c714f29806df1dfc267dcf912a6fa3c8048dd9245e022236aa98faf390b498c\": container with ID starting with 6c714f29806df1dfc267dcf912a6fa3c8048dd9245e022236aa98faf390b498c not found: ID does not exist" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.975904 4950 scope.go:117] "RemoveContainer" containerID="5660ff2159c127514406d6fa0a0e843212c22bba90a9f07c6e263a6e4d71c7c0" Mar 18 20:28:58 crc kubenswrapper[4950]: E0318 20:28:58.976369 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5660ff2159c127514406d6fa0a0e843212c22bba90a9f07c6e263a6e4d71c7c0\": container with ID starting with 5660ff2159c127514406d6fa0a0e843212c22bba90a9f07c6e263a6e4d71c7c0 not found: ID does not exist" containerID="5660ff2159c127514406d6fa0a0e843212c22bba90a9f07c6e263a6e4d71c7c0" Mar 18 20:28:58 crc kubenswrapper[4950]: I0318 20:28:58.976480 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5660ff2159c127514406d6fa0a0e843212c22bba90a9f07c6e263a6e4d71c7c0"} err="failed to get container status \"5660ff2159c127514406d6fa0a0e843212c22bba90a9f07c6e263a6e4d71c7c0\": rpc error: code = NotFound desc = could not find container \"5660ff2159c127514406d6fa0a0e843212c22bba90a9f07c6e263a6e4d71c7c0\": container with ID starting with 5660ff2159c127514406d6fa0a0e843212c22bba90a9f07c6e263a6e4d71c7c0 not found: ID does not exist" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.066403 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-internal-tls-certs\") pod \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.066606 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-config-data\") pod \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.066691 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-combined-ca-bundle\") pod \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.066788 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5vqj\" (UniqueName: \"kubernetes.io/projected/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-kube-api-access-t5vqj\") pod \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.066867 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-logs\") pod \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.066951 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-public-tls-certs\") pod \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\" (UID: \"2fd4b49e-aca5-4160-9bb9-2cf5b4925022\") " Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.069567 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-logs" (OuterVolumeSpecName: "logs") pod "2fd4b49e-aca5-4160-9bb9-2cf5b4925022" (UID: "2fd4b49e-aca5-4160-9bb9-2cf5b4925022"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.073358 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-kube-api-access-t5vqj" (OuterVolumeSpecName: "kube-api-access-t5vqj") pod "2fd4b49e-aca5-4160-9bb9-2cf5b4925022" (UID: "2fd4b49e-aca5-4160-9bb9-2cf5b4925022"). InnerVolumeSpecName "kube-api-access-t5vqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.102348 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-config-data" (OuterVolumeSpecName: "config-data") pod "2fd4b49e-aca5-4160-9bb9-2cf5b4925022" (UID: "2fd4b49e-aca5-4160-9bb9-2cf5b4925022"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.106743 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2fd4b49e-aca5-4160-9bb9-2cf5b4925022" (UID: "2fd4b49e-aca5-4160-9bb9-2cf5b4925022"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.128849 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2fd4b49e-aca5-4160-9bb9-2cf5b4925022" (UID: "2fd4b49e-aca5-4160-9bb9-2cf5b4925022"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.137326 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2fd4b49e-aca5-4160-9bb9-2cf5b4925022" (UID: "2fd4b49e-aca5-4160-9bb9-2cf5b4925022"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.178651 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.179042 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.179066 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5vqj\" (UniqueName: \"kubernetes.io/projected/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-kube-api-access-t5vqj\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.179084 4950 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-logs\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.179099 4950 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.179113 4950 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fd4b49e-aca5-4160-9bb9-2cf5b4925022-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.227782 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.235763 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.252497 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 18 20:28:59 crc kubenswrapper[4950]: E0318 20:28:59.252900 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fd4b49e-aca5-4160-9bb9-2cf5b4925022" containerName="nova-api-api" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.252926 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fd4b49e-aca5-4160-9bb9-2cf5b4925022" containerName="nova-api-api" Mar 18 20:28:59 crc kubenswrapper[4950]: E0318 20:28:59.252961 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fd4b49e-aca5-4160-9bb9-2cf5b4925022" containerName="nova-api-log" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.252972 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fd4b49e-aca5-4160-9bb9-2cf5b4925022" containerName="nova-api-log" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.253189 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fd4b49e-aca5-4160-9bb9-2cf5b4925022" containerName="nova-api-log" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.253219 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fd4b49e-aca5-4160-9bb9-2cf5b4925022" containerName="nova-api-api" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.254329 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.257590 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.257797 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.257993 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.266038 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.399700 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/603bbfd0-1fdf-44f7-bc34-ce9111ee204f-config-data\") pod \"nova-api-0\" (UID: \"603bbfd0-1fdf-44f7-bc34-ce9111ee204f\") " pod="openstack/nova-api-0" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.399787 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4s6j\" (UniqueName: \"kubernetes.io/projected/603bbfd0-1fdf-44f7-bc34-ce9111ee204f-kube-api-access-b4s6j\") pod \"nova-api-0\" (UID: \"603bbfd0-1fdf-44f7-bc34-ce9111ee204f\") " pod="openstack/nova-api-0" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.399807 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/603bbfd0-1fdf-44f7-bc34-ce9111ee204f-logs\") pod \"nova-api-0\" (UID: \"603bbfd0-1fdf-44f7-bc34-ce9111ee204f\") " pod="openstack/nova-api-0" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.399823 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/603bbfd0-1fdf-44f7-bc34-ce9111ee204f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"603bbfd0-1fdf-44f7-bc34-ce9111ee204f\") " pod="openstack/nova-api-0" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.399849 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/603bbfd0-1fdf-44f7-bc34-ce9111ee204f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"603bbfd0-1fdf-44f7-bc34-ce9111ee204f\") " pod="openstack/nova-api-0" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.399881 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/603bbfd0-1fdf-44f7-bc34-ce9111ee204f-public-tls-certs\") pod \"nova-api-0\" (UID: \"603bbfd0-1fdf-44f7-bc34-ce9111ee204f\") " pod="openstack/nova-api-0" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.488262 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fd4b49e-aca5-4160-9bb9-2cf5b4925022" path="/var/lib/kubelet/pods/2fd4b49e-aca5-4160-9bb9-2cf5b4925022/volumes" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.489014 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ade0ed1d-2241-456b-b72f-3236fbc6ed45" path="/var/lib/kubelet/pods/ade0ed1d-2241-456b-b72f-3236fbc6ed45/volumes" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.501903 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4s6j\" (UniqueName: \"kubernetes.io/projected/603bbfd0-1fdf-44f7-bc34-ce9111ee204f-kube-api-access-b4s6j\") pod \"nova-api-0\" (UID: \"603bbfd0-1fdf-44f7-bc34-ce9111ee204f\") " pod="openstack/nova-api-0" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.501935 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/603bbfd0-1fdf-44f7-bc34-ce9111ee204f-logs\") pod \"nova-api-0\" (UID: \"603bbfd0-1fdf-44f7-bc34-ce9111ee204f\") " pod="openstack/nova-api-0" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.501953 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/603bbfd0-1fdf-44f7-bc34-ce9111ee204f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"603bbfd0-1fdf-44f7-bc34-ce9111ee204f\") " pod="openstack/nova-api-0" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.501976 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/603bbfd0-1fdf-44f7-bc34-ce9111ee204f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"603bbfd0-1fdf-44f7-bc34-ce9111ee204f\") " pod="openstack/nova-api-0" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.502014 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/603bbfd0-1fdf-44f7-bc34-ce9111ee204f-public-tls-certs\") pod \"nova-api-0\" (UID: \"603bbfd0-1fdf-44f7-bc34-ce9111ee204f\") " pod="openstack/nova-api-0" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.502080 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/603bbfd0-1fdf-44f7-bc34-ce9111ee204f-config-data\") pod \"nova-api-0\" (UID: \"603bbfd0-1fdf-44f7-bc34-ce9111ee204f\") " pod="openstack/nova-api-0" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.502964 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/603bbfd0-1fdf-44f7-bc34-ce9111ee204f-logs\") pod \"nova-api-0\" (UID: \"603bbfd0-1fdf-44f7-bc34-ce9111ee204f\") " pod="openstack/nova-api-0" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.506508 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/603bbfd0-1fdf-44f7-bc34-ce9111ee204f-config-data\") pod \"nova-api-0\" (UID: \"603bbfd0-1fdf-44f7-bc34-ce9111ee204f\") " pod="openstack/nova-api-0" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.506859 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/603bbfd0-1fdf-44f7-bc34-ce9111ee204f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"603bbfd0-1fdf-44f7-bc34-ce9111ee204f\") " pod="openstack/nova-api-0" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.506957 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/603bbfd0-1fdf-44f7-bc34-ce9111ee204f-public-tls-certs\") pod \"nova-api-0\" (UID: \"603bbfd0-1fdf-44f7-bc34-ce9111ee204f\") " pod="openstack/nova-api-0" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.507564 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/603bbfd0-1fdf-44f7-bc34-ce9111ee204f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"603bbfd0-1fdf-44f7-bc34-ce9111ee204f\") " pod="openstack/nova-api-0" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.524638 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4s6j\" (UniqueName: \"kubernetes.io/projected/603bbfd0-1fdf-44f7-bc34-ce9111ee204f-kube-api-access-b4s6j\") pod \"nova-api-0\" (UID: \"603bbfd0-1fdf-44f7-bc34-ce9111ee204f\") " pod="openstack/nova-api-0" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.660370 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.910133 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f05a94b0-defc-4c6c-af5c-3ce8346261a3","Type":"ContainerStarted","Data":"b35547d212c0b46ba3183ad047c805e3898ee6415c1373eef59164a8eb1ed6a1"} Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.910574 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f05a94b0-defc-4c6c-af5c-3ce8346261a3","Type":"ContainerStarted","Data":"cc69d4b087289d0077dd320557577b75ac6cd43df8c2fb01e7564a214be3e160"} Mar 18 20:28:59 crc kubenswrapper[4950]: I0318 20:28:59.945746 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.94572532 podStartE2EDuration="2.94572532s" podCreationTimestamp="2026-03-18 20:28:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:28:59.928196921 +0000 UTC m=+1353.169038799" watchObservedRunningTime="2026-03-18 20:28:59.94572532 +0000 UTC m=+1353.186567198" Mar 18 20:29:00 crc kubenswrapper[4950]: W0318 20:29:00.138110 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod603bbfd0_1fdf_44f7_bc34_ce9111ee204f.slice/crio-48ef9c4b47c1849839f5843e1a2fecc25e9c9e7f0a624b08654a007452dd4a96 WatchSource:0}: Error finding container 48ef9c4b47c1849839f5843e1a2fecc25e9c9e7f0a624b08654a007452dd4a96: Status 404 returned error can't find the container with id 48ef9c4b47c1849839f5843e1a2fecc25e9c9e7f0a624b08654a007452dd4a96 Mar 18 20:29:00 crc kubenswrapper[4950]: I0318 20:29:00.143288 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 18 20:29:00 crc kubenswrapper[4950]: I0318 20:29:00.923484 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"603bbfd0-1fdf-44f7-bc34-ce9111ee204f","Type":"ContainerStarted","Data":"ec6f083854b31055a1bd9d3a0eb7d1ae145a483ec5d807f2316fb9ba1d4b4f48"} Mar 18 20:29:00 crc kubenswrapper[4950]: I0318 20:29:00.923871 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"603bbfd0-1fdf-44f7-bc34-ce9111ee204f","Type":"ContainerStarted","Data":"6b60dec8a837ee908af3b0d98c9e9ba10819d137db77afb37e7ac897b19768b7"} Mar 18 20:29:00 crc kubenswrapper[4950]: I0318 20:29:00.923894 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"603bbfd0-1fdf-44f7-bc34-ce9111ee204f","Type":"ContainerStarted","Data":"48ef9c4b47c1849839f5843e1a2fecc25e9c9e7f0a624b08654a007452dd4a96"} Mar 18 20:29:00 crc kubenswrapper[4950]: I0318 20:29:00.963972 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.963941645 podStartE2EDuration="1.963941645s" podCreationTimestamp="2026-03-18 20:28:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:29:00.959339184 +0000 UTC m=+1354.200181062" watchObservedRunningTime="2026-03-18 20:29:00.963941645 +0000 UTC m=+1354.204783543" Mar 18 20:29:01 crc kubenswrapper[4950]: I0318 20:29:01.552395 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 18 20:29:03 crc kubenswrapper[4950]: I0318 20:29:03.836652 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:29:03 crc kubenswrapper[4950]: I0318 20:29:03.837163 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:29:03 crc kubenswrapper[4950]: I0318 20:29:03.837234 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:29:03 crc kubenswrapper[4950]: I0318 20:29:03.838452 4950 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4e352dfc9da032a31ef3fccebe090411fdd509a1d2eb6df9588325dd52adecda"} pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 20:29:03 crc kubenswrapper[4950]: I0318 20:29:03.838579 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" containerID="cri-o://4e352dfc9da032a31ef3fccebe090411fdd509a1d2eb6df9588325dd52adecda" gracePeriod=600 Mar 18 20:29:04 crc kubenswrapper[4950]: I0318 20:29:04.968697 4950 generic.go:334] "Generic (PLEG): container finished" podID="4048b439-3266-46e7-9de0-22377efacc46" containerID="4e352dfc9da032a31ef3fccebe090411fdd509a1d2eb6df9588325dd52adecda" exitCode=0 Mar 18 20:29:04 crc kubenswrapper[4950]: I0318 20:29:04.968752 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerDied","Data":"4e352dfc9da032a31ef3fccebe090411fdd509a1d2eb6df9588325dd52adecda"} Mar 18 20:29:04 crc kubenswrapper[4950]: I0318 20:29:04.969068 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerStarted","Data":"02a13f19effbd6df78e6bd43d75fe3a70d3a67fc27d7dba640df7ed8a4afc83f"} Mar 18 20:29:04 crc kubenswrapper[4950]: I0318 20:29:04.969086 4950 scope.go:117] "RemoveContainer" containerID="80bb5b9fd4b42a17689aca26b1eb183f4363b2ddacc1fdf73b829648cfa1e873" Mar 18 20:29:06 crc kubenswrapper[4950]: I0318 20:29:06.551656 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 18 20:29:06 crc kubenswrapper[4950]: I0318 20:29:06.582099 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 18 20:29:07 crc kubenswrapper[4950]: I0318 20:29:07.046537 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 18 20:29:08 crc kubenswrapper[4950]: I0318 20:29:08.300081 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 18 20:29:08 crc kubenswrapper[4950]: I0318 20:29:08.300519 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 18 20:29:09 crc kubenswrapper[4950]: I0318 20:29:09.315625 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f05a94b0-defc-4c6c-af5c-3ce8346261a3" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 18 20:29:09 crc kubenswrapper[4950]: I0318 20:29:09.315632 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f05a94b0-defc-4c6c-af5c-3ce8346261a3" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 18 20:29:09 crc kubenswrapper[4950]: I0318 20:29:09.663264 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 18 20:29:09 crc kubenswrapper[4950]: I0318 20:29:09.663325 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 18 20:29:10 crc kubenswrapper[4950]: I0318 20:29:10.677627 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="603bbfd0-1fdf-44f7-bc34-ce9111ee204f" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.194:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 18 20:29:10 crc kubenswrapper[4950]: I0318 20:29:10.677627 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="603bbfd0-1fdf-44f7-bc34-ce9111ee204f" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.194:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 18 20:29:16 crc kubenswrapper[4950]: I0318 20:29:16.299832 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 18 20:29:16 crc kubenswrapper[4950]: I0318 20:29:16.300448 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 18 20:29:17 crc kubenswrapper[4950]: I0318 20:29:17.661288 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 18 20:29:17 crc kubenswrapper[4950]: I0318 20:29:17.663056 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 18 20:29:18 crc kubenswrapper[4950]: I0318 20:29:18.185950 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 18 20:29:18 crc kubenswrapper[4950]: I0318 20:29:18.308150 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 18 20:29:18 crc kubenswrapper[4950]: I0318 20:29:18.312228 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 18 20:29:18 crc kubenswrapper[4950]: I0318 20:29:18.327184 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 18 20:29:19 crc kubenswrapper[4950]: I0318 20:29:19.164798 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 18 20:29:19 crc kubenswrapper[4950]: I0318 20:29:19.667669 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 18 20:29:19 crc kubenswrapper[4950]: I0318 20:29:19.670209 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 18 20:29:19 crc kubenswrapper[4950]: I0318 20:29:19.677947 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 18 20:29:20 crc kubenswrapper[4950]: I0318 20:29:20.177770 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 18 20:29:21 crc kubenswrapper[4950]: I0318 20:29:21.018795 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 18 20:29:21 crc kubenswrapper[4950]: I0318 20:29:21.019025 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="859fa182-cb03-41ab-b127-20b3564cde97" containerName="kube-state-metrics" containerID="cri-o://cc57df497094b1e0391644a07ee2893fe6ea32b3d25a2060022a9233de18a1f3" gracePeriod=30 Mar 18 20:29:21 crc kubenswrapper[4950]: I0318 20:29:21.169978 4950 generic.go:334] "Generic (PLEG): container finished" podID="859fa182-cb03-41ab-b127-20b3564cde97" containerID="cc57df497094b1e0391644a07ee2893fe6ea32b3d25a2060022a9233de18a1f3" exitCode=2 Mar 18 20:29:21 crc kubenswrapper[4950]: I0318 20:29:21.170019 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"859fa182-cb03-41ab-b127-20b3564cde97","Type":"ContainerDied","Data":"cc57df497094b1e0391644a07ee2893fe6ea32b3d25a2060022a9233de18a1f3"} Mar 18 20:29:21 crc kubenswrapper[4950]: I0318 20:29:21.463068 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 18 20:29:21 crc kubenswrapper[4950]: I0318 20:29:21.524240 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7t8qg\" (UniqueName: \"kubernetes.io/projected/859fa182-cb03-41ab-b127-20b3564cde97-kube-api-access-7t8qg\") pod \"859fa182-cb03-41ab-b127-20b3564cde97\" (UID: \"859fa182-cb03-41ab-b127-20b3564cde97\") " Mar 18 20:29:21 crc kubenswrapper[4950]: I0318 20:29:21.529450 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/859fa182-cb03-41ab-b127-20b3564cde97-kube-api-access-7t8qg" (OuterVolumeSpecName: "kube-api-access-7t8qg") pod "859fa182-cb03-41ab-b127-20b3564cde97" (UID: "859fa182-cb03-41ab-b127-20b3564cde97"). InnerVolumeSpecName "kube-api-access-7t8qg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:29:21 crc kubenswrapper[4950]: I0318 20:29:21.626140 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7t8qg\" (UniqueName: \"kubernetes.io/projected/859fa182-cb03-41ab-b127-20b3564cde97-kube-api-access-7t8qg\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.102174 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.102449 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="00969ec6-eca3-4790-b69f-ba825f809823" containerName="ceilometer-central-agent" containerID="cri-o://8e823354f8410f2242ab0dd6d890021770466371f17c913ba99728c5a0073f32" gracePeriod=30 Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.102558 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="00969ec6-eca3-4790-b69f-ba825f809823" containerName="proxy-httpd" containerID="cri-o://45906ffaab429e682fdd942d7a9b2dd005d7cd0890f1a86ac3b2ce8cd586f984" gracePeriod=30 Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.102596 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="00969ec6-eca3-4790-b69f-ba825f809823" containerName="sg-core" containerID="cri-o://2ecc905e791972c42f485ef6d640aa017d161ec0f620d27f21a7090e6dbb0665" gracePeriod=30 Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.102631 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="00969ec6-eca3-4790-b69f-ba825f809823" containerName="ceilometer-notification-agent" containerID="cri-o://50d2f66b6b7b82c882ff2011cb5fc6653db27f8ede3e1c9d09e985c9cab29fb9" gracePeriod=30 Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.178543 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"859fa182-cb03-41ab-b127-20b3564cde97","Type":"ContainerDied","Data":"e202347533c299d2380522657ce7347670815f7e94f83241af7284fb98a18268"} Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.178576 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.178603 4950 scope.go:117] "RemoveContainer" containerID="cc57df497094b1e0391644a07ee2893fe6ea32b3d25a2060022a9233de18a1f3" Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.213709 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.247139 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.280139 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 18 20:29:22 crc kubenswrapper[4950]: E0318 20:29:22.291697 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="859fa182-cb03-41ab-b127-20b3564cde97" containerName="kube-state-metrics" Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.291737 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="859fa182-cb03-41ab-b127-20b3564cde97" containerName="kube-state-metrics" Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.291913 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="859fa182-cb03-41ab-b127-20b3564cde97" containerName="kube-state-metrics" Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.292538 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.298051 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.298701 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.304792 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.446299 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e\") " pod="openstack/kube-state-metrics-0" Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.446359 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e\") " pod="openstack/kube-state-metrics-0" Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.446403 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e\") " pod="openstack/kube-state-metrics-0" Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.446455 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg89b\" (UniqueName: \"kubernetes.io/projected/23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e-kube-api-access-lg89b\") pod \"kube-state-metrics-0\" (UID: \"23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e\") " pod="openstack/kube-state-metrics-0" Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.548260 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e\") " pod="openstack/kube-state-metrics-0" Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.548361 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e\") " pod="openstack/kube-state-metrics-0" Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.548422 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e\") " pod="openstack/kube-state-metrics-0" Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.548471 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg89b\" (UniqueName: \"kubernetes.io/projected/23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e-kube-api-access-lg89b\") pod \"kube-state-metrics-0\" (UID: \"23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e\") " pod="openstack/kube-state-metrics-0" Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.553369 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e\") " pod="openstack/kube-state-metrics-0" Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.553963 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e\") " pod="openstack/kube-state-metrics-0" Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.553803 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e\") " pod="openstack/kube-state-metrics-0" Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.565949 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg89b\" (UniqueName: \"kubernetes.io/projected/23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e-kube-api-access-lg89b\") pod \"kube-state-metrics-0\" (UID: \"23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e\") " pod="openstack/kube-state-metrics-0" Mar 18 20:29:22 crc kubenswrapper[4950]: I0318 20:29:22.657106 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 18 20:29:23 crc kubenswrapper[4950]: I0318 20:29:23.106165 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 18 20:29:23 crc kubenswrapper[4950]: W0318 20:29:23.113149 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23a67ca2_5df9_4946_a7d0_9e2dd7c0b91e.slice/crio-2f9bd6dd60cc145926a8567bf7ff66650ae834213321f6d880358b65f91761ad WatchSource:0}: Error finding container 2f9bd6dd60cc145926a8567bf7ff66650ae834213321f6d880358b65f91761ad: Status 404 returned error can't find the container with id 2f9bd6dd60cc145926a8567bf7ff66650ae834213321f6d880358b65f91761ad Mar 18 20:29:23 crc kubenswrapper[4950]: I0318 20:29:23.202634 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e","Type":"ContainerStarted","Data":"2f9bd6dd60cc145926a8567bf7ff66650ae834213321f6d880358b65f91761ad"} Mar 18 20:29:23 crc kubenswrapper[4950]: I0318 20:29:23.221314 4950 generic.go:334] "Generic (PLEG): container finished" podID="00969ec6-eca3-4790-b69f-ba825f809823" containerID="45906ffaab429e682fdd942d7a9b2dd005d7cd0890f1a86ac3b2ce8cd586f984" exitCode=0 Mar 18 20:29:23 crc kubenswrapper[4950]: I0318 20:29:23.221360 4950 generic.go:334] "Generic (PLEG): container finished" podID="00969ec6-eca3-4790-b69f-ba825f809823" containerID="2ecc905e791972c42f485ef6d640aa017d161ec0f620d27f21a7090e6dbb0665" exitCode=2 Mar 18 20:29:23 crc kubenswrapper[4950]: I0318 20:29:23.221372 4950 generic.go:334] "Generic (PLEG): container finished" podID="00969ec6-eca3-4790-b69f-ba825f809823" containerID="8e823354f8410f2242ab0dd6d890021770466371f17c913ba99728c5a0073f32" exitCode=0 Mar 18 20:29:23 crc kubenswrapper[4950]: I0318 20:29:23.221399 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00969ec6-eca3-4790-b69f-ba825f809823","Type":"ContainerDied","Data":"45906ffaab429e682fdd942d7a9b2dd005d7cd0890f1a86ac3b2ce8cd586f984"} Mar 18 20:29:23 crc kubenswrapper[4950]: I0318 20:29:23.221457 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00969ec6-eca3-4790-b69f-ba825f809823","Type":"ContainerDied","Data":"2ecc905e791972c42f485ef6d640aa017d161ec0f620d27f21a7090e6dbb0665"} Mar 18 20:29:23 crc kubenswrapper[4950]: I0318 20:29:23.221473 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00969ec6-eca3-4790-b69f-ba825f809823","Type":"ContainerDied","Data":"8e823354f8410f2242ab0dd6d890021770466371f17c913ba99728c5a0073f32"} Mar 18 20:29:23 crc kubenswrapper[4950]: I0318 20:29:23.489372 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="859fa182-cb03-41ab-b127-20b3564cde97" path="/var/lib/kubelet/pods/859fa182-cb03-41ab-b127-20b3564cde97/volumes" Mar 18 20:29:25 crc kubenswrapper[4950]: I0318 20:29:25.243081 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e","Type":"ContainerStarted","Data":"33720dabd27f4d99346890c92af6592341f8325010367bfc25e4d2edcde7d042"} Mar 18 20:29:25 crc kubenswrapper[4950]: I0318 20:29:25.243835 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 18 20:29:25 crc kubenswrapper[4950]: I0318 20:29:25.274890 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.959570134 podStartE2EDuration="3.274837394s" podCreationTimestamp="2026-03-18 20:29:22 +0000 UTC" firstStartedPulling="2026-03-18 20:29:23.116176496 +0000 UTC m=+1376.357018364" lastFinishedPulling="2026-03-18 20:29:24.431443756 +0000 UTC m=+1377.672285624" observedRunningTime="2026-03-18 20:29:25.272288857 +0000 UTC m=+1378.513130765" watchObservedRunningTime="2026-03-18 20:29:25.274837394 +0000 UTC m=+1378.515679282" Mar 18 20:29:26 crc kubenswrapper[4950]: I0318 20:29:26.665718 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:29:26 crc kubenswrapper[4950]: I0318 20:29:26.733363 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-config-data\") pod \"00969ec6-eca3-4790-b69f-ba825f809823\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " Mar 18 20:29:26 crc kubenswrapper[4950]: I0318 20:29:26.733438 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-sg-core-conf-yaml\") pod \"00969ec6-eca3-4790-b69f-ba825f809823\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " Mar 18 20:29:26 crc kubenswrapper[4950]: I0318 20:29:26.733553 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00969ec6-eca3-4790-b69f-ba825f809823-log-httpd\") pod \"00969ec6-eca3-4790-b69f-ba825f809823\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " Mar 18 20:29:26 crc kubenswrapper[4950]: I0318 20:29:26.733642 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00969ec6-eca3-4790-b69f-ba825f809823-run-httpd\") pod \"00969ec6-eca3-4790-b69f-ba825f809823\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " Mar 18 20:29:26 crc kubenswrapper[4950]: I0318 20:29:26.733687 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fg7pr\" (UniqueName: \"kubernetes.io/projected/00969ec6-eca3-4790-b69f-ba825f809823-kube-api-access-fg7pr\") pod \"00969ec6-eca3-4790-b69f-ba825f809823\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " Mar 18 20:29:26 crc kubenswrapper[4950]: I0318 20:29:26.733746 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-combined-ca-bundle\") pod \"00969ec6-eca3-4790-b69f-ba825f809823\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " Mar 18 20:29:26 crc kubenswrapper[4950]: I0318 20:29:26.733814 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-scripts\") pod \"00969ec6-eca3-4790-b69f-ba825f809823\" (UID: \"00969ec6-eca3-4790-b69f-ba825f809823\") " Mar 18 20:29:26 crc kubenswrapper[4950]: I0318 20:29:26.738987 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00969ec6-eca3-4790-b69f-ba825f809823-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "00969ec6-eca3-4790-b69f-ba825f809823" (UID: "00969ec6-eca3-4790-b69f-ba825f809823"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:29:26 crc kubenswrapper[4950]: I0318 20:29:26.739024 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00969ec6-eca3-4790-b69f-ba825f809823-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "00969ec6-eca3-4790-b69f-ba825f809823" (UID: "00969ec6-eca3-4790-b69f-ba825f809823"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:29:26 crc kubenswrapper[4950]: I0318 20:29:26.743625 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00969ec6-eca3-4790-b69f-ba825f809823-kube-api-access-fg7pr" (OuterVolumeSpecName: "kube-api-access-fg7pr") pod "00969ec6-eca3-4790-b69f-ba825f809823" (UID: "00969ec6-eca3-4790-b69f-ba825f809823"). InnerVolumeSpecName "kube-api-access-fg7pr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:29:26 crc kubenswrapper[4950]: I0318 20:29:26.746520 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-scripts" (OuterVolumeSpecName: "scripts") pod "00969ec6-eca3-4790-b69f-ba825f809823" (UID: "00969ec6-eca3-4790-b69f-ba825f809823"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:29:26 crc kubenswrapper[4950]: I0318 20:29:26.794716 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "00969ec6-eca3-4790-b69f-ba825f809823" (UID: "00969ec6-eca3-4790-b69f-ba825f809823"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:29:26 crc kubenswrapper[4950]: I0318 20:29:26.840141 4950 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00969ec6-eca3-4790-b69f-ba825f809823-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:26 crc kubenswrapper[4950]: I0318 20:29:26.840329 4950 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00969ec6-eca3-4790-b69f-ba825f809823-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:26 crc kubenswrapper[4950]: I0318 20:29:26.840357 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fg7pr\" (UniqueName: \"kubernetes.io/projected/00969ec6-eca3-4790-b69f-ba825f809823-kube-api-access-fg7pr\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:26 crc kubenswrapper[4950]: I0318 20:29:26.840542 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:26 crc kubenswrapper[4950]: I0318 20:29:26.840557 4950 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:26 crc kubenswrapper[4950]: I0318 20:29:26.860190 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "00969ec6-eca3-4790-b69f-ba825f809823" (UID: "00969ec6-eca3-4790-b69f-ba825f809823"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:29:26 crc kubenswrapper[4950]: I0318 20:29:26.876196 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-config-data" (OuterVolumeSpecName: "config-data") pod "00969ec6-eca3-4790-b69f-ba825f809823" (UID: "00969ec6-eca3-4790-b69f-ba825f809823"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:29:26 crc kubenswrapper[4950]: I0318 20:29:26.941756 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:26 crc kubenswrapper[4950]: I0318 20:29:26.941784 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00969ec6-eca3-4790-b69f-ba825f809823-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.262023 4950 generic.go:334] "Generic (PLEG): container finished" podID="00969ec6-eca3-4790-b69f-ba825f809823" containerID="50d2f66b6b7b82c882ff2011cb5fc6653db27f8ede3e1c9d09e985c9cab29fb9" exitCode=0 Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.262108 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.262142 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00969ec6-eca3-4790-b69f-ba825f809823","Type":"ContainerDied","Data":"50d2f66b6b7b82c882ff2011cb5fc6653db27f8ede3e1c9d09e985c9cab29fb9"} Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.262586 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00969ec6-eca3-4790-b69f-ba825f809823","Type":"ContainerDied","Data":"6291fa3b694a2d82bf767d88d1e5493818cb5eed30d9ccb6e9c0571df9e1aef2"} Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.262610 4950 scope.go:117] "RemoveContainer" containerID="45906ffaab429e682fdd942d7a9b2dd005d7cd0890f1a86ac3b2ce8cd586f984" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.281277 4950 scope.go:117] "RemoveContainer" containerID="2ecc905e791972c42f485ef6d640aa017d161ec0f620d27f21a7090e6dbb0665" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.304692 4950 scope.go:117] "RemoveContainer" containerID="50d2f66b6b7b82c882ff2011cb5fc6653db27f8ede3e1c9d09e985c9cab29fb9" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.310262 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.321811 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.368135 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:29:27 crc kubenswrapper[4950]: E0318 20:29:27.368921 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00969ec6-eca3-4790-b69f-ba825f809823" containerName="proxy-httpd" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.368941 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="00969ec6-eca3-4790-b69f-ba825f809823" containerName="proxy-httpd" Mar 18 20:29:27 crc kubenswrapper[4950]: E0318 20:29:27.368964 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00969ec6-eca3-4790-b69f-ba825f809823" containerName="ceilometer-central-agent" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.368973 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="00969ec6-eca3-4790-b69f-ba825f809823" containerName="ceilometer-central-agent" Mar 18 20:29:27 crc kubenswrapper[4950]: E0318 20:29:27.368990 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00969ec6-eca3-4790-b69f-ba825f809823" containerName="sg-core" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.368999 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="00969ec6-eca3-4790-b69f-ba825f809823" containerName="sg-core" Mar 18 20:29:27 crc kubenswrapper[4950]: E0318 20:29:27.369034 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00969ec6-eca3-4790-b69f-ba825f809823" containerName="ceilometer-notification-agent" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.369042 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="00969ec6-eca3-4790-b69f-ba825f809823" containerName="ceilometer-notification-agent" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.369274 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="00969ec6-eca3-4790-b69f-ba825f809823" containerName="ceilometer-notification-agent" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.369290 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="00969ec6-eca3-4790-b69f-ba825f809823" containerName="proxy-httpd" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.369319 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="00969ec6-eca3-4790-b69f-ba825f809823" containerName="sg-core" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.369330 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="00969ec6-eca3-4790-b69f-ba825f809823" containerName="ceilometer-central-agent" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.371687 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.374402 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.375054 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.375202 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.386375 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.410342 4950 scope.go:117] "RemoveContainer" containerID="8e823354f8410f2242ab0dd6d890021770466371f17c913ba99728c5a0073f32" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.431630 4950 scope.go:117] "RemoveContainer" containerID="45906ffaab429e682fdd942d7a9b2dd005d7cd0890f1a86ac3b2ce8cd586f984" Mar 18 20:29:27 crc kubenswrapper[4950]: E0318 20:29:27.443723 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45906ffaab429e682fdd942d7a9b2dd005d7cd0890f1a86ac3b2ce8cd586f984\": container with ID starting with 45906ffaab429e682fdd942d7a9b2dd005d7cd0890f1a86ac3b2ce8cd586f984 not found: ID does not exist" containerID="45906ffaab429e682fdd942d7a9b2dd005d7cd0890f1a86ac3b2ce8cd586f984" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.443790 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45906ffaab429e682fdd942d7a9b2dd005d7cd0890f1a86ac3b2ce8cd586f984"} err="failed to get container status \"45906ffaab429e682fdd942d7a9b2dd005d7cd0890f1a86ac3b2ce8cd586f984\": rpc error: code = NotFound desc = could not find container \"45906ffaab429e682fdd942d7a9b2dd005d7cd0890f1a86ac3b2ce8cd586f984\": container with ID starting with 45906ffaab429e682fdd942d7a9b2dd005d7cd0890f1a86ac3b2ce8cd586f984 not found: ID does not exist" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.443829 4950 scope.go:117] "RemoveContainer" containerID="2ecc905e791972c42f485ef6d640aa017d161ec0f620d27f21a7090e6dbb0665" Mar 18 20:29:27 crc kubenswrapper[4950]: E0318 20:29:27.444982 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ecc905e791972c42f485ef6d640aa017d161ec0f620d27f21a7090e6dbb0665\": container with ID starting with 2ecc905e791972c42f485ef6d640aa017d161ec0f620d27f21a7090e6dbb0665 not found: ID does not exist" containerID="2ecc905e791972c42f485ef6d640aa017d161ec0f620d27f21a7090e6dbb0665" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.445015 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ecc905e791972c42f485ef6d640aa017d161ec0f620d27f21a7090e6dbb0665"} err="failed to get container status \"2ecc905e791972c42f485ef6d640aa017d161ec0f620d27f21a7090e6dbb0665\": rpc error: code = NotFound desc = could not find container \"2ecc905e791972c42f485ef6d640aa017d161ec0f620d27f21a7090e6dbb0665\": container with ID starting with 2ecc905e791972c42f485ef6d640aa017d161ec0f620d27f21a7090e6dbb0665 not found: ID does not exist" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.445037 4950 scope.go:117] "RemoveContainer" containerID="50d2f66b6b7b82c882ff2011cb5fc6653db27f8ede3e1c9d09e985c9cab29fb9" Mar 18 20:29:27 crc kubenswrapper[4950]: E0318 20:29:27.446210 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50d2f66b6b7b82c882ff2011cb5fc6653db27f8ede3e1c9d09e985c9cab29fb9\": container with ID starting with 50d2f66b6b7b82c882ff2011cb5fc6653db27f8ede3e1c9d09e985c9cab29fb9 not found: ID does not exist" containerID="50d2f66b6b7b82c882ff2011cb5fc6653db27f8ede3e1c9d09e985c9cab29fb9" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.446253 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50d2f66b6b7b82c882ff2011cb5fc6653db27f8ede3e1c9d09e985c9cab29fb9"} err="failed to get container status \"50d2f66b6b7b82c882ff2011cb5fc6653db27f8ede3e1c9d09e985c9cab29fb9\": rpc error: code = NotFound desc = could not find container \"50d2f66b6b7b82c882ff2011cb5fc6653db27f8ede3e1c9d09e985c9cab29fb9\": container with ID starting with 50d2f66b6b7b82c882ff2011cb5fc6653db27f8ede3e1c9d09e985c9cab29fb9 not found: ID does not exist" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.446275 4950 scope.go:117] "RemoveContainer" containerID="8e823354f8410f2242ab0dd6d890021770466371f17c913ba99728c5a0073f32" Mar 18 20:29:27 crc kubenswrapper[4950]: E0318 20:29:27.446670 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e823354f8410f2242ab0dd6d890021770466371f17c913ba99728c5a0073f32\": container with ID starting with 8e823354f8410f2242ab0dd6d890021770466371f17c913ba99728c5a0073f32 not found: ID does not exist" containerID="8e823354f8410f2242ab0dd6d890021770466371f17c913ba99728c5a0073f32" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.446721 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e823354f8410f2242ab0dd6d890021770466371f17c913ba99728c5a0073f32"} err="failed to get container status \"8e823354f8410f2242ab0dd6d890021770466371f17c913ba99728c5a0073f32\": rpc error: code = NotFound desc = could not find container \"8e823354f8410f2242ab0dd6d890021770466371f17c913ba99728c5a0073f32\": container with ID starting with 8e823354f8410f2242ab0dd6d890021770466371f17c913ba99728c5a0073f32 not found: ID does not exist" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.451454 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9703e257-1e49-46e0-bbdf-754fee621b4b-run-httpd\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.451579 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.451614 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g54r\" (UniqueName: \"kubernetes.io/projected/9703e257-1e49-46e0-bbdf-754fee621b4b-kube-api-access-8g54r\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.451646 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.451672 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.451788 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9703e257-1e49-46e0-bbdf-754fee621b4b-log-httpd\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.452162 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-config-data\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.452386 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-scripts\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.489844 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00969ec6-eca3-4790-b69f-ba825f809823" path="/var/lib/kubelet/pods/00969ec6-eca3-4790-b69f-ba825f809823/volumes" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.553763 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g54r\" (UniqueName: \"kubernetes.io/projected/9703e257-1e49-46e0-bbdf-754fee621b4b-kube-api-access-8g54r\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.554190 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.554218 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.554312 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9703e257-1e49-46e0-bbdf-754fee621b4b-log-httpd\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.554370 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-config-data\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.554433 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-scripts\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.554463 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9703e257-1e49-46e0-bbdf-754fee621b4b-run-httpd\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.554505 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.555451 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9703e257-1e49-46e0-bbdf-754fee621b4b-run-httpd\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.555883 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9703e257-1e49-46e0-bbdf-754fee621b4b-log-httpd\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.559713 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.559846 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-config-data\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.560504 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.563000 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-scripts\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.565388 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.575059 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g54r\" (UniqueName: \"kubernetes.io/projected/9703e257-1e49-46e0-bbdf-754fee621b4b-kube-api-access-8g54r\") pod \"ceilometer-0\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.703623 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 20:29:27 crc kubenswrapper[4950]: I0318 20:29:27.928629 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 18 20:29:28 crc kubenswrapper[4950]: I0318 20:29:28.274922 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 18 20:29:29 crc kubenswrapper[4950]: I0318 20:29:29.068950 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 18 20:29:29 crc kubenswrapper[4950]: I0318 20:29:29.285574 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9703e257-1e49-46e0-bbdf-754fee621b4b","Type":"ContainerStarted","Data":"6ac425705d5c6fa63f048c24cc9ed5e70a1c540588ab51d299b9b060a652d2b2"} Mar 18 20:29:29 crc kubenswrapper[4950]: I0318 20:29:29.286110 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9703e257-1e49-46e0-bbdf-754fee621b4b","Type":"ContainerStarted","Data":"c17a68aca9578a4302873d6a5b8d8c23d033296270f81cf3681f5b96d28da2e5"} Mar 18 20:29:30 crc kubenswrapper[4950]: I0318 20:29:30.294278 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9703e257-1e49-46e0-bbdf-754fee621b4b","Type":"ContainerStarted","Data":"a43c585b19441af8a3824713bad197517a01bc13c492af5db16fd853273c86fc"} Mar 18 20:29:31 crc kubenswrapper[4950]: I0318 20:29:31.303724 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9703e257-1e49-46e0-bbdf-754fee621b4b","Type":"ContainerStarted","Data":"f39bf9b9db435c44a5af5d72d57bfbf948467d00f15a7ae19bd9527a0f06c5c5"} Mar 18 20:29:32 crc kubenswrapper[4950]: I0318 20:29:32.669121 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 18 20:29:33 crc kubenswrapper[4950]: I0318 20:29:33.234905 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="b82f7668-ae1a-4bfe-be27-435cca5df467" containerName="rabbitmq" containerID="cri-o://ffa976652e557de2dbb7eaffaa7cda3491e64db27a9601f202aa674559da0c3b" gracePeriod=604795 Mar 18 20:29:33 crc kubenswrapper[4950]: I0318 20:29:33.334879 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9703e257-1e49-46e0-bbdf-754fee621b4b","Type":"ContainerStarted","Data":"a4960228dd8e0cdea9b485c596699b1a53e857466115f3997b8d6047aa7aecb7"} Mar 18 20:29:33 crc kubenswrapper[4950]: I0318 20:29:33.335138 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 18 20:29:33 crc kubenswrapper[4950]: I0318 20:29:33.359045 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.8037496050000001 podStartE2EDuration="6.359032153s" podCreationTimestamp="2026-03-18 20:29:27 +0000 UTC" firstStartedPulling="2026-03-18 20:29:28.272741139 +0000 UTC m=+1381.513583007" lastFinishedPulling="2026-03-18 20:29:32.828023687 +0000 UTC m=+1386.068865555" observedRunningTime="2026-03-18 20:29:33.355275944 +0000 UTC m=+1386.596117812" watchObservedRunningTime="2026-03-18 20:29:33.359032153 +0000 UTC m=+1386.599874021" Mar 18 20:29:34 crc kubenswrapper[4950]: I0318 20:29:34.168256 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="7502f5ea-31ca-4349-bd79-2219bf3903c0" containerName="rabbitmq" containerID="cri-o://b069af21eaa4256e43fdff47e8842455ff9b0335a0b5b04aec42cea30f1c13fb" gracePeriod=604795 Mar 18 20:29:39 crc kubenswrapper[4950]: I0318 20:29:39.817162 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 18 20:29:39 crc kubenswrapper[4950]: I0318 20:29:39.960513 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="7502f5ea-31ca-4349-bd79-2219bf3903c0" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Mar 18 20:29:39 crc kubenswrapper[4950]: I0318 20:29:39.977231 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b82f7668-ae1a-4bfe-be27-435cca5df467-plugins-conf\") pod \"b82f7668-ae1a-4bfe-be27-435cca5df467\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " Mar 18 20:29:39 crc kubenswrapper[4950]: I0318 20:29:39.977293 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-confd\") pod \"b82f7668-ae1a-4bfe-be27-435cca5df467\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " Mar 18 20:29:39 crc kubenswrapper[4950]: I0318 20:29:39.977342 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5gd4\" (UniqueName: \"kubernetes.io/projected/b82f7668-ae1a-4bfe-be27-435cca5df467-kube-api-access-g5gd4\") pod \"b82f7668-ae1a-4bfe-be27-435cca5df467\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " Mar 18 20:29:39 crc kubenswrapper[4950]: I0318 20:29:39.977372 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b82f7668-ae1a-4bfe-be27-435cca5df467-server-conf\") pod \"b82f7668-ae1a-4bfe-be27-435cca5df467\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " Mar 18 20:29:39 crc kubenswrapper[4950]: I0318 20:29:39.977445 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-erlang-cookie\") pod \"b82f7668-ae1a-4bfe-be27-435cca5df467\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " Mar 18 20:29:39 crc kubenswrapper[4950]: I0318 20:29:39.977800 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b82f7668-ae1a-4bfe-be27-435cca5df467-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "b82f7668-ae1a-4bfe-be27-435cca5df467" (UID: "b82f7668-ae1a-4bfe-be27-435cca5df467"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:29:39 crc kubenswrapper[4950]: I0318 20:29:39.977898 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "b82f7668-ae1a-4bfe-be27-435cca5df467" (UID: "b82f7668-ae1a-4bfe-be27-435cca5df467"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:29:39 crc kubenswrapper[4950]: I0318 20:29:39.978180 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b82f7668-ae1a-4bfe-be27-435cca5df467-pod-info\") pod \"b82f7668-ae1a-4bfe-be27-435cca5df467\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " Mar 18 20:29:39 crc kubenswrapper[4950]: I0318 20:29:39.978220 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-plugins\") pod \"b82f7668-ae1a-4bfe-be27-435cca5df467\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " Mar 18 20:29:39 crc kubenswrapper[4950]: I0318 20:29:39.978308 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b82f7668-ae1a-4bfe-be27-435cca5df467-erlang-cookie-secret\") pod \"b82f7668-ae1a-4bfe-be27-435cca5df467\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " Mar 18 20:29:39 crc kubenswrapper[4950]: I0318 20:29:39.978354 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b82f7668-ae1a-4bfe-be27-435cca5df467-config-data\") pod \"b82f7668-ae1a-4bfe-be27-435cca5df467\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " Mar 18 20:29:39 crc kubenswrapper[4950]: I0318 20:29:39.978390 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"b82f7668-ae1a-4bfe-be27-435cca5df467\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " Mar 18 20:29:39 crc kubenswrapper[4950]: I0318 20:29:39.978446 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-tls\") pod \"b82f7668-ae1a-4bfe-be27-435cca5df467\" (UID: \"b82f7668-ae1a-4bfe-be27-435cca5df467\") " Mar 18 20:29:39 crc kubenswrapper[4950]: I0318 20:29:39.978606 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "b82f7668-ae1a-4bfe-be27-435cca5df467" (UID: "b82f7668-ae1a-4bfe-be27-435cca5df467"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:29:39 crc kubenswrapper[4950]: I0318 20:29:39.979089 4950 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:39 crc kubenswrapper[4950]: I0318 20:29:39.979106 4950 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:39 crc kubenswrapper[4950]: I0318 20:29:39.979114 4950 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b82f7668-ae1a-4bfe-be27-435cca5df467-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:39 crc kubenswrapper[4950]: I0318 20:29:39.984093 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b82f7668-ae1a-4bfe-be27-435cca5df467-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "b82f7668-ae1a-4bfe-be27-435cca5df467" (UID: "b82f7668-ae1a-4bfe-be27-435cca5df467"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:29:39 crc kubenswrapper[4950]: I0318 20:29:39.984423 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "b82f7668-ae1a-4bfe-be27-435cca5df467" (UID: "b82f7668-ae1a-4bfe-be27-435cca5df467"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:29:39 crc kubenswrapper[4950]: I0318 20:29:39.985339 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/b82f7668-ae1a-4bfe-be27-435cca5df467-pod-info" (OuterVolumeSpecName: "pod-info") pod "b82f7668-ae1a-4bfe-be27-435cca5df467" (UID: "b82f7668-ae1a-4bfe-be27-435cca5df467"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.018362 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "b82f7668-ae1a-4bfe-be27-435cca5df467" (UID: "b82f7668-ae1a-4bfe-be27-435cca5df467"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.023056 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b82f7668-ae1a-4bfe-be27-435cca5df467-kube-api-access-g5gd4" (OuterVolumeSpecName: "kube-api-access-g5gd4") pod "b82f7668-ae1a-4bfe-be27-435cca5df467" (UID: "b82f7668-ae1a-4bfe-be27-435cca5df467"). InnerVolumeSpecName "kube-api-access-g5gd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.027279 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b82f7668-ae1a-4bfe-be27-435cca5df467-config-data" (OuterVolumeSpecName: "config-data") pod "b82f7668-ae1a-4bfe-be27-435cca5df467" (UID: "b82f7668-ae1a-4bfe-be27-435cca5df467"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.066548 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b82f7668-ae1a-4bfe-be27-435cca5df467-server-conf" (OuterVolumeSpecName: "server-conf") pod "b82f7668-ae1a-4bfe-be27-435cca5df467" (UID: "b82f7668-ae1a-4bfe-be27-435cca5df467"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.080634 4950 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.080670 4950 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.080685 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5gd4\" (UniqueName: \"kubernetes.io/projected/b82f7668-ae1a-4bfe-be27-435cca5df467-kube-api-access-g5gd4\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.080694 4950 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b82f7668-ae1a-4bfe-be27-435cca5df467-server-conf\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.080702 4950 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b82f7668-ae1a-4bfe-be27-435cca5df467-pod-info\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.080713 4950 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b82f7668-ae1a-4bfe-be27-435cca5df467-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.080721 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b82f7668-ae1a-4bfe-be27-435cca5df467-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.098815 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "b82f7668-ae1a-4bfe-be27-435cca5df467" (UID: "b82f7668-ae1a-4bfe-be27-435cca5df467"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.099507 4950 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.183564 4950 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.183600 4950 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b82f7668-ae1a-4bfe-be27-435cca5df467-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.401225 4950 generic.go:334] "Generic (PLEG): container finished" podID="b82f7668-ae1a-4bfe-be27-435cca5df467" containerID="ffa976652e557de2dbb7eaffaa7cda3491e64db27a9601f202aa674559da0c3b" exitCode=0 Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.401263 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b82f7668-ae1a-4bfe-be27-435cca5df467","Type":"ContainerDied","Data":"ffa976652e557de2dbb7eaffaa7cda3491e64db27a9601f202aa674559da0c3b"} Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.401300 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.401332 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b82f7668-ae1a-4bfe-be27-435cca5df467","Type":"ContainerDied","Data":"87ba0f1995bc9ca4f94deefc7d8d1428ea35680dcfdc49bf02fc667ea2a2d3d0"} Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.401356 4950 scope.go:117] "RemoveContainer" containerID="ffa976652e557de2dbb7eaffaa7cda3491e64db27a9601f202aa674559da0c3b" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.406213 4950 generic.go:334] "Generic (PLEG): container finished" podID="7502f5ea-31ca-4349-bd79-2219bf3903c0" containerID="b069af21eaa4256e43fdff47e8842455ff9b0335a0b5b04aec42cea30f1c13fb" exitCode=0 Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.406284 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7502f5ea-31ca-4349-bd79-2219bf3903c0","Type":"ContainerDied","Data":"b069af21eaa4256e43fdff47e8842455ff9b0335a0b5b04aec42cea30f1c13fb"} Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.428831 4950 scope.go:117] "RemoveContainer" containerID="a9215184422c6bfb281f115311fa646f8d17ed53337255e639a33ab2c4583605" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.439355 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.480216 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.497031 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 18 20:29:40 crc kubenswrapper[4950]: E0318 20:29:40.497427 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b82f7668-ae1a-4bfe-be27-435cca5df467" containerName="rabbitmq" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.497444 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="b82f7668-ae1a-4bfe-be27-435cca5df467" containerName="rabbitmq" Mar 18 20:29:40 crc kubenswrapper[4950]: E0318 20:29:40.497502 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b82f7668-ae1a-4bfe-be27-435cca5df467" containerName="setup-container" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.497509 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="b82f7668-ae1a-4bfe-be27-435cca5df467" containerName="setup-container" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.497669 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="b82f7668-ae1a-4bfe-be27-435cca5df467" containerName="rabbitmq" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.498606 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.504966 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.505136 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.505322 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.505448 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.505828 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.506074 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.506275 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-xm9qz" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.511479 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.512170 4950 scope.go:117] "RemoveContainer" containerID="ffa976652e557de2dbb7eaffaa7cda3491e64db27a9601f202aa674559da0c3b" Mar 18 20:29:40 crc kubenswrapper[4950]: E0318 20:29:40.517206 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffa976652e557de2dbb7eaffaa7cda3491e64db27a9601f202aa674559da0c3b\": container with ID starting with ffa976652e557de2dbb7eaffaa7cda3491e64db27a9601f202aa674559da0c3b not found: ID does not exist" containerID="ffa976652e557de2dbb7eaffaa7cda3491e64db27a9601f202aa674559da0c3b" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.517265 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffa976652e557de2dbb7eaffaa7cda3491e64db27a9601f202aa674559da0c3b"} err="failed to get container status \"ffa976652e557de2dbb7eaffaa7cda3491e64db27a9601f202aa674559da0c3b\": rpc error: code = NotFound desc = could not find container \"ffa976652e557de2dbb7eaffaa7cda3491e64db27a9601f202aa674559da0c3b\": container with ID starting with ffa976652e557de2dbb7eaffaa7cda3491e64db27a9601f202aa674559da0c3b not found: ID does not exist" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.517298 4950 scope.go:117] "RemoveContainer" containerID="a9215184422c6bfb281f115311fa646f8d17ed53337255e639a33ab2c4583605" Mar 18 20:29:40 crc kubenswrapper[4950]: E0318 20:29:40.523700 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9215184422c6bfb281f115311fa646f8d17ed53337255e639a33ab2c4583605\": container with ID starting with a9215184422c6bfb281f115311fa646f8d17ed53337255e639a33ab2c4583605 not found: ID does not exist" containerID="a9215184422c6bfb281f115311fa646f8d17ed53337255e639a33ab2c4583605" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.523761 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9215184422c6bfb281f115311fa646f8d17ed53337255e639a33ab2c4583605"} err="failed to get container status \"a9215184422c6bfb281f115311fa646f8d17ed53337255e639a33ab2c4583605\": rpc error: code = NotFound desc = could not find container \"a9215184422c6bfb281f115311fa646f8d17ed53337255e639a33ab2c4583605\": container with ID starting with a9215184422c6bfb281f115311fa646f8d17ed53337255e639a33ab2c4583605 not found: ID does not exist" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.691905 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.691958 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-config-data\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.691993 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.692015 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjhgv\" (UniqueName: \"kubernetes.io/projected/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-kube-api-access-xjhgv\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.692038 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.692068 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-pod-info\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.692087 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.692106 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.692129 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-server-conf\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.692144 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.692189 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.796546 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.797041 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.797080 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.797116 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-server-conf\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.797135 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.797186 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.798936 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-server-conf\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.799440 4950 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.800606 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.800798 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-config-data\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.800919 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.801001 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjhgv\" (UniqueName: \"kubernetes.io/projected/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-kube-api-access-xjhgv\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.801069 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.801176 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-pod-info\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.803747 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.803994 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.804018 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-config-data\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.807347 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.809205 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.816366 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-pod-info\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.817057 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.819217 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjhgv\" (UniqueName: \"kubernetes.io/projected/27be57d1-fc6d-4fe7-a6b3-c7276d5ce547-kube-api-access-xjhgv\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.856357 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:40 crc kubenswrapper[4950]: I0318 20:29:40.865097 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547\") " pod="openstack/rabbitmq-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.010240 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtmb7\" (UniqueName: \"kubernetes.io/projected/7502f5ea-31ca-4349-bd79-2219bf3903c0-kube-api-access-gtmb7\") pod \"7502f5ea-31ca-4349-bd79-2219bf3903c0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.010296 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-erlang-cookie\") pod \"7502f5ea-31ca-4349-bd79-2219bf3903c0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.010320 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7502f5ea-31ca-4349-bd79-2219bf3903c0-plugins-conf\") pod \"7502f5ea-31ca-4349-bd79-2219bf3903c0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.010361 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-confd\") pod \"7502f5ea-31ca-4349-bd79-2219bf3903c0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.010384 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-plugins\") pod \"7502f5ea-31ca-4349-bd79-2219bf3903c0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.010790 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7502f5ea-31ca-4349-bd79-2219bf3903c0-config-data\") pod \"7502f5ea-31ca-4349-bd79-2219bf3903c0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.010841 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-tls\") pod \"7502f5ea-31ca-4349-bd79-2219bf3903c0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.010880 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7502f5ea-31ca-4349-bd79-2219bf3903c0-server-conf\") pod \"7502f5ea-31ca-4349-bd79-2219bf3903c0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.010936 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7502f5ea-31ca-4349-bd79-2219bf3903c0-pod-info\") pod \"7502f5ea-31ca-4349-bd79-2219bf3903c0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.010971 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"7502f5ea-31ca-4349-bd79-2219bf3903c0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.011007 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7502f5ea-31ca-4349-bd79-2219bf3903c0-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "7502f5ea-31ca-4349-bd79-2219bf3903c0" (UID: "7502f5ea-31ca-4349-bd79-2219bf3903c0"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.011023 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7502f5ea-31ca-4349-bd79-2219bf3903c0-erlang-cookie-secret\") pod \"7502f5ea-31ca-4349-bd79-2219bf3903c0\" (UID: \"7502f5ea-31ca-4349-bd79-2219bf3903c0\") " Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.011069 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "7502f5ea-31ca-4349-bd79-2219bf3903c0" (UID: "7502f5ea-31ca-4349-bd79-2219bf3903c0"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.011787 4950 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7502f5ea-31ca-4349-bd79-2219bf3903c0-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.011804 4950 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.012460 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "7502f5ea-31ca-4349-bd79-2219bf3903c0" (UID: "7502f5ea-31ca-4349-bd79-2219bf3903c0"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.015906 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7502f5ea-31ca-4349-bd79-2219bf3903c0-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "7502f5ea-31ca-4349-bd79-2219bf3903c0" (UID: "7502f5ea-31ca-4349-bd79-2219bf3903c0"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.015974 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7502f5ea-31ca-4349-bd79-2219bf3903c0-kube-api-access-gtmb7" (OuterVolumeSpecName: "kube-api-access-gtmb7") pod "7502f5ea-31ca-4349-bd79-2219bf3903c0" (UID: "7502f5ea-31ca-4349-bd79-2219bf3903c0"). InnerVolumeSpecName "kube-api-access-gtmb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.017117 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "7502f5ea-31ca-4349-bd79-2219bf3903c0" (UID: "7502f5ea-31ca-4349-bd79-2219bf3903c0"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.017637 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "7502f5ea-31ca-4349-bd79-2219bf3903c0" (UID: "7502f5ea-31ca-4349-bd79-2219bf3903c0"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.018559 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/7502f5ea-31ca-4349-bd79-2219bf3903c0-pod-info" (OuterVolumeSpecName: "pod-info") pod "7502f5ea-31ca-4349-bd79-2219bf3903c0" (UID: "7502f5ea-31ca-4349-bd79-2219bf3903c0"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.052144 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7502f5ea-31ca-4349-bd79-2219bf3903c0-config-data" (OuterVolumeSpecName: "config-data") pod "7502f5ea-31ca-4349-bd79-2219bf3903c0" (UID: "7502f5ea-31ca-4349-bd79-2219bf3903c0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.088637 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7502f5ea-31ca-4349-bd79-2219bf3903c0-server-conf" (OuterVolumeSpecName: "server-conf") pod "7502f5ea-31ca-4349-bd79-2219bf3903c0" (UID: "7502f5ea-31ca-4349-bd79-2219bf3903c0"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.102386 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "7502f5ea-31ca-4349-bd79-2219bf3903c0" (UID: "7502f5ea-31ca-4349-bd79-2219bf3903c0"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.113222 4950 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7502f5ea-31ca-4349-bd79-2219bf3903c0-pod-info\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.113400 4950 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.113504 4950 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7502f5ea-31ca-4349-bd79-2219bf3903c0-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.113589 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtmb7\" (UniqueName: \"kubernetes.io/projected/7502f5ea-31ca-4349-bd79-2219bf3903c0-kube-api-access-gtmb7\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.113673 4950 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.113741 4950 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.113811 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7502f5ea-31ca-4349-bd79-2219bf3903c0-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.113878 4950 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7502f5ea-31ca-4349-bd79-2219bf3903c0-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.113941 4950 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7502f5ea-31ca-4349-bd79-2219bf3903c0-server-conf\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.131883 4950 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.154768 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.215777 4950 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.420219 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"7502f5ea-31ca-4349-bd79-2219bf3903c0","Type":"ContainerDied","Data":"499d73cbee8abb443a1869c77ddd26c53518424c0040f187778794b996492a62"} Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.420572 4950 scope.go:117] "RemoveContainer" containerID="b069af21eaa4256e43fdff47e8842455ff9b0335a0b5b04aec42cea30f1c13fb" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.420374 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.452717 4950 scope.go:117] "RemoveContainer" containerID="c87e384b986ec489013da0da075d69803e2eecf7ea2534fbb8b55233e4c8990e" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.490719 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b82f7668-ae1a-4bfe-be27-435cca5df467" path="/var/lib/kubelet/pods/b82f7668-ae1a-4bfe-be27-435cca5df467/volumes" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.519613 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.528770 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.556396 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 18 20:29:41 crc kubenswrapper[4950]: E0318 20:29:41.556873 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7502f5ea-31ca-4349-bd79-2219bf3903c0" containerName="rabbitmq" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.556894 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="7502f5ea-31ca-4349-bd79-2219bf3903c0" containerName="rabbitmq" Mar 18 20:29:41 crc kubenswrapper[4950]: E0318 20:29:41.556920 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7502f5ea-31ca-4349-bd79-2219bf3903c0" containerName="setup-container" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.556930 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="7502f5ea-31ca-4349-bd79-2219bf3903c0" containerName="setup-container" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.557189 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="7502f5ea-31ca-4349-bd79-2219bf3903c0" containerName="rabbitmq" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.558407 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.561103 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.561308 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.561799 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.561954 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-bjdw7" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.562063 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.562170 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.562305 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.581612 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.638675 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.728996 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/435fb3ca-c4ad-4b0a-a832-88f4a7203def-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.729360 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/435fb3ca-c4ad-4b0a-a832-88f4a7203def-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.729424 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkz8x\" (UniqueName: \"kubernetes.io/projected/435fb3ca-c4ad-4b0a-a832-88f4a7203def-kube-api-access-nkz8x\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.729448 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/435fb3ca-c4ad-4b0a-a832-88f4a7203def-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.729477 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/435fb3ca-c4ad-4b0a-a832-88f4a7203def-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.729529 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/435fb3ca-c4ad-4b0a-a832-88f4a7203def-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.729559 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/435fb3ca-c4ad-4b0a-a832-88f4a7203def-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.729578 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/435fb3ca-c4ad-4b0a-a832-88f4a7203def-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.729600 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/435fb3ca-c4ad-4b0a-a832-88f4a7203def-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.729638 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/435fb3ca-c4ad-4b0a-a832-88f4a7203def-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.729661 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.831393 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/435fb3ca-c4ad-4b0a-a832-88f4a7203def-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.831489 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/435fb3ca-c4ad-4b0a-a832-88f4a7203def-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.831509 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/435fb3ca-c4ad-4b0a-a832-88f4a7203def-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.831531 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/435fb3ca-c4ad-4b0a-a832-88f4a7203def-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.831570 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/435fb3ca-c4ad-4b0a-a832-88f4a7203def-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.831593 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.831618 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/435fb3ca-c4ad-4b0a-a832-88f4a7203def-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.831638 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/435fb3ca-c4ad-4b0a-a832-88f4a7203def-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.831677 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkz8x\" (UniqueName: \"kubernetes.io/projected/435fb3ca-c4ad-4b0a-a832-88f4a7203def-kube-api-access-nkz8x\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.832188 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/435fb3ca-c4ad-4b0a-a832-88f4a7203def-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.832444 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/435fb3ca-c4ad-4b0a-a832-88f4a7203def-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.832488 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/435fb3ca-c4ad-4b0a-a832-88f4a7203def-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.832779 4950 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.832810 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/435fb3ca-c4ad-4b0a-a832-88f4a7203def-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.832797 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/435fb3ca-c4ad-4b0a-a832-88f4a7203def-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.833317 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/435fb3ca-c4ad-4b0a-a832-88f4a7203def-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.833755 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/435fb3ca-c4ad-4b0a-a832-88f4a7203def-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.836087 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/435fb3ca-c4ad-4b0a-a832-88f4a7203def-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.836370 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/435fb3ca-c4ad-4b0a-a832-88f4a7203def-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.836982 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/435fb3ca-c4ad-4b0a-a832-88f4a7203def-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.838624 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/435fb3ca-c4ad-4b0a-a832-88f4a7203def-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.848526 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkz8x\" (UniqueName: \"kubernetes.io/projected/435fb3ca-c4ad-4b0a-a832-88f4a7203def-kube-api-access-nkz8x\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.859676 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"435fb3ca-c4ad-4b0a-a832-88f4a7203def\") " pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:41 crc kubenswrapper[4950]: I0318 20:29:41.883669 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:29:42 crc kubenswrapper[4950]: I0318 20:29:42.394342 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 18 20:29:42 crc kubenswrapper[4950]: W0318 20:29:42.396302 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod435fb3ca_c4ad_4b0a_a832_88f4a7203def.slice/crio-4d44f3b23bf6e12335412f9f0cbabf174474f45f71ff46d762750da9cade1ed6 WatchSource:0}: Error finding container 4d44f3b23bf6e12335412f9f0cbabf174474f45f71ff46d762750da9cade1ed6: Status 404 returned error can't find the container with id 4d44f3b23bf6e12335412f9f0cbabf174474f45f71ff46d762750da9cade1ed6 Mar 18 20:29:42 crc kubenswrapper[4950]: I0318 20:29:42.431555 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547","Type":"ContainerStarted","Data":"6a4d8c69a0048bab234f527bde622cd732e6ae6319d79084249e727e320353c1"} Mar 18 20:29:42 crc kubenswrapper[4950]: I0318 20:29:42.436252 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"435fb3ca-c4ad-4b0a-a832-88f4a7203def","Type":"ContainerStarted","Data":"4d44f3b23bf6e12335412f9f0cbabf174474f45f71ff46d762750da9cade1ed6"} Mar 18 20:29:43 crc kubenswrapper[4950]: I0318 20:29:43.452828 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547","Type":"ContainerStarted","Data":"6c2f214a17b18bdd07c3ec6a2b8f14d50b88b9e3ec7996079136b453ab29f051"} Mar 18 20:29:43 crc kubenswrapper[4950]: I0318 20:29:43.498680 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7502f5ea-31ca-4349-bd79-2219bf3903c0" path="/var/lib/kubelet/pods/7502f5ea-31ca-4349-bd79-2219bf3903c0/volumes" Mar 18 20:29:43 crc kubenswrapper[4950]: I0318 20:29:43.640401 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77d56d79f-55rjc"] Mar 18 20:29:43 crc kubenswrapper[4950]: I0318 20:29:43.642027 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:29:43 crc kubenswrapper[4950]: W0318 20:29:43.647847 4950 reflector.go:561] object-"openstack"/"openstack-edpm-ipam": failed to list *v1.ConfigMap: configmaps "openstack-edpm-ipam" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Mar 18 20:29:43 crc kubenswrapper[4950]: E0318 20:29:43.648116 4950 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"openstack-edpm-ipam\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openstack-edpm-ipam\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 18 20:29:43 crc kubenswrapper[4950]: I0318 20:29:43.724070 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77d56d79f-55rjc"] Mar 18 20:29:43 crc kubenswrapper[4950]: I0318 20:29:43.772357 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-openstack-edpm-ipam\") pod \"dnsmasq-dns-77d56d79f-55rjc\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:29:43 crc kubenswrapper[4950]: I0318 20:29:43.772424 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-ovsdbserver-sb\") pod \"dnsmasq-dns-77d56d79f-55rjc\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:29:43 crc kubenswrapper[4950]: I0318 20:29:43.772481 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldfjg\" (UniqueName: \"kubernetes.io/projected/08adefb7-c2e8-4221-be40-29dc95dae4ef-kube-api-access-ldfjg\") pod \"dnsmasq-dns-77d56d79f-55rjc\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:29:43 crc kubenswrapper[4950]: I0318 20:29:43.772539 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-dns-svc\") pod \"dnsmasq-dns-77d56d79f-55rjc\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:29:43 crc kubenswrapper[4950]: I0318 20:29:43.772562 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-config\") pod \"dnsmasq-dns-77d56d79f-55rjc\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:29:43 crc kubenswrapper[4950]: I0318 20:29:43.772583 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-ovsdbserver-nb\") pod \"dnsmasq-dns-77d56d79f-55rjc\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:29:43 crc kubenswrapper[4950]: I0318 20:29:43.874294 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-ovsdbserver-sb\") pod \"dnsmasq-dns-77d56d79f-55rjc\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:29:43 crc kubenswrapper[4950]: I0318 20:29:43.874455 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldfjg\" (UniqueName: \"kubernetes.io/projected/08adefb7-c2e8-4221-be40-29dc95dae4ef-kube-api-access-ldfjg\") pod \"dnsmasq-dns-77d56d79f-55rjc\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:29:43 crc kubenswrapper[4950]: I0318 20:29:43.874538 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-dns-svc\") pod \"dnsmasq-dns-77d56d79f-55rjc\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:29:43 crc kubenswrapper[4950]: I0318 20:29:43.874567 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-config\") pod \"dnsmasq-dns-77d56d79f-55rjc\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:29:43 crc kubenswrapper[4950]: I0318 20:29:43.874593 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-ovsdbserver-nb\") pod \"dnsmasq-dns-77d56d79f-55rjc\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:29:43 crc kubenswrapper[4950]: I0318 20:29:43.874683 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-openstack-edpm-ipam\") pod \"dnsmasq-dns-77d56d79f-55rjc\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:29:43 crc kubenswrapper[4950]: I0318 20:29:43.875337 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-ovsdbserver-sb\") pod \"dnsmasq-dns-77d56d79f-55rjc\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:29:43 crc kubenswrapper[4950]: I0318 20:29:43.875448 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-dns-svc\") pod \"dnsmasq-dns-77d56d79f-55rjc\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:29:43 crc kubenswrapper[4950]: I0318 20:29:43.875725 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-ovsdbserver-nb\") pod \"dnsmasq-dns-77d56d79f-55rjc\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:29:43 crc kubenswrapper[4950]: I0318 20:29:43.876120 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-config\") pod \"dnsmasq-dns-77d56d79f-55rjc\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:29:43 crc kubenswrapper[4950]: I0318 20:29:43.931850 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldfjg\" (UniqueName: \"kubernetes.io/projected/08adefb7-c2e8-4221-be40-29dc95dae4ef-kube-api-access-ldfjg\") pod \"dnsmasq-dns-77d56d79f-55rjc\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:29:44 crc kubenswrapper[4950]: I0318 20:29:44.481927 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"435fb3ca-c4ad-4b0a-a832-88f4a7203def","Type":"ContainerStarted","Data":"7e4e4d3c6e3e2f4708f38f07018e53ac7ff666378a0e3dce5f1b4f204855da9c"} Mar 18 20:29:44 crc kubenswrapper[4950]: E0318 20:29:44.876509 4950 configmap.go:193] Couldn't get configMap openstack/openstack-edpm-ipam: failed to sync configmap cache: timed out waiting for the condition Mar 18 20:29:44 crc kubenswrapper[4950]: E0318 20:29:44.876603 4950 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-openstack-edpm-ipam podName:08adefb7-c2e8-4221-be40-29dc95dae4ef nodeName:}" failed. No retries permitted until 2026-03-18 20:29:45.376578493 +0000 UTC m=+1398.617420361 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "openstack-edpm-ipam" (UniqueName: "kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-openstack-edpm-ipam") pod "dnsmasq-dns-77d56d79f-55rjc" (UID: "08adefb7-c2e8-4221-be40-29dc95dae4ef") : failed to sync configmap cache: timed out waiting for the condition Mar 18 20:29:45 crc kubenswrapper[4950]: I0318 20:29:45.152242 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Mar 18 20:29:45 crc kubenswrapper[4950]: I0318 20:29:45.403633 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-openstack-edpm-ipam\") pod \"dnsmasq-dns-77d56d79f-55rjc\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:29:45 crc kubenswrapper[4950]: I0318 20:29:45.405701 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-openstack-edpm-ipam\") pod \"dnsmasq-dns-77d56d79f-55rjc\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:29:45 crc kubenswrapper[4950]: I0318 20:29:45.467515 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:29:45 crc kubenswrapper[4950]: I0318 20:29:45.953639 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77d56d79f-55rjc"] Mar 18 20:29:45 crc kubenswrapper[4950]: W0318 20:29:45.964615 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08adefb7_c2e8_4221_be40_29dc95dae4ef.slice/crio-efe46f74b5745b7b2a73a937285b6767b45c820821eda38c1c6d77f82f7522b3 WatchSource:0}: Error finding container efe46f74b5745b7b2a73a937285b6767b45c820821eda38c1c6d77f82f7522b3: Status 404 returned error can't find the container with id efe46f74b5745b7b2a73a937285b6767b45c820821eda38c1c6d77f82f7522b3 Mar 18 20:29:46 crc kubenswrapper[4950]: I0318 20:29:46.502911 4950 generic.go:334] "Generic (PLEG): container finished" podID="08adefb7-c2e8-4221-be40-29dc95dae4ef" containerID="4418bcb48659f340c8faec5d4308b37417e9d894523f8668a770147089a55635" exitCode=0 Mar 18 20:29:46 crc kubenswrapper[4950]: I0318 20:29:46.502966 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d56d79f-55rjc" event={"ID":"08adefb7-c2e8-4221-be40-29dc95dae4ef","Type":"ContainerDied","Data":"4418bcb48659f340c8faec5d4308b37417e9d894523f8668a770147089a55635"} Mar 18 20:29:46 crc kubenswrapper[4950]: I0318 20:29:46.503179 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d56d79f-55rjc" event={"ID":"08adefb7-c2e8-4221-be40-29dc95dae4ef","Type":"ContainerStarted","Data":"efe46f74b5745b7b2a73a937285b6767b45c820821eda38c1c6d77f82f7522b3"} Mar 18 20:29:47 crc kubenswrapper[4950]: I0318 20:29:47.520239 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d56d79f-55rjc" event={"ID":"08adefb7-c2e8-4221-be40-29dc95dae4ef","Type":"ContainerStarted","Data":"69a22266979859c648a8e2939b3675d436b13e6c386d29ca1daa7af2f12daf2f"} Mar 18 20:29:47 crc kubenswrapper[4950]: I0318 20:29:47.520707 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:29:47 crc kubenswrapper[4950]: I0318 20:29:47.564171 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77d56d79f-55rjc" podStartSLOduration=4.564148992 podStartE2EDuration="4.564148992s" podCreationTimestamp="2026-03-18 20:29:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:29:47.544790043 +0000 UTC m=+1400.785631931" watchObservedRunningTime="2026-03-18 20:29:47.564148992 +0000 UTC m=+1400.804990870" Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.469648 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.548923 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-776c49c95f-d9656"] Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.549205 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-776c49c95f-d9656" podUID="e7feba91-0f11-4825-8b4c-a24bc2807566" containerName="dnsmasq-dns" containerID="cri-o://a2e15ed2c693569d6521c74966e17c04ee48d9d64e059b59d8a9d355ac2cbd14" gracePeriod=10 Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.741944 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84f6c65c87-mmngf"] Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.760443 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84f6c65c87-mmngf"] Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.760547 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.853616 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzbxj\" (UniqueName: \"kubernetes.io/projected/74de5806-1889-472c-b423-dfbdc2a63020-kube-api-access-hzbxj\") pod \"dnsmasq-dns-84f6c65c87-mmngf\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.853657 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-openstack-edpm-ipam\") pod \"dnsmasq-dns-84f6c65c87-mmngf\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.853684 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-ovsdbserver-sb\") pod \"dnsmasq-dns-84f6c65c87-mmngf\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.853785 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-config\") pod \"dnsmasq-dns-84f6c65c87-mmngf\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.853806 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-dns-svc\") pod \"dnsmasq-dns-84f6c65c87-mmngf\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.853861 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-ovsdbserver-nb\") pod \"dnsmasq-dns-84f6c65c87-mmngf\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.955424 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-config\") pod \"dnsmasq-dns-84f6c65c87-mmngf\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.956530 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-dns-svc\") pod \"dnsmasq-dns-84f6c65c87-mmngf\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.956604 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-ovsdbserver-nb\") pod \"dnsmasq-dns-84f6c65c87-mmngf\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.956646 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzbxj\" (UniqueName: \"kubernetes.io/projected/74de5806-1889-472c-b423-dfbdc2a63020-kube-api-access-hzbxj\") pod \"dnsmasq-dns-84f6c65c87-mmngf\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.956662 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-openstack-edpm-ipam\") pod \"dnsmasq-dns-84f6c65c87-mmngf\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.956692 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-ovsdbserver-sb\") pod \"dnsmasq-dns-84f6c65c87-mmngf\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.957222 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-ovsdbserver-sb\") pod \"dnsmasq-dns-84f6c65c87-mmngf\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.956479 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-config\") pod \"dnsmasq-dns-84f6c65c87-mmngf\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.957743 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-dns-svc\") pod \"dnsmasq-dns-84f6c65c87-mmngf\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.958596 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-ovsdbserver-nb\") pod \"dnsmasq-dns-84f6c65c87-mmngf\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.958887 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-openstack-edpm-ipam\") pod \"dnsmasq-dns-84f6c65c87-mmngf\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 20:29:55 crc kubenswrapper[4950]: I0318 20:29:55.976450 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzbxj\" (UniqueName: \"kubernetes.io/projected/74de5806-1889-472c-b423-dfbdc2a63020-kube-api-access-hzbxj\") pod \"dnsmasq-dns-84f6c65c87-mmngf\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.095293 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.187775 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-776c49c95f-d9656" Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.261290 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-ovsdbserver-nb\") pod \"e7feba91-0f11-4825-8b4c-a24bc2807566\" (UID: \"e7feba91-0f11-4825-8b4c-a24bc2807566\") " Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.261450 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-config\") pod \"e7feba91-0f11-4825-8b4c-a24bc2807566\" (UID: \"e7feba91-0f11-4825-8b4c-a24bc2807566\") " Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.261481 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-ovsdbserver-sb\") pod \"e7feba91-0f11-4825-8b4c-a24bc2807566\" (UID: \"e7feba91-0f11-4825-8b4c-a24bc2807566\") " Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.261520 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-dns-svc\") pod \"e7feba91-0f11-4825-8b4c-a24bc2807566\" (UID: \"e7feba91-0f11-4825-8b4c-a24bc2807566\") " Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.261594 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sppj5\" (UniqueName: \"kubernetes.io/projected/e7feba91-0f11-4825-8b4c-a24bc2807566-kube-api-access-sppj5\") pod \"e7feba91-0f11-4825-8b4c-a24bc2807566\" (UID: \"e7feba91-0f11-4825-8b4c-a24bc2807566\") " Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.284907 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7feba91-0f11-4825-8b4c-a24bc2807566-kube-api-access-sppj5" (OuterVolumeSpecName: "kube-api-access-sppj5") pod "e7feba91-0f11-4825-8b4c-a24bc2807566" (UID: "e7feba91-0f11-4825-8b4c-a24bc2807566"). InnerVolumeSpecName "kube-api-access-sppj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.330966 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-config" (OuterVolumeSpecName: "config") pod "e7feba91-0f11-4825-8b4c-a24bc2807566" (UID: "e7feba91-0f11-4825-8b4c-a24bc2807566"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.331924 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e7feba91-0f11-4825-8b4c-a24bc2807566" (UID: "e7feba91-0f11-4825-8b4c-a24bc2807566"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.333067 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e7feba91-0f11-4825-8b4c-a24bc2807566" (UID: "e7feba91-0f11-4825-8b4c-a24bc2807566"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.335774 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e7feba91-0f11-4825-8b4c-a24bc2807566" (UID: "e7feba91-0f11-4825-8b4c-a24bc2807566"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.363750 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.363775 4950 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.363786 4950 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.363795 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sppj5\" (UniqueName: \"kubernetes.io/projected/e7feba91-0f11-4825-8b4c-a24bc2807566-kube-api-access-sppj5\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.363807 4950 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7feba91-0f11-4825-8b4c-a24bc2807566-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.593278 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84f6c65c87-mmngf"] Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.640397 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" event={"ID":"74de5806-1889-472c-b423-dfbdc2a63020","Type":"ContainerStarted","Data":"c217b96a0bde9026a07c7184e2b4075f8923056f16782388425c19c5532aaab3"} Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.642141 4950 generic.go:334] "Generic (PLEG): container finished" podID="e7feba91-0f11-4825-8b4c-a24bc2807566" containerID="a2e15ed2c693569d6521c74966e17c04ee48d9d64e059b59d8a9d355ac2cbd14" exitCode=0 Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.642173 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-776c49c95f-d9656" event={"ID":"e7feba91-0f11-4825-8b4c-a24bc2807566","Type":"ContainerDied","Data":"a2e15ed2c693569d6521c74966e17c04ee48d9d64e059b59d8a9d355ac2cbd14"} Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.642187 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-776c49c95f-d9656" event={"ID":"e7feba91-0f11-4825-8b4c-a24bc2807566","Type":"ContainerDied","Data":"9ff88862cbebfcf5f39958385155b309205c42a27c9067d471683a08c158315e"} Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.642207 4950 scope.go:117] "RemoveContainer" containerID="a2e15ed2c693569d6521c74966e17c04ee48d9d64e059b59d8a9d355ac2cbd14" Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.642306 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-776c49c95f-d9656" Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.681698 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-776c49c95f-d9656"] Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.705163 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-776c49c95f-d9656"] Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.705866 4950 scope.go:117] "RemoveContainer" containerID="6d08377bb6463505044677f3066c51b22fb5f47c8d7c25b14a1faa87637e6f89" Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.799758 4950 scope.go:117] "RemoveContainer" containerID="a2e15ed2c693569d6521c74966e17c04ee48d9d64e059b59d8a9d355ac2cbd14" Mar 18 20:29:56 crc kubenswrapper[4950]: E0318 20:29:56.800478 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2e15ed2c693569d6521c74966e17c04ee48d9d64e059b59d8a9d355ac2cbd14\": container with ID starting with a2e15ed2c693569d6521c74966e17c04ee48d9d64e059b59d8a9d355ac2cbd14 not found: ID does not exist" containerID="a2e15ed2c693569d6521c74966e17c04ee48d9d64e059b59d8a9d355ac2cbd14" Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.800509 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2e15ed2c693569d6521c74966e17c04ee48d9d64e059b59d8a9d355ac2cbd14"} err="failed to get container status \"a2e15ed2c693569d6521c74966e17c04ee48d9d64e059b59d8a9d355ac2cbd14\": rpc error: code = NotFound desc = could not find container \"a2e15ed2c693569d6521c74966e17c04ee48d9d64e059b59d8a9d355ac2cbd14\": container with ID starting with a2e15ed2c693569d6521c74966e17c04ee48d9d64e059b59d8a9d355ac2cbd14 not found: ID does not exist" Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.800528 4950 scope.go:117] "RemoveContainer" containerID="6d08377bb6463505044677f3066c51b22fb5f47c8d7c25b14a1faa87637e6f89" Mar 18 20:29:56 crc kubenswrapper[4950]: E0318 20:29:56.800975 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d08377bb6463505044677f3066c51b22fb5f47c8d7c25b14a1faa87637e6f89\": container with ID starting with 6d08377bb6463505044677f3066c51b22fb5f47c8d7c25b14a1faa87637e6f89 not found: ID does not exist" containerID="6d08377bb6463505044677f3066c51b22fb5f47c8d7c25b14a1faa87637e6f89" Mar 18 20:29:56 crc kubenswrapper[4950]: I0318 20:29:56.801025 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d08377bb6463505044677f3066c51b22fb5f47c8d7c25b14a1faa87637e6f89"} err="failed to get container status \"6d08377bb6463505044677f3066c51b22fb5f47c8d7c25b14a1faa87637e6f89\": rpc error: code = NotFound desc = could not find container \"6d08377bb6463505044677f3066c51b22fb5f47c8d7c25b14a1faa87637e6f89\": container with ID starting with 6d08377bb6463505044677f3066c51b22fb5f47c8d7c25b14a1faa87637e6f89 not found: ID does not exist" Mar 18 20:29:57 crc kubenswrapper[4950]: I0318 20:29:57.493473 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7feba91-0f11-4825-8b4c-a24bc2807566" path="/var/lib/kubelet/pods/e7feba91-0f11-4825-8b4c-a24bc2807566/volumes" Mar 18 20:29:57 crc kubenswrapper[4950]: I0318 20:29:57.650744 4950 generic.go:334] "Generic (PLEG): container finished" podID="74de5806-1889-472c-b423-dfbdc2a63020" containerID="654f0d6c8a9b433b09cd88437a2df12e449a86f79c0b22d19f564e39e30c5d5c" exitCode=0 Mar 18 20:29:57 crc kubenswrapper[4950]: I0318 20:29:57.650831 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" event={"ID":"74de5806-1889-472c-b423-dfbdc2a63020","Type":"ContainerDied","Data":"654f0d6c8a9b433b09cd88437a2df12e449a86f79c0b22d19f564e39e30c5d5c"} Mar 18 20:29:57 crc kubenswrapper[4950]: I0318 20:29:57.728342 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 18 20:29:58 crc kubenswrapper[4950]: I0318 20:29:58.665275 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" event={"ID":"74de5806-1889-472c-b423-dfbdc2a63020","Type":"ContainerStarted","Data":"6ebacc2249dba5b2f473a4da950e2dd3ee04679510a00ffe911f4046844db248"} Mar 18 20:29:58 crc kubenswrapper[4950]: I0318 20:29:58.665588 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 20:29:58 crc kubenswrapper[4950]: I0318 20:29:58.699176 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" podStartSLOduration=3.699146158 podStartE2EDuration="3.699146158s" podCreationTimestamp="2026-03-18 20:29:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:29:58.684498873 +0000 UTC m=+1411.925340771" watchObservedRunningTime="2026-03-18 20:29:58.699146158 +0000 UTC m=+1411.939988056" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.160933 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564430-hrm5q"] Mar 18 20:30:00 crc kubenswrapper[4950]: E0318 20:30:00.162025 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7feba91-0f11-4825-8b4c-a24bc2807566" containerName="dnsmasq-dns" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.162046 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7feba91-0f11-4825-8b4c-a24bc2807566" containerName="dnsmasq-dns" Mar 18 20:30:00 crc kubenswrapper[4950]: E0318 20:30:00.162069 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7feba91-0f11-4825-8b4c-a24bc2807566" containerName="init" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.162082 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7feba91-0f11-4825-8b4c-a24bc2807566" containerName="init" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.162397 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7feba91-0f11-4825-8b4c-a24bc2807566" containerName="dnsmasq-dns" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.163474 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564430-hrm5q" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.182562 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564430-hrm5q"] Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.202467 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.202586 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.233049 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86rhd\" (UniqueName: \"kubernetes.io/projected/87931a42-2463-4633-a253-f7a01eee1935-kube-api-access-86rhd\") pod \"collect-profiles-29564430-hrm5q\" (UID: \"87931a42-2463-4633-a253-f7a01eee1935\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564430-hrm5q" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.233131 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87931a42-2463-4633-a253-f7a01eee1935-config-volume\") pod \"collect-profiles-29564430-hrm5q\" (UID: \"87931a42-2463-4633-a253-f7a01eee1935\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564430-hrm5q" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.233173 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/87931a42-2463-4633-a253-f7a01eee1935-secret-volume\") pod \"collect-profiles-29564430-hrm5q\" (UID: \"87931a42-2463-4633-a253-f7a01eee1935\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564430-hrm5q" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.256764 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564430-gxf7d"] Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.258029 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564430-gxf7d" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.266124 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.266371 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.266614 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.267791 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564430-gxf7d"] Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.334323 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86rhd\" (UniqueName: \"kubernetes.io/projected/87931a42-2463-4633-a253-f7a01eee1935-kube-api-access-86rhd\") pod \"collect-profiles-29564430-hrm5q\" (UID: \"87931a42-2463-4633-a253-f7a01eee1935\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564430-hrm5q" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.334420 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87931a42-2463-4633-a253-f7a01eee1935-config-volume\") pod \"collect-profiles-29564430-hrm5q\" (UID: \"87931a42-2463-4633-a253-f7a01eee1935\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564430-hrm5q" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.334459 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74fs9\" (UniqueName: \"kubernetes.io/projected/d021afc1-77fb-4064-8b4c-a81f25c293eb-kube-api-access-74fs9\") pod \"auto-csr-approver-29564430-gxf7d\" (UID: \"d021afc1-77fb-4064-8b4c-a81f25c293eb\") " pod="openshift-infra/auto-csr-approver-29564430-gxf7d" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.334488 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/87931a42-2463-4633-a253-f7a01eee1935-secret-volume\") pod \"collect-profiles-29564430-hrm5q\" (UID: \"87931a42-2463-4633-a253-f7a01eee1935\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564430-hrm5q" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.335458 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87931a42-2463-4633-a253-f7a01eee1935-config-volume\") pod \"collect-profiles-29564430-hrm5q\" (UID: \"87931a42-2463-4633-a253-f7a01eee1935\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564430-hrm5q" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.349950 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/87931a42-2463-4633-a253-f7a01eee1935-secret-volume\") pod \"collect-profiles-29564430-hrm5q\" (UID: \"87931a42-2463-4633-a253-f7a01eee1935\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564430-hrm5q" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.350681 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86rhd\" (UniqueName: \"kubernetes.io/projected/87931a42-2463-4633-a253-f7a01eee1935-kube-api-access-86rhd\") pod \"collect-profiles-29564430-hrm5q\" (UID: \"87931a42-2463-4633-a253-f7a01eee1935\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564430-hrm5q" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.436026 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74fs9\" (UniqueName: \"kubernetes.io/projected/d021afc1-77fb-4064-8b4c-a81f25c293eb-kube-api-access-74fs9\") pod \"auto-csr-approver-29564430-gxf7d\" (UID: \"d021afc1-77fb-4064-8b4c-a81f25c293eb\") " pod="openshift-infra/auto-csr-approver-29564430-gxf7d" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.456312 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74fs9\" (UniqueName: \"kubernetes.io/projected/d021afc1-77fb-4064-8b4c-a81f25c293eb-kube-api-access-74fs9\") pod \"auto-csr-approver-29564430-gxf7d\" (UID: \"d021afc1-77fb-4064-8b4c-a81f25c293eb\") " pod="openshift-infra/auto-csr-approver-29564430-gxf7d" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.522641 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564430-hrm5q" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.578071 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564430-gxf7d" Mar 18 20:30:00 crc kubenswrapper[4950]: I0318 20:30:00.981612 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564430-gxf7d"] Mar 18 20:30:01 crc kubenswrapper[4950]: I0318 20:30:01.168950 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564430-hrm5q"] Mar 18 20:30:01 crc kubenswrapper[4950]: W0318 20:30:01.171618 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87931a42_2463_4633_a253_f7a01eee1935.slice/crio-05f3b93553b7b22b1246c0f55896545290f8192928cfe3aacab3ee124cdbc503 WatchSource:0}: Error finding container 05f3b93553b7b22b1246c0f55896545290f8192928cfe3aacab3ee124cdbc503: Status 404 returned error can't find the container with id 05f3b93553b7b22b1246c0f55896545290f8192928cfe3aacab3ee124cdbc503 Mar 18 20:30:01 crc kubenswrapper[4950]: I0318 20:30:01.694328 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564430-gxf7d" event={"ID":"d021afc1-77fb-4064-8b4c-a81f25c293eb","Type":"ContainerStarted","Data":"e29e2064a2cab061af0035c2a86cc425b4f70477dfcccfb1ec4dc782ff8b5f85"} Mar 18 20:30:01 crc kubenswrapper[4950]: I0318 20:30:01.695672 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564430-hrm5q" event={"ID":"87931a42-2463-4633-a253-f7a01eee1935","Type":"ContainerStarted","Data":"18d06b3d74085c796f46cf0af33ebf90c012776cc061dba2a1c1175506c6e48e"} Mar 18 20:30:01 crc kubenswrapper[4950]: I0318 20:30:01.695697 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564430-hrm5q" event={"ID":"87931a42-2463-4633-a253-f7a01eee1935","Type":"ContainerStarted","Data":"05f3b93553b7b22b1246c0f55896545290f8192928cfe3aacab3ee124cdbc503"} Mar 18 20:30:01 crc kubenswrapper[4950]: I0318 20:30:01.712607 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29564430-hrm5q" podStartSLOduration=1.712586993 podStartE2EDuration="1.712586993s" podCreationTimestamp="2026-03-18 20:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:30:01.711923395 +0000 UTC m=+1414.952765273" watchObservedRunningTime="2026-03-18 20:30:01.712586993 +0000 UTC m=+1414.953428861" Mar 18 20:30:02 crc kubenswrapper[4950]: I0318 20:30:02.707098 4950 generic.go:334] "Generic (PLEG): container finished" podID="87931a42-2463-4633-a253-f7a01eee1935" containerID="18d06b3d74085c796f46cf0af33ebf90c012776cc061dba2a1c1175506c6e48e" exitCode=0 Mar 18 20:30:02 crc kubenswrapper[4950]: I0318 20:30:02.708947 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564430-hrm5q" event={"ID":"87931a42-2463-4633-a253-f7a01eee1935","Type":"ContainerDied","Data":"18d06b3d74085c796f46cf0af33ebf90c012776cc061dba2a1c1175506c6e48e"} Mar 18 20:30:05 crc kubenswrapper[4950]: I0318 20:30:04.113342 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564430-hrm5q" Mar 18 20:30:05 crc kubenswrapper[4950]: I0318 20:30:04.314251 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/87931a42-2463-4633-a253-f7a01eee1935-secret-volume\") pod \"87931a42-2463-4633-a253-f7a01eee1935\" (UID: \"87931a42-2463-4633-a253-f7a01eee1935\") " Mar 18 20:30:05 crc kubenswrapper[4950]: I0318 20:30:04.314383 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86rhd\" (UniqueName: \"kubernetes.io/projected/87931a42-2463-4633-a253-f7a01eee1935-kube-api-access-86rhd\") pod \"87931a42-2463-4633-a253-f7a01eee1935\" (UID: \"87931a42-2463-4633-a253-f7a01eee1935\") " Mar 18 20:30:05 crc kubenswrapper[4950]: I0318 20:30:04.314407 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87931a42-2463-4633-a253-f7a01eee1935-config-volume\") pod \"87931a42-2463-4633-a253-f7a01eee1935\" (UID: \"87931a42-2463-4633-a253-f7a01eee1935\") " Mar 18 20:30:05 crc kubenswrapper[4950]: I0318 20:30:04.314974 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87931a42-2463-4633-a253-f7a01eee1935-config-volume" (OuterVolumeSpecName: "config-volume") pod "87931a42-2463-4633-a253-f7a01eee1935" (UID: "87931a42-2463-4633-a253-f7a01eee1935"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:30:05 crc kubenswrapper[4950]: I0318 20:30:04.416265 4950 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87931a42-2463-4633-a253-f7a01eee1935-config-volume\") on node \"crc\" DevicePath \"\"" Mar 18 20:30:05 crc kubenswrapper[4950]: I0318 20:30:04.613024 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87931a42-2463-4633-a253-f7a01eee1935-kube-api-access-86rhd" (OuterVolumeSpecName: "kube-api-access-86rhd") pod "87931a42-2463-4633-a253-f7a01eee1935" (UID: "87931a42-2463-4633-a253-f7a01eee1935"). InnerVolumeSpecName "kube-api-access-86rhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:30:05 crc kubenswrapper[4950]: I0318 20:30:04.613436 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87931a42-2463-4633-a253-f7a01eee1935-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "87931a42-2463-4633-a253-f7a01eee1935" (UID: "87931a42-2463-4633-a253-f7a01eee1935"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:30:05 crc kubenswrapper[4950]: I0318 20:30:04.619557 4950 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/87931a42-2463-4633-a253-f7a01eee1935-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 18 20:30:05 crc kubenswrapper[4950]: I0318 20:30:04.619587 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86rhd\" (UniqueName: \"kubernetes.io/projected/87931a42-2463-4633-a253-f7a01eee1935-kube-api-access-86rhd\") on node \"crc\" DevicePath \"\"" Mar 18 20:30:05 crc kubenswrapper[4950]: I0318 20:30:04.733812 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564430-hrm5q" event={"ID":"87931a42-2463-4633-a253-f7a01eee1935","Type":"ContainerDied","Data":"05f3b93553b7b22b1246c0f55896545290f8192928cfe3aacab3ee124cdbc503"} Mar 18 20:30:05 crc kubenswrapper[4950]: I0318 20:30:04.734131 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05f3b93553b7b22b1246c0f55896545290f8192928cfe3aacab3ee124cdbc503" Mar 18 20:30:05 crc kubenswrapper[4950]: I0318 20:30:04.733952 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564430-hrm5q" Mar 18 20:30:05 crc kubenswrapper[4950]: I0318 20:30:05.751524 4950 generic.go:334] "Generic (PLEG): container finished" podID="d021afc1-77fb-4064-8b4c-a81f25c293eb" containerID="67ac8553eb5aa5ab32a8c742ca36189fb2548b29cc3cfaab3867d8931285085c" exitCode=0 Mar 18 20:30:05 crc kubenswrapper[4950]: I0318 20:30:05.751596 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564430-gxf7d" event={"ID":"d021afc1-77fb-4064-8b4c-a81f25c293eb","Type":"ContainerDied","Data":"67ac8553eb5aa5ab32a8c742ca36189fb2548b29cc3cfaab3867d8931285085c"} Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.097540 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.214335 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77d56d79f-55rjc"] Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.214653 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77d56d79f-55rjc" podUID="08adefb7-c2e8-4221-be40-29dc95dae4ef" containerName="dnsmasq-dns" containerID="cri-o://69a22266979859c648a8e2939b3675d436b13e6c386d29ca1daa7af2f12daf2f" gracePeriod=10 Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.675255 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.760323 4950 generic.go:334] "Generic (PLEG): container finished" podID="08adefb7-c2e8-4221-be40-29dc95dae4ef" containerID="69a22266979859c648a8e2939b3675d436b13e6c386d29ca1daa7af2f12daf2f" exitCode=0 Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.760595 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77d56d79f-55rjc" Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.762095 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d56d79f-55rjc" event={"ID":"08adefb7-c2e8-4221-be40-29dc95dae4ef","Type":"ContainerDied","Data":"69a22266979859c648a8e2939b3675d436b13e6c386d29ca1daa7af2f12daf2f"} Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.762126 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77d56d79f-55rjc" event={"ID":"08adefb7-c2e8-4221-be40-29dc95dae4ef","Type":"ContainerDied","Data":"efe46f74b5745b7b2a73a937285b6767b45c820821eda38c1c6d77f82f7522b3"} Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.762142 4950 scope.go:117] "RemoveContainer" containerID="69a22266979859c648a8e2939b3675d436b13e6c386d29ca1daa7af2f12daf2f" Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.777657 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-config\") pod \"08adefb7-c2e8-4221-be40-29dc95dae4ef\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.777703 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-openstack-edpm-ipam\") pod \"08adefb7-c2e8-4221-be40-29dc95dae4ef\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.777815 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldfjg\" (UniqueName: \"kubernetes.io/projected/08adefb7-c2e8-4221-be40-29dc95dae4ef-kube-api-access-ldfjg\") pod \"08adefb7-c2e8-4221-be40-29dc95dae4ef\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.777844 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-ovsdbserver-nb\") pod \"08adefb7-c2e8-4221-be40-29dc95dae4ef\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.777967 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-ovsdbserver-sb\") pod \"08adefb7-c2e8-4221-be40-29dc95dae4ef\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.777999 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-dns-svc\") pod \"08adefb7-c2e8-4221-be40-29dc95dae4ef\" (UID: \"08adefb7-c2e8-4221-be40-29dc95dae4ef\") " Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.782211 4950 scope.go:117] "RemoveContainer" containerID="4418bcb48659f340c8faec5d4308b37417e9d894523f8668a770147089a55635" Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.798820 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08adefb7-c2e8-4221-be40-29dc95dae4ef-kube-api-access-ldfjg" (OuterVolumeSpecName: "kube-api-access-ldfjg") pod "08adefb7-c2e8-4221-be40-29dc95dae4ef" (UID: "08adefb7-c2e8-4221-be40-29dc95dae4ef"). InnerVolumeSpecName "kube-api-access-ldfjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.832345 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "08adefb7-c2e8-4221-be40-29dc95dae4ef" (UID: "08adefb7-c2e8-4221-be40-29dc95dae4ef"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.834594 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-config" (OuterVolumeSpecName: "config") pod "08adefb7-c2e8-4221-be40-29dc95dae4ef" (UID: "08adefb7-c2e8-4221-be40-29dc95dae4ef"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.843030 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "08adefb7-c2e8-4221-be40-29dc95dae4ef" (UID: "08adefb7-c2e8-4221-be40-29dc95dae4ef"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.846035 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "08adefb7-c2e8-4221-be40-29dc95dae4ef" (UID: "08adefb7-c2e8-4221-be40-29dc95dae4ef"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.847370 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "08adefb7-c2e8-4221-be40-29dc95dae4ef" (UID: "08adefb7-c2e8-4221-be40-29dc95dae4ef"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.880477 4950 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.880509 4950 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.880519 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-config\") on node \"crc\" DevicePath \"\"" Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.880529 4950 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.880539 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldfjg\" (UniqueName: \"kubernetes.io/projected/08adefb7-c2e8-4221-be40-29dc95dae4ef-kube-api-access-ldfjg\") on node \"crc\" DevicePath \"\"" Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.880547 4950 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08adefb7-c2e8-4221-be40-29dc95dae4ef-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.900855 4950 scope.go:117] "RemoveContainer" containerID="69a22266979859c648a8e2939b3675d436b13e6c386d29ca1daa7af2f12daf2f" Mar 18 20:30:06 crc kubenswrapper[4950]: E0318 20:30:06.901288 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69a22266979859c648a8e2939b3675d436b13e6c386d29ca1daa7af2f12daf2f\": container with ID starting with 69a22266979859c648a8e2939b3675d436b13e6c386d29ca1daa7af2f12daf2f not found: ID does not exist" containerID="69a22266979859c648a8e2939b3675d436b13e6c386d29ca1daa7af2f12daf2f" Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.901316 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69a22266979859c648a8e2939b3675d436b13e6c386d29ca1daa7af2f12daf2f"} err="failed to get container status \"69a22266979859c648a8e2939b3675d436b13e6c386d29ca1daa7af2f12daf2f\": rpc error: code = NotFound desc = could not find container \"69a22266979859c648a8e2939b3675d436b13e6c386d29ca1daa7af2f12daf2f\": container with ID starting with 69a22266979859c648a8e2939b3675d436b13e6c386d29ca1daa7af2f12daf2f not found: ID does not exist" Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.901335 4950 scope.go:117] "RemoveContainer" containerID="4418bcb48659f340c8faec5d4308b37417e9d894523f8668a770147089a55635" Mar 18 20:30:06 crc kubenswrapper[4950]: E0318 20:30:06.901768 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4418bcb48659f340c8faec5d4308b37417e9d894523f8668a770147089a55635\": container with ID starting with 4418bcb48659f340c8faec5d4308b37417e9d894523f8668a770147089a55635 not found: ID does not exist" containerID="4418bcb48659f340c8faec5d4308b37417e9d894523f8668a770147089a55635" Mar 18 20:30:06 crc kubenswrapper[4950]: I0318 20:30:06.901785 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4418bcb48659f340c8faec5d4308b37417e9d894523f8668a770147089a55635"} err="failed to get container status \"4418bcb48659f340c8faec5d4308b37417e9d894523f8668a770147089a55635\": rpc error: code = NotFound desc = could not find container \"4418bcb48659f340c8faec5d4308b37417e9d894523f8668a770147089a55635\": container with ID starting with 4418bcb48659f340c8faec5d4308b37417e9d894523f8668a770147089a55635 not found: ID does not exist" Mar 18 20:30:07 crc kubenswrapper[4950]: I0318 20:30:07.002915 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564430-gxf7d" Mar 18 20:30:07 crc kubenswrapper[4950]: I0318 20:30:07.085449 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74fs9\" (UniqueName: \"kubernetes.io/projected/d021afc1-77fb-4064-8b4c-a81f25c293eb-kube-api-access-74fs9\") pod \"d021afc1-77fb-4064-8b4c-a81f25c293eb\" (UID: \"d021afc1-77fb-4064-8b4c-a81f25c293eb\") " Mar 18 20:30:07 crc kubenswrapper[4950]: I0318 20:30:07.093328 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d021afc1-77fb-4064-8b4c-a81f25c293eb-kube-api-access-74fs9" (OuterVolumeSpecName: "kube-api-access-74fs9") pod "d021afc1-77fb-4064-8b4c-a81f25c293eb" (UID: "d021afc1-77fb-4064-8b4c-a81f25c293eb"). InnerVolumeSpecName "kube-api-access-74fs9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:30:07 crc kubenswrapper[4950]: I0318 20:30:07.103174 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77d56d79f-55rjc"] Mar 18 20:30:07 crc kubenswrapper[4950]: I0318 20:30:07.111121 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77d56d79f-55rjc"] Mar 18 20:30:07 crc kubenswrapper[4950]: I0318 20:30:07.189163 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74fs9\" (UniqueName: \"kubernetes.io/projected/d021afc1-77fb-4064-8b4c-a81f25c293eb-kube-api-access-74fs9\") on node \"crc\" DevicePath \"\"" Mar 18 20:30:07 crc kubenswrapper[4950]: I0318 20:30:07.496144 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08adefb7-c2e8-4221-be40-29dc95dae4ef" path="/var/lib/kubelet/pods/08adefb7-c2e8-4221-be40-29dc95dae4ef/volumes" Mar 18 20:30:07 crc kubenswrapper[4950]: I0318 20:30:07.769480 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564430-gxf7d" Mar 18 20:30:07 crc kubenswrapper[4950]: I0318 20:30:07.770190 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564430-gxf7d" event={"ID":"d021afc1-77fb-4064-8b4c-a81f25c293eb","Type":"ContainerDied","Data":"e29e2064a2cab061af0035c2a86cc425b4f70477dfcccfb1ec4dc782ff8b5f85"} Mar 18 20:30:07 crc kubenswrapper[4950]: I0318 20:30:07.770229 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e29e2064a2cab061af0035c2a86cc425b4f70477dfcccfb1ec4dc782ff8b5f85" Mar 18 20:30:08 crc kubenswrapper[4950]: I0318 20:30:08.092209 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564424-q58sw"] Mar 18 20:30:08 crc kubenswrapper[4950]: I0318 20:30:08.100459 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564424-q58sw"] Mar 18 20:30:09 crc kubenswrapper[4950]: I0318 20:30:09.493913 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ff7fb3b-ac2f-4718-8c93-65b9a6291a20" path="/var/lib/kubelet/pods/2ff7fb3b-ac2f-4718-8c93-65b9a6291a20/volumes" Mar 18 20:30:11 crc kubenswrapper[4950]: I0318 20:30:11.888475 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv"] Mar 18 20:30:11 crc kubenswrapper[4950]: E0318 20:30:11.889193 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08adefb7-c2e8-4221-be40-29dc95dae4ef" containerName="init" Mar 18 20:30:11 crc kubenswrapper[4950]: I0318 20:30:11.889207 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="08adefb7-c2e8-4221-be40-29dc95dae4ef" containerName="init" Mar 18 20:30:11 crc kubenswrapper[4950]: E0318 20:30:11.889220 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87931a42-2463-4633-a253-f7a01eee1935" containerName="collect-profiles" Mar 18 20:30:11 crc kubenswrapper[4950]: I0318 20:30:11.889229 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="87931a42-2463-4633-a253-f7a01eee1935" containerName="collect-profiles" Mar 18 20:30:11 crc kubenswrapper[4950]: E0318 20:30:11.889252 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d021afc1-77fb-4064-8b4c-a81f25c293eb" containerName="oc" Mar 18 20:30:11 crc kubenswrapper[4950]: I0318 20:30:11.889260 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="d021afc1-77fb-4064-8b4c-a81f25c293eb" containerName="oc" Mar 18 20:30:11 crc kubenswrapper[4950]: E0318 20:30:11.889287 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08adefb7-c2e8-4221-be40-29dc95dae4ef" containerName="dnsmasq-dns" Mar 18 20:30:11 crc kubenswrapper[4950]: I0318 20:30:11.889295 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="08adefb7-c2e8-4221-be40-29dc95dae4ef" containerName="dnsmasq-dns" Mar 18 20:30:11 crc kubenswrapper[4950]: I0318 20:30:11.889499 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="08adefb7-c2e8-4221-be40-29dc95dae4ef" containerName="dnsmasq-dns" Mar 18 20:30:11 crc kubenswrapper[4950]: I0318 20:30:11.889511 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="87931a42-2463-4633-a253-f7a01eee1935" containerName="collect-profiles" Mar 18 20:30:11 crc kubenswrapper[4950]: I0318 20:30:11.889531 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="d021afc1-77fb-4064-8b4c-a81f25c293eb" containerName="oc" Mar 18 20:30:11 crc kubenswrapper[4950]: I0318 20:30:11.890149 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv" Mar 18 20:30:11 crc kubenswrapper[4950]: I0318 20:30:11.895012 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:30:11 crc kubenswrapper[4950]: I0318 20:30:11.895114 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:30:11 crc kubenswrapper[4950]: I0318 20:30:11.895478 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:30:11 crc kubenswrapper[4950]: I0318 20:30:11.895982 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:30:11 crc kubenswrapper[4950]: I0318 20:30:11.907717 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv"] Mar 18 20:30:11 crc kubenswrapper[4950]: I0318 20:30:11.980781 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv\" (UID: \"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv" Mar 18 20:30:11 crc kubenswrapper[4950]: I0318 20:30:11.980925 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv\" (UID: \"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv" Mar 18 20:30:11 crc kubenswrapper[4950]: I0318 20:30:11.980981 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv\" (UID: \"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv" Mar 18 20:30:11 crc kubenswrapper[4950]: I0318 20:30:11.981032 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l68j\" (UniqueName: \"kubernetes.io/projected/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-kube-api-access-4l68j\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv\" (UID: \"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv" Mar 18 20:30:12 crc kubenswrapper[4950]: I0318 20:30:12.084712 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv\" (UID: \"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv" Mar 18 20:30:12 crc kubenswrapper[4950]: I0318 20:30:12.084801 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv\" (UID: \"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv" Mar 18 20:30:12 crc kubenswrapper[4950]: I0318 20:30:12.084882 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l68j\" (UniqueName: \"kubernetes.io/projected/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-kube-api-access-4l68j\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv\" (UID: \"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv" Mar 18 20:30:12 crc kubenswrapper[4950]: I0318 20:30:12.084948 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv\" (UID: \"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv" Mar 18 20:30:12 crc kubenswrapper[4950]: I0318 20:30:12.092102 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv\" (UID: \"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv" Mar 18 20:30:12 crc kubenswrapper[4950]: I0318 20:30:12.093859 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv\" (UID: \"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv" Mar 18 20:30:12 crc kubenswrapper[4950]: I0318 20:30:12.098173 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv\" (UID: \"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv" Mar 18 20:30:12 crc kubenswrapper[4950]: I0318 20:30:12.100871 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l68j\" (UniqueName: \"kubernetes.io/projected/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-kube-api-access-4l68j\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv\" (UID: \"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv" Mar 18 20:30:12 crc kubenswrapper[4950]: I0318 20:30:12.243311 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv" Mar 18 20:30:12 crc kubenswrapper[4950]: I0318 20:30:12.825529 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv"] Mar 18 20:30:13 crc kubenswrapper[4950]: I0318 20:30:13.833043 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv" event={"ID":"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0","Type":"ContainerStarted","Data":"dd34109a4ff124b97ebb801e54f6ea4a30b467af3e6367f31dd0ef8e3f4a7538"} Mar 18 20:30:15 crc kubenswrapper[4950]: I0318 20:30:15.856851 4950 generic.go:334] "Generic (PLEG): container finished" podID="27be57d1-fc6d-4fe7-a6b3-c7276d5ce547" containerID="6c2f214a17b18bdd07c3ec6a2b8f14d50b88b9e3ec7996079136b453ab29f051" exitCode=0 Mar 18 20:30:15 crc kubenswrapper[4950]: I0318 20:30:15.858055 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547","Type":"ContainerDied","Data":"6c2f214a17b18bdd07c3ec6a2b8f14d50b88b9e3ec7996079136b453ab29f051"} Mar 18 20:30:16 crc kubenswrapper[4950]: I0318 20:30:16.869041 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"435fb3ca-c4ad-4b0a-a832-88f4a7203def","Type":"ContainerDied","Data":"7e4e4d3c6e3e2f4708f38f07018e53ac7ff666378a0e3dce5f1b4f204855da9c"} Mar 18 20:30:16 crc kubenswrapper[4950]: I0318 20:30:16.869068 4950 generic.go:334] "Generic (PLEG): container finished" podID="435fb3ca-c4ad-4b0a-a832-88f4a7203def" containerID="7e4e4d3c6e3e2f4708f38f07018e53ac7ff666378a0e3dce5f1b4f204855da9c" exitCode=0 Mar 18 20:30:22 crc kubenswrapper[4950]: I0318 20:30:22.922353 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv" event={"ID":"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0","Type":"ContainerStarted","Data":"177a725128594a13952a18229398d0c915fcbbb80992adf5d9a78770c3435cb3"} Mar 18 20:30:22 crc kubenswrapper[4950]: I0318 20:30:22.925593 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"435fb3ca-c4ad-4b0a-a832-88f4a7203def","Type":"ContainerStarted","Data":"a73c1f88e3e65ecda4cd7d7c87227e119c0859ffb6d560f648314a6de3662069"} Mar 18 20:30:22 crc kubenswrapper[4950]: I0318 20:30:22.925967 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:30:22 crc kubenswrapper[4950]: I0318 20:30:22.927821 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"27be57d1-fc6d-4fe7-a6b3-c7276d5ce547","Type":"ContainerStarted","Data":"c0dd9f3a1890f68002bf81d9b5cd156b9f3dc274f7d5d6afffccd173de44e492"} Mar 18 20:30:22 crc kubenswrapper[4950]: I0318 20:30:22.928874 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 18 20:30:22 crc kubenswrapper[4950]: I0318 20:30:22.984028 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=42.984006257 podStartE2EDuration="42.984006257s" podCreationTimestamp="2026-03-18 20:29:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:30:22.976247353 +0000 UTC m=+1436.217089221" watchObservedRunningTime="2026-03-18 20:30:22.984006257 +0000 UTC m=+1436.224848125" Mar 18 20:30:22 crc kubenswrapper[4950]: I0318 20:30:22.995654 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv" podStartSLOduration=2.913210045 podStartE2EDuration="11.995636122s" podCreationTimestamp="2026-03-18 20:30:11 +0000 UTC" firstStartedPulling="2026-03-18 20:30:12.83666387 +0000 UTC m=+1426.077505738" lastFinishedPulling="2026-03-18 20:30:21.919089937 +0000 UTC m=+1435.159931815" observedRunningTime="2026-03-18 20:30:22.952601881 +0000 UTC m=+1436.193443759" watchObservedRunningTime="2026-03-18 20:30:22.995636122 +0000 UTC m=+1436.236477990" Mar 18 20:30:23 crc kubenswrapper[4950]: I0318 20:30:23.014956 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=42.01493568 podStartE2EDuration="42.01493568s" podCreationTimestamp="2026-03-18 20:29:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:30:23.011672644 +0000 UTC m=+1436.252514512" watchObservedRunningTime="2026-03-18 20:30:23.01493568 +0000 UTC m=+1436.255777548" Mar 18 20:30:36 crc kubenswrapper[4950]: I0318 20:30:36.039963 4950 generic.go:334] "Generic (PLEG): container finished" podID="7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0" containerID="177a725128594a13952a18229398d0c915fcbbb80992adf5d9a78770c3435cb3" exitCode=0 Mar 18 20:30:36 crc kubenswrapper[4950]: I0318 20:30:36.040032 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv" event={"ID":"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0","Type":"ContainerDied","Data":"177a725128594a13952a18229398d0c915fcbbb80992adf5d9a78770c3435cb3"} Mar 18 20:30:37 crc kubenswrapper[4950]: I0318 20:30:37.586159 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv" Mar 18 20:30:37 crc kubenswrapper[4950]: I0318 20:30:37.736185 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4l68j\" (UniqueName: \"kubernetes.io/projected/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-kube-api-access-4l68j\") pod \"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0\" (UID: \"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0\") " Mar 18 20:30:37 crc kubenswrapper[4950]: I0318 20:30:37.736288 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-ssh-key-openstack-edpm-ipam\") pod \"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0\" (UID: \"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0\") " Mar 18 20:30:37 crc kubenswrapper[4950]: I0318 20:30:37.736367 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-repo-setup-combined-ca-bundle\") pod \"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0\" (UID: \"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0\") " Mar 18 20:30:37 crc kubenswrapper[4950]: I0318 20:30:37.736492 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-inventory\") pod \"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0\" (UID: \"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0\") " Mar 18 20:30:37 crc kubenswrapper[4950]: I0318 20:30:37.741003 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0" (UID: "7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:30:37 crc kubenswrapper[4950]: I0318 20:30:37.746780 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-kube-api-access-4l68j" (OuterVolumeSpecName: "kube-api-access-4l68j") pod "7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0" (UID: "7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0"). InnerVolumeSpecName "kube-api-access-4l68j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:30:37 crc kubenswrapper[4950]: I0318 20:30:37.763275 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-inventory" (OuterVolumeSpecName: "inventory") pod "7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0" (UID: "7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:30:37 crc kubenswrapper[4950]: I0318 20:30:37.763300 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0" (UID: "7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:30:37 crc kubenswrapper[4950]: I0318 20:30:37.838498 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4l68j\" (UniqueName: \"kubernetes.io/projected/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-kube-api-access-4l68j\") on node \"crc\" DevicePath \"\"" Mar 18 20:30:37 crc kubenswrapper[4950]: I0318 20:30:37.838526 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:30:37 crc kubenswrapper[4950]: I0318 20:30:37.838536 4950 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:30:37 crc kubenswrapper[4950]: I0318 20:30:37.838545 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.060018 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv" event={"ID":"7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0","Type":"ContainerDied","Data":"dd34109a4ff124b97ebb801e54f6ea4a30b467af3e6367f31dd0ef8e3f4a7538"} Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.060063 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd34109a4ff124b97ebb801e54f6ea4a30b467af3e6367f31dd0ef8e3f4a7538" Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.060506 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv" Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.162512 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh"] Mar 18 20:30:38 crc kubenswrapper[4950]: E0318 20:30:38.162878 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.162901 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.163075 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.163668 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh" Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.166521 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.167139 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.167169 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.167523 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.174276 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh"] Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.247374 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1b58e52-1dea-4216-9eec-ebf5c09fba32-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh\" (UID: \"a1b58e52-1dea-4216-9eec-ebf5c09fba32\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh" Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.247989 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b58e52-1dea-4216-9eec-ebf5c09fba32-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh\" (UID: \"a1b58e52-1dea-4216-9eec-ebf5c09fba32\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh" Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.248230 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jv69q\" (UniqueName: \"kubernetes.io/projected/a1b58e52-1dea-4216-9eec-ebf5c09fba32-kube-api-access-jv69q\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh\" (UID: \"a1b58e52-1dea-4216-9eec-ebf5c09fba32\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh" Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.248448 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a1b58e52-1dea-4216-9eec-ebf5c09fba32-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh\" (UID: \"a1b58e52-1dea-4216-9eec-ebf5c09fba32\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh" Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.349179 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a1b58e52-1dea-4216-9eec-ebf5c09fba32-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh\" (UID: \"a1b58e52-1dea-4216-9eec-ebf5c09fba32\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh" Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.349485 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1b58e52-1dea-4216-9eec-ebf5c09fba32-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh\" (UID: \"a1b58e52-1dea-4216-9eec-ebf5c09fba32\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh" Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.349686 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b58e52-1dea-4216-9eec-ebf5c09fba32-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh\" (UID: \"a1b58e52-1dea-4216-9eec-ebf5c09fba32\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh" Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.349836 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jv69q\" (UniqueName: \"kubernetes.io/projected/a1b58e52-1dea-4216-9eec-ebf5c09fba32-kube-api-access-jv69q\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh\" (UID: \"a1b58e52-1dea-4216-9eec-ebf5c09fba32\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh" Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.356259 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a1b58e52-1dea-4216-9eec-ebf5c09fba32-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh\" (UID: \"a1b58e52-1dea-4216-9eec-ebf5c09fba32\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh" Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.358703 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1b58e52-1dea-4216-9eec-ebf5c09fba32-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh\" (UID: \"a1b58e52-1dea-4216-9eec-ebf5c09fba32\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh" Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.361232 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b58e52-1dea-4216-9eec-ebf5c09fba32-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh\" (UID: \"a1b58e52-1dea-4216-9eec-ebf5c09fba32\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh" Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.372707 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jv69q\" (UniqueName: \"kubernetes.io/projected/a1b58e52-1dea-4216-9eec-ebf5c09fba32-kube-api-access-jv69q\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh\" (UID: \"a1b58e52-1dea-4216-9eec-ebf5c09fba32\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh" Mar 18 20:30:38 crc kubenswrapper[4950]: I0318 20:30:38.485541 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh" Mar 18 20:30:39 crc kubenswrapper[4950]: I0318 20:30:39.161662 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh"] Mar 18 20:30:39 crc kubenswrapper[4950]: W0318 20:30:39.173797 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1b58e52_1dea_4216_9eec_ebf5c09fba32.slice/crio-565f6461e9ec1a067a2c3c2219485a95f19eae8ad4dee6a328e4cb0189098849 WatchSource:0}: Error finding container 565f6461e9ec1a067a2c3c2219485a95f19eae8ad4dee6a328e4cb0189098849: Status 404 returned error can't find the container with id 565f6461e9ec1a067a2c3c2219485a95f19eae8ad4dee6a328e4cb0189098849 Mar 18 20:30:40 crc kubenswrapper[4950]: I0318 20:30:40.085973 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh" event={"ID":"a1b58e52-1dea-4216-9eec-ebf5c09fba32","Type":"ContainerStarted","Data":"74242ce84c406b044ae39cf55d1dea785c786d4e82138ad9b68e2eedf3cc8061"} Mar 18 20:30:40 crc kubenswrapper[4950]: I0318 20:30:40.086306 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh" event={"ID":"a1b58e52-1dea-4216-9eec-ebf5c09fba32","Type":"ContainerStarted","Data":"565f6461e9ec1a067a2c3c2219485a95f19eae8ad4dee6a328e4cb0189098849"} Mar 18 20:30:40 crc kubenswrapper[4950]: I0318 20:30:40.106717 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh" podStartSLOduration=1.6975410050000002 podStartE2EDuration="2.106695249s" podCreationTimestamp="2026-03-18 20:30:38 +0000 UTC" firstStartedPulling="2026-03-18 20:30:39.181041749 +0000 UTC m=+1452.421883617" lastFinishedPulling="2026-03-18 20:30:39.590195953 +0000 UTC m=+1452.831037861" observedRunningTime="2026-03-18 20:30:40.099571161 +0000 UTC m=+1453.340413049" watchObservedRunningTime="2026-03-18 20:30:40.106695249 +0000 UTC m=+1453.347537137" Mar 18 20:30:41 crc kubenswrapper[4950]: I0318 20:30:41.158617 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 18 20:30:41 crc kubenswrapper[4950]: I0318 20:30:41.886576 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 18 20:30:46 crc kubenswrapper[4950]: I0318 20:30:46.121149 4950 scope.go:117] "RemoveContainer" containerID="df6c20cda351b2931e39b19a9a5094c1f74accc06feb7285ed7e0985dee7e485" Mar 18 20:31:33 crc kubenswrapper[4950]: I0318 20:31:33.836778 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:31:33 crc kubenswrapper[4950]: I0318 20:31:33.837397 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:31:46 crc kubenswrapper[4950]: I0318 20:31:46.228603 4950 scope.go:117] "RemoveContainer" containerID="0aad004978d3ce977decea721b811314e837f4da8975d2e599ca41ff0b3339b0" Mar 18 20:31:46 crc kubenswrapper[4950]: I0318 20:31:46.266068 4950 scope.go:117] "RemoveContainer" containerID="e9069b3bc780d7e4876c5f477247669b5903a0aaf4a440d0d609149fdf78e0e1" Mar 18 20:32:00 crc kubenswrapper[4950]: I0318 20:32:00.165629 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564432-qs5nb"] Mar 18 20:32:00 crc kubenswrapper[4950]: I0318 20:32:00.167441 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564432-qs5nb" Mar 18 20:32:00 crc kubenswrapper[4950]: I0318 20:32:00.169552 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:32:00 crc kubenswrapper[4950]: I0318 20:32:00.172857 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:32:00 crc kubenswrapper[4950]: I0318 20:32:00.172863 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:32:00 crc kubenswrapper[4950]: I0318 20:32:00.185963 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8p89w\" (UniqueName: \"kubernetes.io/projected/3d3a6361-b1e7-4192-b620-71bb92c2844a-kube-api-access-8p89w\") pod \"auto-csr-approver-29564432-qs5nb\" (UID: \"3d3a6361-b1e7-4192-b620-71bb92c2844a\") " pod="openshift-infra/auto-csr-approver-29564432-qs5nb" Mar 18 20:32:00 crc kubenswrapper[4950]: I0318 20:32:00.195594 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564432-qs5nb"] Mar 18 20:32:00 crc kubenswrapper[4950]: I0318 20:32:00.289022 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8p89w\" (UniqueName: \"kubernetes.io/projected/3d3a6361-b1e7-4192-b620-71bb92c2844a-kube-api-access-8p89w\") pod \"auto-csr-approver-29564432-qs5nb\" (UID: \"3d3a6361-b1e7-4192-b620-71bb92c2844a\") " pod="openshift-infra/auto-csr-approver-29564432-qs5nb" Mar 18 20:32:00 crc kubenswrapper[4950]: I0318 20:32:00.322120 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8p89w\" (UniqueName: \"kubernetes.io/projected/3d3a6361-b1e7-4192-b620-71bb92c2844a-kube-api-access-8p89w\") pod \"auto-csr-approver-29564432-qs5nb\" (UID: \"3d3a6361-b1e7-4192-b620-71bb92c2844a\") " pod="openshift-infra/auto-csr-approver-29564432-qs5nb" Mar 18 20:32:00 crc kubenswrapper[4950]: I0318 20:32:00.499284 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564432-qs5nb" Mar 18 20:32:00 crc kubenswrapper[4950]: I0318 20:32:00.985345 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564432-qs5nb"] Mar 18 20:32:01 crc kubenswrapper[4950]: W0318 20:32:01.000845 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d3a6361_b1e7_4192_b620_71bb92c2844a.slice/crio-db27c9c7ac199f50c456422aa8eb4db30c3f3cf31701c0448c47bc9dae0219f8 WatchSource:0}: Error finding container db27c9c7ac199f50c456422aa8eb4db30c3f3cf31701c0448c47bc9dae0219f8: Status 404 returned error can't find the container with id db27c9c7ac199f50c456422aa8eb4db30c3f3cf31701c0448c47bc9dae0219f8 Mar 18 20:32:01 crc kubenswrapper[4950]: I0318 20:32:01.006121 4950 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 20:32:01 crc kubenswrapper[4950]: I0318 20:32:01.679594 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564432-qs5nb" event={"ID":"3d3a6361-b1e7-4192-b620-71bb92c2844a","Type":"ContainerStarted","Data":"db27c9c7ac199f50c456422aa8eb4db30c3f3cf31701c0448c47bc9dae0219f8"} Mar 18 20:32:02 crc kubenswrapper[4950]: I0318 20:32:02.691707 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564432-qs5nb" event={"ID":"3d3a6361-b1e7-4192-b620-71bb92c2844a","Type":"ContainerStarted","Data":"1c4256450e87d1329579e44017780b5ac367ecc5a866b74fcd756d459d685bd9"} Mar 18 20:32:02 crc kubenswrapper[4950]: I0318 20:32:02.715698 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29564432-qs5nb" podStartSLOduration=1.767198155 podStartE2EDuration="2.715682105s" podCreationTimestamp="2026-03-18 20:32:00 +0000 UTC" firstStartedPulling="2026-03-18 20:32:01.005781422 +0000 UTC m=+1534.246623300" lastFinishedPulling="2026-03-18 20:32:01.954265342 +0000 UTC m=+1535.195107250" observedRunningTime="2026-03-18 20:32:02.707672524 +0000 UTC m=+1535.948514402" watchObservedRunningTime="2026-03-18 20:32:02.715682105 +0000 UTC m=+1535.956523973" Mar 18 20:32:03 crc kubenswrapper[4950]: I0318 20:32:03.705718 4950 generic.go:334] "Generic (PLEG): container finished" podID="3d3a6361-b1e7-4192-b620-71bb92c2844a" containerID="1c4256450e87d1329579e44017780b5ac367ecc5a866b74fcd756d459d685bd9" exitCode=0 Mar 18 20:32:03 crc kubenswrapper[4950]: I0318 20:32:03.705821 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564432-qs5nb" event={"ID":"3d3a6361-b1e7-4192-b620-71bb92c2844a","Type":"ContainerDied","Data":"1c4256450e87d1329579e44017780b5ac367ecc5a866b74fcd756d459d685bd9"} Mar 18 20:32:03 crc kubenswrapper[4950]: I0318 20:32:03.836489 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:32:03 crc kubenswrapper[4950]: I0318 20:32:03.836614 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:32:05 crc kubenswrapper[4950]: I0318 20:32:05.107461 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564432-qs5nb" Mar 18 20:32:05 crc kubenswrapper[4950]: I0318 20:32:05.195194 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8p89w\" (UniqueName: \"kubernetes.io/projected/3d3a6361-b1e7-4192-b620-71bb92c2844a-kube-api-access-8p89w\") pod \"3d3a6361-b1e7-4192-b620-71bb92c2844a\" (UID: \"3d3a6361-b1e7-4192-b620-71bb92c2844a\") " Mar 18 20:32:05 crc kubenswrapper[4950]: I0318 20:32:05.202331 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d3a6361-b1e7-4192-b620-71bb92c2844a-kube-api-access-8p89w" (OuterVolumeSpecName: "kube-api-access-8p89w") pod "3d3a6361-b1e7-4192-b620-71bb92c2844a" (UID: "3d3a6361-b1e7-4192-b620-71bb92c2844a"). InnerVolumeSpecName "kube-api-access-8p89w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:32:05 crc kubenswrapper[4950]: I0318 20:32:05.298728 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8p89w\" (UniqueName: \"kubernetes.io/projected/3d3a6361-b1e7-4192-b620-71bb92c2844a-kube-api-access-8p89w\") on node \"crc\" DevicePath \"\"" Mar 18 20:32:05 crc kubenswrapper[4950]: I0318 20:32:05.728190 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564432-qs5nb" event={"ID":"3d3a6361-b1e7-4192-b620-71bb92c2844a","Type":"ContainerDied","Data":"db27c9c7ac199f50c456422aa8eb4db30c3f3cf31701c0448c47bc9dae0219f8"} Mar 18 20:32:05 crc kubenswrapper[4950]: I0318 20:32:05.728238 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db27c9c7ac199f50c456422aa8eb4db30c3f3cf31701c0448c47bc9dae0219f8" Mar 18 20:32:05 crc kubenswrapper[4950]: I0318 20:32:05.728248 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564432-qs5nb" Mar 18 20:32:05 crc kubenswrapper[4950]: I0318 20:32:05.787626 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564426-4lcpm"] Mar 18 20:32:05 crc kubenswrapper[4950]: I0318 20:32:05.797601 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564426-4lcpm"] Mar 18 20:32:07 crc kubenswrapper[4950]: I0318 20:32:07.498483 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="619d4317-c521-4032-a59f-067f9914bb44" path="/var/lib/kubelet/pods/619d4317-c521-4032-a59f-067f9914bb44/volumes" Mar 18 20:32:33 crc kubenswrapper[4950]: I0318 20:32:33.835967 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:32:33 crc kubenswrapper[4950]: I0318 20:32:33.836593 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:32:33 crc kubenswrapper[4950]: I0318 20:32:33.836638 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:32:33 crc kubenswrapper[4950]: I0318 20:32:33.837488 4950 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"02a13f19effbd6df78e6bd43d75fe3a70d3a67fc27d7dba640df7ed8a4afc83f"} pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 20:32:33 crc kubenswrapper[4950]: I0318 20:32:33.837542 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" containerID="cri-o://02a13f19effbd6df78e6bd43d75fe3a70d3a67fc27d7dba640df7ed8a4afc83f" gracePeriod=600 Mar 18 20:32:34 crc kubenswrapper[4950]: I0318 20:32:34.068931 4950 generic.go:334] "Generic (PLEG): container finished" podID="4048b439-3266-46e7-9de0-22377efacc46" containerID="02a13f19effbd6df78e6bd43d75fe3a70d3a67fc27d7dba640df7ed8a4afc83f" exitCode=0 Mar 18 20:32:34 crc kubenswrapper[4950]: I0318 20:32:34.068998 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerDied","Data":"02a13f19effbd6df78e6bd43d75fe3a70d3a67fc27d7dba640df7ed8a4afc83f"} Mar 18 20:32:34 crc kubenswrapper[4950]: I0318 20:32:34.069619 4950 scope.go:117] "RemoveContainer" containerID="4e352dfc9da032a31ef3fccebe090411fdd509a1d2eb6df9588325dd52adecda" Mar 18 20:32:35 crc kubenswrapper[4950]: I0318 20:32:35.088503 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerStarted","Data":"f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e"} Mar 18 20:32:46 crc kubenswrapper[4950]: I0318 20:32:46.388296 4950 scope.go:117] "RemoveContainer" containerID="fafb42832d2dcda2f28bda37a748b550b176d5f3fdcdc2a73e44ebdaf4cb9e3c" Mar 18 20:32:46 crc kubenswrapper[4950]: I0318 20:32:46.458621 4950 scope.go:117] "RemoveContainer" containerID="a1291c01d23d49a5770ca05d1127cad33c511df6e4bffd22c195d2178762aebb" Mar 18 20:32:46 crc kubenswrapper[4950]: I0318 20:32:46.482233 4950 scope.go:117] "RemoveContainer" containerID="1a47134f941a07b6ccc64ee6978cdd2910a4f674c4f2e2038ee93b17c562ff9c" Mar 18 20:32:46 crc kubenswrapper[4950]: I0318 20:32:46.532287 4950 scope.go:117] "RemoveContainer" containerID="9b8971bb26768fa3b1732ed9bb879b77e747608497828e2f327f5cbfc179b4fc" Mar 18 20:33:13 crc kubenswrapper[4950]: I0318 20:33:13.430117 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dtmzz"] Mar 18 20:33:13 crc kubenswrapper[4950]: E0318 20:33:13.430914 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d3a6361-b1e7-4192-b620-71bb92c2844a" containerName="oc" Mar 18 20:33:13 crc kubenswrapper[4950]: I0318 20:33:13.430925 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d3a6361-b1e7-4192-b620-71bb92c2844a" containerName="oc" Mar 18 20:33:13 crc kubenswrapper[4950]: I0318 20:33:13.431119 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d3a6361-b1e7-4192-b620-71bb92c2844a" containerName="oc" Mar 18 20:33:13 crc kubenswrapper[4950]: I0318 20:33:13.432300 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtmzz" Mar 18 20:33:13 crc kubenswrapper[4950]: I0318 20:33:13.445106 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dtmzz"] Mar 18 20:33:13 crc kubenswrapper[4950]: I0318 20:33:13.508927 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxh5z\" (UniqueName: \"kubernetes.io/projected/9a9d2338-9268-424b-80a8-6401f09165e4-kube-api-access-mxh5z\") pod \"certified-operators-dtmzz\" (UID: \"9a9d2338-9268-424b-80a8-6401f09165e4\") " pod="openshift-marketplace/certified-operators-dtmzz" Mar 18 20:33:13 crc kubenswrapper[4950]: I0318 20:33:13.509012 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a9d2338-9268-424b-80a8-6401f09165e4-catalog-content\") pod \"certified-operators-dtmzz\" (UID: \"9a9d2338-9268-424b-80a8-6401f09165e4\") " pod="openshift-marketplace/certified-operators-dtmzz" Mar 18 20:33:13 crc kubenswrapper[4950]: I0318 20:33:13.509058 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a9d2338-9268-424b-80a8-6401f09165e4-utilities\") pod \"certified-operators-dtmzz\" (UID: \"9a9d2338-9268-424b-80a8-6401f09165e4\") " pod="openshift-marketplace/certified-operators-dtmzz" Mar 18 20:33:13 crc kubenswrapper[4950]: I0318 20:33:13.610831 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxh5z\" (UniqueName: \"kubernetes.io/projected/9a9d2338-9268-424b-80a8-6401f09165e4-kube-api-access-mxh5z\") pod \"certified-operators-dtmzz\" (UID: \"9a9d2338-9268-424b-80a8-6401f09165e4\") " pod="openshift-marketplace/certified-operators-dtmzz" Mar 18 20:33:13 crc kubenswrapper[4950]: I0318 20:33:13.611189 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a9d2338-9268-424b-80a8-6401f09165e4-catalog-content\") pod \"certified-operators-dtmzz\" (UID: \"9a9d2338-9268-424b-80a8-6401f09165e4\") " pod="openshift-marketplace/certified-operators-dtmzz" Mar 18 20:33:13 crc kubenswrapper[4950]: I0318 20:33:13.611211 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a9d2338-9268-424b-80a8-6401f09165e4-utilities\") pod \"certified-operators-dtmzz\" (UID: \"9a9d2338-9268-424b-80a8-6401f09165e4\") " pod="openshift-marketplace/certified-operators-dtmzz" Mar 18 20:33:13 crc kubenswrapper[4950]: I0318 20:33:13.611653 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a9d2338-9268-424b-80a8-6401f09165e4-catalog-content\") pod \"certified-operators-dtmzz\" (UID: \"9a9d2338-9268-424b-80a8-6401f09165e4\") " pod="openshift-marketplace/certified-operators-dtmzz" Mar 18 20:33:13 crc kubenswrapper[4950]: I0318 20:33:13.611698 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a9d2338-9268-424b-80a8-6401f09165e4-utilities\") pod \"certified-operators-dtmzz\" (UID: \"9a9d2338-9268-424b-80a8-6401f09165e4\") " pod="openshift-marketplace/certified-operators-dtmzz" Mar 18 20:33:13 crc kubenswrapper[4950]: I0318 20:33:13.639056 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxh5z\" (UniqueName: \"kubernetes.io/projected/9a9d2338-9268-424b-80a8-6401f09165e4-kube-api-access-mxh5z\") pod \"certified-operators-dtmzz\" (UID: \"9a9d2338-9268-424b-80a8-6401f09165e4\") " pod="openshift-marketplace/certified-operators-dtmzz" Mar 18 20:33:13 crc kubenswrapper[4950]: I0318 20:33:13.765562 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtmzz" Mar 18 20:33:14 crc kubenswrapper[4950]: I0318 20:33:14.269783 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dtmzz"] Mar 18 20:33:15 crc kubenswrapper[4950]: I0318 20:33:15.008705 4950 generic.go:334] "Generic (PLEG): container finished" podID="9a9d2338-9268-424b-80a8-6401f09165e4" containerID="136bdb163ee980abd587ee1ac2f6778fb8f5e05778d765db37ea924c3da5a39d" exitCode=0 Mar 18 20:33:15 crc kubenswrapper[4950]: I0318 20:33:15.008748 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtmzz" event={"ID":"9a9d2338-9268-424b-80a8-6401f09165e4","Type":"ContainerDied","Data":"136bdb163ee980abd587ee1ac2f6778fb8f5e05778d765db37ea924c3da5a39d"} Mar 18 20:33:15 crc kubenswrapper[4950]: I0318 20:33:15.008787 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtmzz" event={"ID":"9a9d2338-9268-424b-80a8-6401f09165e4","Type":"ContainerStarted","Data":"c66fab53955ba6ac3ba710c8c41e290d74eee2093c7c18da2ce1494fab78b939"} Mar 18 20:33:16 crc kubenswrapper[4950]: I0318 20:33:16.020874 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtmzz" event={"ID":"9a9d2338-9268-424b-80a8-6401f09165e4","Type":"ContainerStarted","Data":"99d8a9f7f4b4eea973de2ca04a543f18b8c7b05c51f4a8b83ea35c93997a363f"} Mar 18 20:33:18 crc kubenswrapper[4950]: I0318 20:33:18.040608 4950 generic.go:334] "Generic (PLEG): container finished" podID="9a9d2338-9268-424b-80a8-6401f09165e4" containerID="99d8a9f7f4b4eea973de2ca04a543f18b8c7b05c51f4a8b83ea35c93997a363f" exitCode=0 Mar 18 20:33:18 crc kubenswrapper[4950]: I0318 20:33:18.040669 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtmzz" event={"ID":"9a9d2338-9268-424b-80a8-6401f09165e4","Type":"ContainerDied","Data":"99d8a9f7f4b4eea973de2ca04a543f18b8c7b05c51f4a8b83ea35c93997a363f"} Mar 18 20:33:19 crc kubenswrapper[4950]: I0318 20:33:19.053981 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtmzz" event={"ID":"9a9d2338-9268-424b-80a8-6401f09165e4","Type":"ContainerStarted","Data":"7051a0b1b60cba686f8aaf0ccbb5a8d229ee79710358fb1a5e2950086f2dfcd5"} Mar 18 20:33:19 crc kubenswrapper[4950]: I0318 20:33:19.085983 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dtmzz" podStartSLOduration=2.632000536 podStartE2EDuration="6.085958188s" podCreationTimestamp="2026-03-18 20:33:13 +0000 UTC" firstStartedPulling="2026-03-18 20:33:15.0119333 +0000 UTC m=+1608.252775208" lastFinishedPulling="2026-03-18 20:33:18.465890982 +0000 UTC m=+1611.706732860" observedRunningTime="2026-03-18 20:33:19.07535983 +0000 UTC m=+1612.316201718" watchObservedRunningTime="2026-03-18 20:33:19.085958188 +0000 UTC m=+1612.326800076" Mar 18 20:33:20 crc kubenswrapper[4950]: I0318 20:33:20.806385 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2x2dv"] Mar 18 20:33:20 crc kubenswrapper[4950]: I0318 20:33:20.808021 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2x2dv" Mar 18 20:33:20 crc kubenswrapper[4950]: I0318 20:33:20.816382 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2x2dv"] Mar 18 20:33:20 crc kubenswrapper[4950]: I0318 20:33:20.949954 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhmm7\" (UniqueName: \"kubernetes.io/projected/e1bc2013-dd47-41c7-a055-d7740d82edfa-kube-api-access-xhmm7\") pod \"redhat-marketplace-2x2dv\" (UID: \"e1bc2013-dd47-41c7-a055-d7740d82edfa\") " pod="openshift-marketplace/redhat-marketplace-2x2dv" Mar 18 20:33:20 crc kubenswrapper[4950]: I0318 20:33:20.950250 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1bc2013-dd47-41c7-a055-d7740d82edfa-catalog-content\") pod \"redhat-marketplace-2x2dv\" (UID: \"e1bc2013-dd47-41c7-a055-d7740d82edfa\") " pod="openshift-marketplace/redhat-marketplace-2x2dv" Mar 18 20:33:20 crc kubenswrapper[4950]: I0318 20:33:20.950386 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1bc2013-dd47-41c7-a055-d7740d82edfa-utilities\") pod \"redhat-marketplace-2x2dv\" (UID: \"e1bc2013-dd47-41c7-a055-d7740d82edfa\") " pod="openshift-marketplace/redhat-marketplace-2x2dv" Mar 18 20:33:21 crc kubenswrapper[4950]: I0318 20:33:21.051655 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1bc2013-dd47-41c7-a055-d7740d82edfa-utilities\") pod \"redhat-marketplace-2x2dv\" (UID: \"e1bc2013-dd47-41c7-a055-d7740d82edfa\") " pod="openshift-marketplace/redhat-marketplace-2x2dv" Mar 18 20:33:21 crc kubenswrapper[4950]: I0318 20:33:21.051732 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhmm7\" (UniqueName: \"kubernetes.io/projected/e1bc2013-dd47-41c7-a055-d7740d82edfa-kube-api-access-xhmm7\") pod \"redhat-marketplace-2x2dv\" (UID: \"e1bc2013-dd47-41c7-a055-d7740d82edfa\") " pod="openshift-marketplace/redhat-marketplace-2x2dv" Mar 18 20:33:21 crc kubenswrapper[4950]: I0318 20:33:21.051757 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1bc2013-dd47-41c7-a055-d7740d82edfa-catalog-content\") pod \"redhat-marketplace-2x2dv\" (UID: \"e1bc2013-dd47-41c7-a055-d7740d82edfa\") " pod="openshift-marketplace/redhat-marketplace-2x2dv" Mar 18 20:33:21 crc kubenswrapper[4950]: I0318 20:33:21.052277 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1bc2013-dd47-41c7-a055-d7740d82edfa-catalog-content\") pod \"redhat-marketplace-2x2dv\" (UID: \"e1bc2013-dd47-41c7-a055-d7740d82edfa\") " pod="openshift-marketplace/redhat-marketplace-2x2dv" Mar 18 20:33:21 crc kubenswrapper[4950]: I0318 20:33:21.052320 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1bc2013-dd47-41c7-a055-d7740d82edfa-utilities\") pod \"redhat-marketplace-2x2dv\" (UID: \"e1bc2013-dd47-41c7-a055-d7740d82edfa\") " pod="openshift-marketplace/redhat-marketplace-2x2dv" Mar 18 20:33:21 crc kubenswrapper[4950]: I0318 20:33:21.070449 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhmm7\" (UniqueName: \"kubernetes.io/projected/e1bc2013-dd47-41c7-a055-d7740d82edfa-kube-api-access-xhmm7\") pod \"redhat-marketplace-2x2dv\" (UID: \"e1bc2013-dd47-41c7-a055-d7740d82edfa\") " pod="openshift-marketplace/redhat-marketplace-2x2dv" Mar 18 20:33:21 crc kubenswrapper[4950]: I0318 20:33:21.146347 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2x2dv" Mar 18 20:33:21 crc kubenswrapper[4950]: W0318 20:33:21.603672 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1bc2013_dd47_41c7_a055_d7740d82edfa.slice/crio-bba3353e8ab3d82b35e6cc6d0a46598219b129f27e74c9c33de549ee4f91e240 WatchSource:0}: Error finding container bba3353e8ab3d82b35e6cc6d0a46598219b129f27e74c9c33de549ee4f91e240: Status 404 returned error can't find the container with id bba3353e8ab3d82b35e6cc6d0a46598219b129f27e74c9c33de549ee4f91e240 Mar 18 20:33:21 crc kubenswrapper[4950]: I0318 20:33:21.613095 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2x2dv"] Mar 18 20:33:22 crc kubenswrapper[4950]: I0318 20:33:22.082815 4950 generic.go:334] "Generic (PLEG): container finished" podID="e1bc2013-dd47-41c7-a055-d7740d82edfa" containerID="ed8389cff8ddb6c191e4a1ccb3c3943c9b1d1558190b98d8e496b2cfddb01cff" exitCode=0 Mar 18 20:33:22 crc kubenswrapper[4950]: I0318 20:33:22.082862 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2x2dv" event={"ID":"e1bc2013-dd47-41c7-a055-d7740d82edfa","Type":"ContainerDied","Data":"ed8389cff8ddb6c191e4a1ccb3c3943c9b1d1558190b98d8e496b2cfddb01cff"} Mar 18 20:33:22 crc kubenswrapper[4950]: I0318 20:33:22.082891 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2x2dv" event={"ID":"e1bc2013-dd47-41c7-a055-d7740d82edfa","Type":"ContainerStarted","Data":"bba3353e8ab3d82b35e6cc6d0a46598219b129f27e74c9c33de549ee4f91e240"} Mar 18 20:33:23 crc kubenswrapper[4950]: I0318 20:33:23.092011 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2x2dv" event={"ID":"e1bc2013-dd47-41c7-a055-d7740d82edfa","Type":"ContainerStarted","Data":"17bca0bf43a96467f418a84e68f1edeb106897dcc9b56afa049f1ff0258f6d6b"} Mar 18 20:33:23 crc kubenswrapper[4950]: I0318 20:33:23.765931 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dtmzz" Mar 18 20:33:23 crc kubenswrapper[4950]: I0318 20:33:23.766499 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dtmzz" Mar 18 20:33:23 crc kubenswrapper[4950]: I0318 20:33:23.812965 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dtmzz" Mar 18 20:33:24 crc kubenswrapper[4950]: I0318 20:33:24.150952 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dtmzz" Mar 18 20:33:25 crc kubenswrapper[4950]: I0318 20:33:25.111709 4950 generic.go:334] "Generic (PLEG): container finished" podID="e1bc2013-dd47-41c7-a055-d7740d82edfa" containerID="17bca0bf43a96467f418a84e68f1edeb106897dcc9b56afa049f1ff0258f6d6b" exitCode=0 Mar 18 20:33:25 crc kubenswrapper[4950]: I0318 20:33:25.111771 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2x2dv" event={"ID":"e1bc2013-dd47-41c7-a055-d7740d82edfa","Type":"ContainerDied","Data":"17bca0bf43a96467f418a84e68f1edeb106897dcc9b56afa049f1ff0258f6d6b"} Mar 18 20:33:26 crc kubenswrapper[4950]: I0318 20:33:26.121135 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2x2dv" event={"ID":"e1bc2013-dd47-41c7-a055-d7740d82edfa","Type":"ContainerStarted","Data":"1c7aae6dcfc0be59d918f371b10ecb048268beffcd3cdd93c5d08ec12467ca41"} Mar 18 20:33:26 crc kubenswrapper[4950]: I0318 20:33:26.146423 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2x2dv" podStartSLOduration=2.734432634 podStartE2EDuration="6.146391233s" podCreationTimestamp="2026-03-18 20:33:20 +0000 UTC" firstStartedPulling="2026-03-18 20:33:22.083977887 +0000 UTC m=+1615.324819755" lastFinishedPulling="2026-03-18 20:33:25.495936486 +0000 UTC m=+1618.736778354" observedRunningTime="2026-03-18 20:33:26.145803657 +0000 UTC m=+1619.386645525" watchObservedRunningTime="2026-03-18 20:33:26.146391233 +0000 UTC m=+1619.387233101" Mar 18 20:33:26 crc kubenswrapper[4950]: I0318 20:33:26.207947 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dtmzz"] Mar 18 20:33:27 crc kubenswrapper[4950]: I0318 20:33:27.130639 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dtmzz" podUID="9a9d2338-9268-424b-80a8-6401f09165e4" containerName="registry-server" containerID="cri-o://7051a0b1b60cba686f8aaf0ccbb5a8d229ee79710358fb1a5e2950086f2dfcd5" gracePeriod=2 Mar 18 20:33:27 crc kubenswrapper[4950]: I0318 20:33:27.604469 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtmzz" Mar 18 20:33:27 crc kubenswrapper[4950]: I0318 20:33:27.681829 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a9d2338-9268-424b-80a8-6401f09165e4-utilities\") pod \"9a9d2338-9268-424b-80a8-6401f09165e4\" (UID: \"9a9d2338-9268-424b-80a8-6401f09165e4\") " Mar 18 20:33:27 crc kubenswrapper[4950]: I0318 20:33:27.682017 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a9d2338-9268-424b-80a8-6401f09165e4-catalog-content\") pod \"9a9d2338-9268-424b-80a8-6401f09165e4\" (UID: \"9a9d2338-9268-424b-80a8-6401f09165e4\") " Mar 18 20:33:27 crc kubenswrapper[4950]: I0318 20:33:27.682110 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxh5z\" (UniqueName: \"kubernetes.io/projected/9a9d2338-9268-424b-80a8-6401f09165e4-kube-api-access-mxh5z\") pod \"9a9d2338-9268-424b-80a8-6401f09165e4\" (UID: \"9a9d2338-9268-424b-80a8-6401f09165e4\") " Mar 18 20:33:27 crc kubenswrapper[4950]: I0318 20:33:27.684495 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a9d2338-9268-424b-80a8-6401f09165e4-utilities" (OuterVolumeSpecName: "utilities") pod "9a9d2338-9268-424b-80a8-6401f09165e4" (UID: "9a9d2338-9268-424b-80a8-6401f09165e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:33:27 crc kubenswrapper[4950]: I0318 20:33:27.686792 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a9d2338-9268-424b-80a8-6401f09165e4-kube-api-access-mxh5z" (OuterVolumeSpecName: "kube-api-access-mxh5z") pod "9a9d2338-9268-424b-80a8-6401f09165e4" (UID: "9a9d2338-9268-424b-80a8-6401f09165e4"). InnerVolumeSpecName "kube-api-access-mxh5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:33:27 crc kubenswrapper[4950]: I0318 20:33:27.738929 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a9d2338-9268-424b-80a8-6401f09165e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9a9d2338-9268-424b-80a8-6401f09165e4" (UID: "9a9d2338-9268-424b-80a8-6401f09165e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:33:27 crc kubenswrapper[4950]: I0318 20:33:27.784008 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxh5z\" (UniqueName: \"kubernetes.io/projected/9a9d2338-9268-424b-80a8-6401f09165e4-kube-api-access-mxh5z\") on node \"crc\" DevicePath \"\"" Mar 18 20:33:27 crc kubenswrapper[4950]: I0318 20:33:27.784043 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a9d2338-9268-424b-80a8-6401f09165e4-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:33:27 crc kubenswrapper[4950]: I0318 20:33:27.784052 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a9d2338-9268-424b-80a8-6401f09165e4-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:33:28 crc kubenswrapper[4950]: I0318 20:33:28.140503 4950 generic.go:334] "Generic (PLEG): container finished" podID="9a9d2338-9268-424b-80a8-6401f09165e4" containerID="7051a0b1b60cba686f8aaf0ccbb5a8d229ee79710358fb1a5e2950086f2dfcd5" exitCode=0 Mar 18 20:33:28 crc kubenswrapper[4950]: I0318 20:33:28.140548 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtmzz" event={"ID":"9a9d2338-9268-424b-80a8-6401f09165e4","Type":"ContainerDied","Data":"7051a0b1b60cba686f8aaf0ccbb5a8d229ee79710358fb1a5e2950086f2dfcd5"} Mar 18 20:33:28 crc kubenswrapper[4950]: I0318 20:33:28.140563 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtmzz" Mar 18 20:33:28 crc kubenswrapper[4950]: I0318 20:33:28.140577 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtmzz" event={"ID":"9a9d2338-9268-424b-80a8-6401f09165e4","Type":"ContainerDied","Data":"c66fab53955ba6ac3ba710c8c41e290d74eee2093c7c18da2ce1494fab78b939"} Mar 18 20:33:28 crc kubenswrapper[4950]: I0318 20:33:28.140597 4950 scope.go:117] "RemoveContainer" containerID="7051a0b1b60cba686f8aaf0ccbb5a8d229ee79710358fb1a5e2950086f2dfcd5" Mar 18 20:33:28 crc kubenswrapper[4950]: I0318 20:33:28.163776 4950 scope.go:117] "RemoveContainer" containerID="99d8a9f7f4b4eea973de2ca04a543f18b8c7b05c51f4a8b83ea35c93997a363f" Mar 18 20:33:28 crc kubenswrapper[4950]: I0318 20:33:28.175192 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dtmzz"] Mar 18 20:33:28 crc kubenswrapper[4950]: I0318 20:33:28.183357 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dtmzz"] Mar 18 20:33:28 crc kubenswrapper[4950]: I0318 20:33:28.194486 4950 scope.go:117] "RemoveContainer" containerID="136bdb163ee980abd587ee1ac2f6778fb8f5e05778d765db37ea924c3da5a39d" Mar 18 20:33:28 crc kubenswrapper[4950]: I0318 20:33:28.227246 4950 scope.go:117] "RemoveContainer" containerID="7051a0b1b60cba686f8aaf0ccbb5a8d229ee79710358fb1a5e2950086f2dfcd5" Mar 18 20:33:28 crc kubenswrapper[4950]: E0318 20:33:28.228149 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7051a0b1b60cba686f8aaf0ccbb5a8d229ee79710358fb1a5e2950086f2dfcd5\": container with ID starting with 7051a0b1b60cba686f8aaf0ccbb5a8d229ee79710358fb1a5e2950086f2dfcd5 not found: ID does not exist" containerID="7051a0b1b60cba686f8aaf0ccbb5a8d229ee79710358fb1a5e2950086f2dfcd5" Mar 18 20:33:28 crc kubenswrapper[4950]: I0318 20:33:28.228181 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7051a0b1b60cba686f8aaf0ccbb5a8d229ee79710358fb1a5e2950086f2dfcd5"} err="failed to get container status \"7051a0b1b60cba686f8aaf0ccbb5a8d229ee79710358fb1a5e2950086f2dfcd5\": rpc error: code = NotFound desc = could not find container \"7051a0b1b60cba686f8aaf0ccbb5a8d229ee79710358fb1a5e2950086f2dfcd5\": container with ID starting with 7051a0b1b60cba686f8aaf0ccbb5a8d229ee79710358fb1a5e2950086f2dfcd5 not found: ID does not exist" Mar 18 20:33:28 crc kubenswrapper[4950]: I0318 20:33:28.228204 4950 scope.go:117] "RemoveContainer" containerID="99d8a9f7f4b4eea973de2ca04a543f18b8c7b05c51f4a8b83ea35c93997a363f" Mar 18 20:33:28 crc kubenswrapper[4950]: E0318 20:33:28.228796 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99d8a9f7f4b4eea973de2ca04a543f18b8c7b05c51f4a8b83ea35c93997a363f\": container with ID starting with 99d8a9f7f4b4eea973de2ca04a543f18b8c7b05c51f4a8b83ea35c93997a363f not found: ID does not exist" containerID="99d8a9f7f4b4eea973de2ca04a543f18b8c7b05c51f4a8b83ea35c93997a363f" Mar 18 20:33:28 crc kubenswrapper[4950]: I0318 20:33:28.228819 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99d8a9f7f4b4eea973de2ca04a543f18b8c7b05c51f4a8b83ea35c93997a363f"} err="failed to get container status \"99d8a9f7f4b4eea973de2ca04a543f18b8c7b05c51f4a8b83ea35c93997a363f\": rpc error: code = NotFound desc = could not find container \"99d8a9f7f4b4eea973de2ca04a543f18b8c7b05c51f4a8b83ea35c93997a363f\": container with ID starting with 99d8a9f7f4b4eea973de2ca04a543f18b8c7b05c51f4a8b83ea35c93997a363f not found: ID does not exist" Mar 18 20:33:28 crc kubenswrapper[4950]: I0318 20:33:28.228832 4950 scope.go:117] "RemoveContainer" containerID="136bdb163ee980abd587ee1ac2f6778fb8f5e05778d765db37ea924c3da5a39d" Mar 18 20:33:28 crc kubenswrapper[4950]: E0318 20:33:28.229218 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"136bdb163ee980abd587ee1ac2f6778fb8f5e05778d765db37ea924c3da5a39d\": container with ID starting with 136bdb163ee980abd587ee1ac2f6778fb8f5e05778d765db37ea924c3da5a39d not found: ID does not exist" containerID="136bdb163ee980abd587ee1ac2f6778fb8f5e05778d765db37ea924c3da5a39d" Mar 18 20:33:28 crc kubenswrapper[4950]: I0318 20:33:28.229245 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"136bdb163ee980abd587ee1ac2f6778fb8f5e05778d765db37ea924c3da5a39d"} err="failed to get container status \"136bdb163ee980abd587ee1ac2f6778fb8f5e05778d765db37ea924c3da5a39d\": rpc error: code = NotFound desc = could not find container \"136bdb163ee980abd587ee1ac2f6778fb8f5e05778d765db37ea924c3da5a39d\": container with ID starting with 136bdb163ee980abd587ee1ac2f6778fb8f5e05778d765db37ea924c3da5a39d not found: ID does not exist" Mar 18 20:33:29 crc kubenswrapper[4950]: I0318 20:33:29.491281 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a9d2338-9268-424b-80a8-6401f09165e4" path="/var/lib/kubelet/pods/9a9d2338-9268-424b-80a8-6401f09165e4/volumes" Mar 18 20:33:31 crc kubenswrapper[4950]: I0318 20:33:31.147295 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2x2dv" Mar 18 20:33:31 crc kubenswrapper[4950]: I0318 20:33:31.147353 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2x2dv" Mar 18 20:33:31 crc kubenswrapper[4950]: I0318 20:33:31.221717 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2x2dv" Mar 18 20:33:31 crc kubenswrapper[4950]: I0318 20:33:31.293708 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2x2dv" Mar 18 20:33:32 crc kubenswrapper[4950]: I0318 20:33:32.200506 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2x2dv"] Mar 18 20:33:33 crc kubenswrapper[4950]: I0318 20:33:33.199385 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2x2dv" podUID="e1bc2013-dd47-41c7-a055-d7740d82edfa" containerName="registry-server" containerID="cri-o://1c7aae6dcfc0be59d918f371b10ecb048268beffcd3cdd93c5d08ec12467ca41" gracePeriod=2 Mar 18 20:33:33 crc kubenswrapper[4950]: I0318 20:33:33.686492 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2x2dv" Mar 18 20:33:33 crc kubenswrapper[4950]: I0318 20:33:33.805578 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1bc2013-dd47-41c7-a055-d7740d82edfa-utilities\") pod \"e1bc2013-dd47-41c7-a055-d7740d82edfa\" (UID: \"e1bc2013-dd47-41c7-a055-d7740d82edfa\") " Mar 18 20:33:33 crc kubenswrapper[4950]: I0318 20:33:33.805652 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1bc2013-dd47-41c7-a055-d7740d82edfa-catalog-content\") pod \"e1bc2013-dd47-41c7-a055-d7740d82edfa\" (UID: \"e1bc2013-dd47-41c7-a055-d7740d82edfa\") " Mar 18 20:33:33 crc kubenswrapper[4950]: I0318 20:33:33.805683 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhmm7\" (UniqueName: \"kubernetes.io/projected/e1bc2013-dd47-41c7-a055-d7740d82edfa-kube-api-access-xhmm7\") pod \"e1bc2013-dd47-41c7-a055-d7740d82edfa\" (UID: \"e1bc2013-dd47-41c7-a055-d7740d82edfa\") " Mar 18 20:33:33 crc kubenswrapper[4950]: I0318 20:33:33.807073 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1bc2013-dd47-41c7-a055-d7740d82edfa-utilities" (OuterVolumeSpecName: "utilities") pod "e1bc2013-dd47-41c7-a055-d7740d82edfa" (UID: "e1bc2013-dd47-41c7-a055-d7740d82edfa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:33:33 crc kubenswrapper[4950]: I0318 20:33:33.824261 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1bc2013-dd47-41c7-a055-d7740d82edfa-kube-api-access-xhmm7" (OuterVolumeSpecName: "kube-api-access-xhmm7") pod "e1bc2013-dd47-41c7-a055-d7740d82edfa" (UID: "e1bc2013-dd47-41c7-a055-d7740d82edfa"). InnerVolumeSpecName "kube-api-access-xhmm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:33:33 crc kubenswrapper[4950]: I0318 20:33:33.841239 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1bc2013-dd47-41c7-a055-d7740d82edfa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e1bc2013-dd47-41c7-a055-d7740d82edfa" (UID: "e1bc2013-dd47-41c7-a055-d7740d82edfa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:33:33 crc kubenswrapper[4950]: I0318 20:33:33.908507 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1bc2013-dd47-41c7-a055-d7740d82edfa-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:33:33 crc kubenswrapper[4950]: I0318 20:33:33.908572 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1bc2013-dd47-41c7-a055-d7740d82edfa-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:33:33 crc kubenswrapper[4950]: I0318 20:33:33.908600 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhmm7\" (UniqueName: \"kubernetes.io/projected/e1bc2013-dd47-41c7-a055-d7740d82edfa-kube-api-access-xhmm7\") on node \"crc\" DevicePath \"\"" Mar 18 20:33:34 crc kubenswrapper[4950]: I0318 20:33:34.214697 4950 generic.go:334] "Generic (PLEG): container finished" podID="e1bc2013-dd47-41c7-a055-d7740d82edfa" containerID="1c7aae6dcfc0be59d918f371b10ecb048268beffcd3cdd93c5d08ec12467ca41" exitCode=0 Mar 18 20:33:34 crc kubenswrapper[4950]: I0318 20:33:34.214744 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2x2dv" event={"ID":"e1bc2013-dd47-41c7-a055-d7740d82edfa","Type":"ContainerDied","Data":"1c7aae6dcfc0be59d918f371b10ecb048268beffcd3cdd93c5d08ec12467ca41"} Mar 18 20:33:34 crc kubenswrapper[4950]: I0318 20:33:34.214792 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2x2dv" event={"ID":"e1bc2013-dd47-41c7-a055-d7740d82edfa","Type":"ContainerDied","Data":"bba3353e8ab3d82b35e6cc6d0a46598219b129f27e74c9c33de549ee4f91e240"} Mar 18 20:33:34 crc kubenswrapper[4950]: I0318 20:33:34.214813 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2x2dv" Mar 18 20:33:34 crc kubenswrapper[4950]: I0318 20:33:34.214815 4950 scope.go:117] "RemoveContainer" containerID="1c7aae6dcfc0be59d918f371b10ecb048268beffcd3cdd93c5d08ec12467ca41" Mar 18 20:33:34 crc kubenswrapper[4950]: I0318 20:33:34.239784 4950 scope.go:117] "RemoveContainer" containerID="17bca0bf43a96467f418a84e68f1edeb106897dcc9b56afa049f1ff0258f6d6b" Mar 18 20:33:34 crc kubenswrapper[4950]: I0318 20:33:34.315333 4950 scope.go:117] "RemoveContainer" containerID="ed8389cff8ddb6c191e4a1ccb3c3943c9b1d1558190b98d8e496b2cfddb01cff" Mar 18 20:33:34 crc kubenswrapper[4950]: I0318 20:33:34.317683 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2x2dv"] Mar 18 20:33:34 crc kubenswrapper[4950]: I0318 20:33:34.336997 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2x2dv"] Mar 18 20:33:34 crc kubenswrapper[4950]: I0318 20:33:34.373189 4950 scope.go:117] "RemoveContainer" containerID="1c7aae6dcfc0be59d918f371b10ecb048268beffcd3cdd93c5d08ec12467ca41" Mar 18 20:33:34 crc kubenswrapper[4950]: E0318 20:33:34.378877 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c7aae6dcfc0be59d918f371b10ecb048268beffcd3cdd93c5d08ec12467ca41\": container with ID starting with 1c7aae6dcfc0be59d918f371b10ecb048268beffcd3cdd93c5d08ec12467ca41 not found: ID does not exist" containerID="1c7aae6dcfc0be59d918f371b10ecb048268beffcd3cdd93c5d08ec12467ca41" Mar 18 20:33:34 crc kubenswrapper[4950]: I0318 20:33:34.378949 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c7aae6dcfc0be59d918f371b10ecb048268beffcd3cdd93c5d08ec12467ca41"} err="failed to get container status \"1c7aae6dcfc0be59d918f371b10ecb048268beffcd3cdd93c5d08ec12467ca41\": rpc error: code = NotFound desc = could not find container \"1c7aae6dcfc0be59d918f371b10ecb048268beffcd3cdd93c5d08ec12467ca41\": container with ID starting with 1c7aae6dcfc0be59d918f371b10ecb048268beffcd3cdd93c5d08ec12467ca41 not found: ID does not exist" Mar 18 20:33:34 crc kubenswrapper[4950]: I0318 20:33:34.378977 4950 scope.go:117] "RemoveContainer" containerID="17bca0bf43a96467f418a84e68f1edeb106897dcc9b56afa049f1ff0258f6d6b" Mar 18 20:33:34 crc kubenswrapper[4950]: E0318 20:33:34.379372 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17bca0bf43a96467f418a84e68f1edeb106897dcc9b56afa049f1ff0258f6d6b\": container with ID starting with 17bca0bf43a96467f418a84e68f1edeb106897dcc9b56afa049f1ff0258f6d6b not found: ID does not exist" containerID="17bca0bf43a96467f418a84e68f1edeb106897dcc9b56afa049f1ff0258f6d6b" Mar 18 20:33:34 crc kubenswrapper[4950]: I0318 20:33:34.379483 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17bca0bf43a96467f418a84e68f1edeb106897dcc9b56afa049f1ff0258f6d6b"} err="failed to get container status \"17bca0bf43a96467f418a84e68f1edeb106897dcc9b56afa049f1ff0258f6d6b\": rpc error: code = NotFound desc = could not find container \"17bca0bf43a96467f418a84e68f1edeb106897dcc9b56afa049f1ff0258f6d6b\": container with ID starting with 17bca0bf43a96467f418a84e68f1edeb106897dcc9b56afa049f1ff0258f6d6b not found: ID does not exist" Mar 18 20:33:34 crc kubenswrapper[4950]: I0318 20:33:34.379573 4950 scope.go:117] "RemoveContainer" containerID="ed8389cff8ddb6c191e4a1ccb3c3943c9b1d1558190b98d8e496b2cfddb01cff" Mar 18 20:33:34 crc kubenswrapper[4950]: E0318 20:33:34.379895 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed8389cff8ddb6c191e4a1ccb3c3943c9b1d1558190b98d8e496b2cfddb01cff\": container with ID starting with ed8389cff8ddb6c191e4a1ccb3c3943c9b1d1558190b98d8e496b2cfddb01cff not found: ID does not exist" containerID="ed8389cff8ddb6c191e4a1ccb3c3943c9b1d1558190b98d8e496b2cfddb01cff" Mar 18 20:33:34 crc kubenswrapper[4950]: I0318 20:33:34.379931 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed8389cff8ddb6c191e4a1ccb3c3943c9b1d1558190b98d8e496b2cfddb01cff"} err="failed to get container status \"ed8389cff8ddb6c191e4a1ccb3c3943c9b1d1558190b98d8e496b2cfddb01cff\": rpc error: code = NotFound desc = could not find container \"ed8389cff8ddb6c191e4a1ccb3c3943c9b1d1558190b98d8e496b2cfddb01cff\": container with ID starting with ed8389cff8ddb6c191e4a1ccb3c3943c9b1d1558190b98d8e496b2cfddb01cff not found: ID does not exist" Mar 18 20:33:35 crc kubenswrapper[4950]: I0318 20:33:35.499268 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1bc2013-dd47-41c7-a055-d7740d82edfa" path="/var/lib/kubelet/pods/e1bc2013-dd47-41c7-a055-d7740d82edfa/volumes" Mar 18 20:33:40 crc kubenswrapper[4950]: I0318 20:33:40.448059 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zgzvl"] Mar 18 20:33:40 crc kubenswrapper[4950]: E0318 20:33:40.449089 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a9d2338-9268-424b-80a8-6401f09165e4" containerName="registry-server" Mar 18 20:33:40 crc kubenswrapper[4950]: I0318 20:33:40.449115 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a9d2338-9268-424b-80a8-6401f09165e4" containerName="registry-server" Mar 18 20:33:40 crc kubenswrapper[4950]: E0318 20:33:40.449130 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1bc2013-dd47-41c7-a055-d7740d82edfa" containerName="extract-content" Mar 18 20:33:40 crc kubenswrapper[4950]: I0318 20:33:40.449140 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1bc2013-dd47-41c7-a055-d7740d82edfa" containerName="extract-content" Mar 18 20:33:40 crc kubenswrapper[4950]: E0318 20:33:40.449156 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a9d2338-9268-424b-80a8-6401f09165e4" containerName="extract-content" Mar 18 20:33:40 crc kubenswrapper[4950]: I0318 20:33:40.449167 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a9d2338-9268-424b-80a8-6401f09165e4" containerName="extract-content" Mar 18 20:33:40 crc kubenswrapper[4950]: E0318 20:33:40.449211 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1bc2013-dd47-41c7-a055-d7740d82edfa" containerName="registry-server" Mar 18 20:33:40 crc kubenswrapper[4950]: I0318 20:33:40.449221 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1bc2013-dd47-41c7-a055-d7740d82edfa" containerName="registry-server" Mar 18 20:33:40 crc kubenswrapper[4950]: E0318 20:33:40.449235 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1bc2013-dd47-41c7-a055-d7740d82edfa" containerName="extract-utilities" Mar 18 20:33:40 crc kubenswrapper[4950]: I0318 20:33:40.449245 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1bc2013-dd47-41c7-a055-d7740d82edfa" containerName="extract-utilities" Mar 18 20:33:40 crc kubenswrapper[4950]: E0318 20:33:40.449263 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a9d2338-9268-424b-80a8-6401f09165e4" containerName="extract-utilities" Mar 18 20:33:40 crc kubenswrapper[4950]: I0318 20:33:40.449273 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a9d2338-9268-424b-80a8-6401f09165e4" containerName="extract-utilities" Mar 18 20:33:40 crc kubenswrapper[4950]: I0318 20:33:40.449550 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1bc2013-dd47-41c7-a055-d7740d82edfa" containerName="registry-server" Mar 18 20:33:40 crc kubenswrapper[4950]: I0318 20:33:40.449590 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a9d2338-9268-424b-80a8-6401f09165e4" containerName="registry-server" Mar 18 20:33:40 crc kubenswrapper[4950]: I0318 20:33:40.451499 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zgzvl" Mar 18 20:33:40 crc kubenswrapper[4950]: I0318 20:33:40.471155 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zgzvl"] Mar 18 20:33:40 crc kubenswrapper[4950]: I0318 20:33:40.536904 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5acbd9d-4cd5-4704-97f9-6350c12021a3-catalog-content\") pod \"community-operators-zgzvl\" (UID: \"d5acbd9d-4cd5-4704-97f9-6350c12021a3\") " pod="openshift-marketplace/community-operators-zgzvl" Mar 18 20:33:40 crc kubenswrapper[4950]: I0318 20:33:40.536977 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gh2lb\" (UniqueName: \"kubernetes.io/projected/d5acbd9d-4cd5-4704-97f9-6350c12021a3-kube-api-access-gh2lb\") pod \"community-operators-zgzvl\" (UID: \"d5acbd9d-4cd5-4704-97f9-6350c12021a3\") " pod="openshift-marketplace/community-operators-zgzvl" Mar 18 20:33:40 crc kubenswrapper[4950]: I0318 20:33:40.537283 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5acbd9d-4cd5-4704-97f9-6350c12021a3-utilities\") pod \"community-operators-zgzvl\" (UID: \"d5acbd9d-4cd5-4704-97f9-6350c12021a3\") " pod="openshift-marketplace/community-operators-zgzvl" Mar 18 20:33:40 crc kubenswrapper[4950]: I0318 20:33:40.639000 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5acbd9d-4cd5-4704-97f9-6350c12021a3-catalog-content\") pod \"community-operators-zgzvl\" (UID: \"d5acbd9d-4cd5-4704-97f9-6350c12021a3\") " pod="openshift-marketplace/community-operators-zgzvl" Mar 18 20:33:40 crc kubenswrapper[4950]: I0318 20:33:40.639098 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gh2lb\" (UniqueName: \"kubernetes.io/projected/d5acbd9d-4cd5-4704-97f9-6350c12021a3-kube-api-access-gh2lb\") pod \"community-operators-zgzvl\" (UID: \"d5acbd9d-4cd5-4704-97f9-6350c12021a3\") " pod="openshift-marketplace/community-operators-zgzvl" Mar 18 20:33:40 crc kubenswrapper[4950]: I0318 20:33:40.639214 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5acbd9d-4cd5-4704-97f9-6350c12021a3-utilities\") pod \"community-operators-zgzvl\" (UID: \"d5acbd9d-4cd5-4704-97f9-6350c12021a3\") " pod="openshift-marketplace/community-operators-zgzvl" Mar 18 20:33:40 crc kubenswrapper[4950]: I0318 20:33:40.639744 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5acbd9d-4cd5-4704-97f9-6350c12021a3-utilities\") pod \"community-operators-zgzvl\" (UID: \"d5acbd9d-4cd5-4704-97f9-6350c12021a3\") " pod="openshift-marketplace/community-operators-zgzvl" Mar 18 20:33:40 crc kubenswrapper[4950]: I0318 20:33:40.639739 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5acbd9d-4cd5-4704-97f9-6350c12021a3-catalog-content\") pod \"community-operators-zgzvl\" (UID: \"d5acbd9d-4cd5-4704-97f9-6350c12021a3\") " pod="openshift-marketplace/community-operators-zgzvl" Mar 18 20:33:40 crc kubenswrapper[4950]: I0318 20:33:40.663462 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gh2lb\" (UniqueName: \"kubernetes.io/projected/d5acbd9d-4cd5-4704-97f9-6350c12021a3-kube-api-access-gh2lb\") pod \"community-operators-zgzvl\" (UID: \"d5acbd9d-4cd5-4704-97f9-6350c12021a3\") " pod="openshift-marketplace/community-operators-zgzvl" Mar 18 20:33:40 crc kubenswrapper[4950]: I0318 20:33:40.770322 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zgzvl" Mar 18 20:33:41 crc kubenswrapper[4950]: I0318 20:33:41.256342 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zgzvl"] Mar 18 20:33:41 crc kubenswrapper[4950]: I0318 20:33:41.342794 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zgzvl" event={"ID":"d5acbd9d-4cd5-4704-97f9-6350c12021a3","Type":"ContainerStarted","Data":"6b9eb92b0bdcfdd9f869b0a2fd79681ae3fa7c4281dac58ee3026f905dd1d49c"} Mar 18 20:33:42 crc kubenswrapper[4950]: I0318 20:33:42.357711 4950 generic.go:334] "Generic (PLEG): container finished" podID="d5acbd9d-4cd5-4704-97f9-6350c12021a3" containerID="f09e273eacb261e1308c1333e2994fc9be95a63aeb2fba77857e16ff7dc1e169" exitCode=0 Mar 18 20:33:42 crc kubenswrapper[4950]: I0318 20:33:42.357959 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zgzvl" event={"ID":"d5acbd9d-4cd5-4704-97f9-6350c12021a3","Type":"ContainerDied","Data":"f09e273eacb261e1308c1333e2994fc9be95a63aeb2fba77857e16ff7dc1e169"} Mar 18 20:33:44 crc kubenswrapper[4950]: I0318 20:33:44.379065 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zgzvl" event={"ID":"d5acbd9d-4cd5-4704-97f9-6350c12021a3","Type":"ContainerStarted","Data":"b77e8e2a1b9a6b69b9b6f396d3585488d650ecdfdcea4b178feeea1894ce9556"} Mar 18 20:33:46 crc kubenswrapper[4950]: I0318 20:33:46.400480 4950 generic.go:334] "Generic (PLEG): container finished" podID="d5acbd9d-4cd5-4704-97f9-6350c12021a3" containerID="b77e8e2a1b9a6b69b9b6f396d3585488d650ecdfdcea4b178feeea1894ce9556" exitCode=0 Mar 18 20:33:46 crc kubenswrapper[4950]: I0318 20:33:46.400545 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zgzvl" event={"ID":"d5acbd9d-4cd5-4704-97f9-6350c12021a3","Type":"ContainerDied","Data":"b77e8e2a1b9a6b69b9b6f396d3585488d650ecdfdcea4b178feeea1894ce9556"} Mar 18 20:33:47 crc kubenswrapper[4950]: I0318 20:33:47.411081 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zgzvl" event={"ID":"d5acbd9d-4cd5-4704-97f9-6350c12021a3","Type":"ContainerStarted","Data":"e18444048975137296e8f2f7731cbc06d50b6b4b54a9135f7e0e100ba4007a7a"} Mar 18 20:33:47 crc kubenswrapper[4950]: I0318 20:33:47.438487 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zgzvl" podStartSLOduration=2.926922047 podStartE2EDuration="7.438463635s" podCreationTimestamp="2026-03-18 20:33:40 +0000 UTC" firstStartedPulling="2026-03-18 20:33:42.360202783 +0000 UTC m=+1635.601044651" lastFinishedPulling="2026-03-18 20:33:46.871744371 +0000 UTC m=+1640.112586239" observedRunningTime="2026-03-18 20:33:47.42913278 +0000 UTC m=+1640.669974648" watchObservedRunningTime="2026-03-18 20:33:47.438463635 +0000 UTC m=+1640.679305503" Mar 18 20:33:50 crc kubenswrapper[4950]: I0318 20:33:50.770928 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zgzvl" Mar 18 20:33:50 crc kubenswrapper[4950]: I0318 20:33:50.772609 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zgzvl" Mar 18 20:33:50 crc kubenswrapper[4950]: I0318 20:33:50.815353 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zgzvl" Mar 18 20:33:52 crc kubenswrapper[4950]: I0318 20:33:52.529397 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zgzvl" Mar 18 20:33:52 crc kubenswrapper[4950]: I0318 20:33:52.596731 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zgzvl"] Mar 18 20:33:54 crc kubenswrapper[4950]: I0318 20:33:54.494999 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zgzvl" podUID="d5acbd9d-4cd5-4704-97f9-6350c12021a3" containerName="registry-server" containerID="cri-o://e18444048975137296e8f2f7731cbc06d50b6b4b54a9135f7e0e100ba4007a7a" gracePeriod=2 Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.016093 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zgzvl" Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.147433 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5acbd9d-4cd5-4704-97f9-6350c12021a3-catalog-content\") pod \"d5acbd9d-4cd5-4704-97f9-6350c12021a3\" (UID: \"d5acbd9d-4cd5-4704-97f9-6350c12021a3\") " Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.147527 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gh2lb\" (UniqueName: \"kubernetes.io/projected/d5acbd9d-4cd5-4704-97f9-6350c12021a3-kube-api-access-gh2lb\") pod \"d5acbd9d-4cd5-4704-97f9-6350c12021a3\" (UID: \"d5acbd9d-4cd5-4704-97f9-6350c12021a3\") " Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.147556 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5acbd9d-4cd5-4704-97f9-6350c12021a3-utilities\") pod \"d5acbd9d-4cd5-4704-97f9-6350c12021a3\" (UID: \"d5acbd9d-4cd5-4704-97f9-6350c12021a3\") " Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.148376 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5acbd9d-4cd5-4704-97f9-6350c12021a3-utilities" (OuterVolumeSpecName: "utilities") pod "d5acbd9d-4cd5-4704-97f9-6350c12021a3" (UID: "d5acbd9d-4cd5-4704-97f9-6350c12021a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.153573 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5acbd9d-4cd5-4704-97f9-6350c12021a3-kube-api-access-gh2lb" (OuterVolumeSpecName: "kube-api-access-gh2lb") pod "d5acbd9d-4cd5-4704-97f9-6350c12021a3" (UID: "d5acbd9d-4cd5-4704-97f9-6350c12021a3"). InnerVolumeSpecName "kube-api-access-gh2lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.205263 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5acbd9d-4cd5-4704-97f9-6350c12021a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d5acbd9d-4cd5-4704-97f9-6350c12021a3" (UID: "d5acbd9d-4cd5-4704-97f9-6350c12021a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.249122 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5acbd9d-4cd5-4704-97f9-6350c12021a3-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.249150 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gh2lb\" (UniqueName: \"kubernetes.io/projected/d5acbd9d-4cd5-4704-97f9-6350c12021a3-kube-api-access-gh2lb\") on node \"crc\" DevicePath \"\"" Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.249164 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5acbd9d-4cd5-4704-97f9-6350c12021a3-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.506962 4950 generic.go:334] "Generic (PLEG): container finished" podID="d5acbd9d-4cd5-4704-97f9-6350c12021a3" containerID="e18444048975137296e8f2f7731cbc06d50b6b4b54a9135f7e0e100ba4007a7a" exitCode=0 Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.507004 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zgzvl" event={"ID":"d5acbd9d-4cd5-4704-97f9-6350c12021a3","Type":"ContainerDied","Data":"e18444048975137296e8f2f7731cbc06d50b6b4b54a9135f7e0e100ba4007a7a"} Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.507028 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zgzvl" event={"ID":"d5acbd9d-4cd5-4704-97f9-6350c12021a3","Type":"ContainerDied","Data":"6b9eb92b0bdcfdd9f869b0a2fd79681ae3fa7c4281dac58ee3026f905dd1d49c"} Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.507044 4950 scope.go:117] "RemoveContainer" containerID="e18444048975137296e8f2f7731cbc06d50b6b4b54a9135f7e0e100ba4007a7a" Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.507165 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zgzvl" Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.549673 4950 scope.go:117] "RemoveContainer" containerID="b77e8e2a1b9a6b69b9b6f396d3585488d650ecdfdcea4b178feeea1894ce9556" Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.568376 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zgzvl"] Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.575564 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zgzvl"] Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.586945 4950 scope.go:117] "RemoveContainer" containerID="f09e273eacb261e1308c1333e2994fc9be95a63aeb2fba77857e16ff7dc1e169" Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.633910 4950 scope.go:117] "RemoveContainer" containerID="e18444048975137296e8f2f7731cbc06d50b6b4b54a9135f7e0e100ba4007a7a" Mar 18 20:33:55 crc kubenswrapper[4950]: E0318 20:33:55.634996 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e18444048975137296e8f2f7731cbc06d50b6b4b54a9135f7e0e100ba4007a7a\": container with ID starting with e18444048975137296e8f2f7731cbc06d50b6b4b54a9135f7e0e100ba4007a7a not found: ID does not exist" containerID="e18444048975137296e8f2f7731cbc06d50b6b4b54a9135f7e0e100ba4007a7a" Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.635039 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e18444048975137296e8f2f7731cbc06d50b6b4b54a9135f7e0e100ba4007a7a"} err="failed to get container status \"e18444048975137296e8f2f7731cbc06d50b6b4b54a9135f7e0e100ba4007a7a\": rpc error: code = NotFound desc = could not find container \"e18444048975137296e8f2f7731cbc06d50b6b4b54a9135f7e0e100ba4007a7a\": container with ID starting with e18444048975137296e8f2f7731cbc06d50b6b4b54a9135f7e0e100ba4007a7a not found: ID does not exist" Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.635066 4950 scope.go:117] "RemoveContainer" containerID="b77e8e2a1b9a6b69b9b6f396d3585488d650ecdfdcea4b178feeea1894ce9556" Mar 18 20:33:55 crc kubenswrapper[4950]: E0318 20:33:55.635560 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b77e8e2a1b9a6b69b9b6f396d3585488d650ecdfdcea4b178feeea1894ce9556\": container with ID starting with b77e8e2a1b9a6b69b9b6f396d3585488d650ecdfdcea4b178feeea1894ce9556 not found: ID does not exist" containerID="b77e8e2a1b9a6b69b9b6f396d3585488d650ecdfdcea4b178feeea1894ce9556" Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.635597 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b77e8e2a1b9a6b69b9b6f396d3585488d650ecdfdcea4b178feeea1894ce9556"} err="failed to get container status \"b77e8e2a1b9a6b69b9b6f396d3585488d650ecdfdcea4b178feeea1894ce9556\": rpc error: code = NotFound desc = could not find container \"b77e8e2a1b9a6b69b9b6f396d3585488d650ecdfdcea4b178feeea1894ce9556\": container with ID starting with b77e8e2a1b9a6b69b9b6f396d3585488d650ecdfdcea4b178feeea1894ce9556 not found: ID does not exist" Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.635622 4950 scope.go:117] "RemoveContainer" containerID="f09e273eacb261e1308c1333e2994fc9be95a63aeb2fba77857e16ff7dc1e169" Mar 18 20:33:55 crc kubenswrapper[4950]: E0318 20:33:55.635893 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f09e273eacb261e1308c1333e2994fc9be95a63aeb2fba77857e16ff7dc1e169\": container with ID starting with f09e273eacb261e1308c1333e2994fc9be95a63aeb2fba77857e16ff7dc1e169 not found: ID does not exist" containerID="f09e273eacb261e1308c1333e2994fc9be95a63aeb2fba77857e16ff7dc1e169" Mar 18 20:33:55 crc kubenswrapper[4950]: I0318 20:33:55.635919 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f09e273eacb261e1308c1333e2994fc9be95a63aeb2fba77857e16ff7dc1e169"} err="failed to get container status \"f09e273eacb261e1308c1333e2994fc9be95a63aeb2fba77857e16ff7dc1e169\": rpc error: code = NotFound desc = could not find container \"f09e273eacb261e1308c1333e2994fc9be95a63aeb2fba77857e16ff7dc1e169\": container with ID starting with f09e273eacb261e1308c1333e2994fc9be95a63aeb2fba77857e16ff7dc1e169 not found: ID does not exist" Mar 18 20:33:57 crc kubenswrapper[4950]: I0318 20:33:57.495351 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5acbd9d-4cd5-4704-97f9-6350c12021a3" path="/var/lib/kubelet/pods/d5acbd9d-4cd5-4704-97f9-6350c12021a3/volumes" Mar 18 20:33:58 crc kubenswrapper[4950]: I0318 20:33:58.544019 4950 generic.go:334] "Generic (PLEG): container finished" podID="a1b58e52-1dea-4216-9eec-ebf5c09fba32" containerID="74242ce84c406b044ae39cf55d1dea785c786d4e82138ad9b68e2eedf3cc8061" exitCode=0 Mar 18 20:33:58 crc kubenswrapper[4950]: I0318 20:33:58.544393 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh" event={"ID":"a1b58e52-1dea-4216-9eec-ebf5c09fba32","Type":"ContainerDied","Data":"74242ce84c406b044ae39cf55d1dea785c786d4e82138ad9b68e2eedf3cc8061"} Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.014340 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.141475 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564434-ps98x"] Mar 18 20:34:00 crc kubenswrapper[4950]: E0318 20:34:00.142216 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5acbd9d-4cd5-4704-97f9-6350c12021a3" containerName="extract-content" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.142311 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5acbd9d-4cd5-4704-97f9-6350c12021a3" containerName="extract-content" Mar 18 20:34:00 crc kubenswrapper[4950]: E0318 20:34:00.142404 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5acbd9d-4cd5-4704-97f9-6350c12021a3" containerName="extract-utilities" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.142520 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5acbd9d-4cd5-4704-97f9-6350c12021a3" containerName="extract-utilities" Mar 18 20:34:00 crc kubenswrapper[4950]: E0318 20:34:00.142597 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1b58e52-1dea-4216-9eec-ebf5c09fba32" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.142699 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1b58e52-1dea-4216-9eec-ebf5c09fba32" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 18 20:34:00 crc kubenswrapper[4950]: E0318 20:34:00.142776 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5acbd9d-4cd5-4704-97f9-6350c12021a3" containerName="registry-server" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.142845 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5acbd9d-4cd5-4704-97f9-6350c12021a3" containerName="registry-server" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.141720 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1b58e52-1dea-4216-9eec-ebf5c09fba32-inventory\") pod \"a1b58e52-1dea-4216-9eec-ebf5c09fba32\" (UID: \"a1b58e52-1dea-4216-9eec-ebf5c09fba32\") " Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.143086 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a1b58e52-1dea-4216-9eec-ebf5c09fba32-ssh-key-openstack-edpm-ipam\") pod \"a1b58e52-1dea-4216-9eec-ebf5c09fba32\" (UID: \"a1b58e52-1dea-4216-9eec-ebf5c09fba32\") " Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.143135 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jv69q\" (UniqueName: \"kubernetes.io/projected/a1b58e52-1dea-4216-9eec-ebf5c09fba32-kube-api-access-jv69q\") pod \"a1b58e52-1dea-4216-9eec-ebf5c09fba32\" (UID: \"a1b58e52-1dea-4216-9eec-ebf5c09fba32\") " Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.143196 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b58e52-1dea-4216-9eec-ebf5c09fba32-bootstrap-combined-ca-bundle\") pod \"a1b58e52-1dea-4216-9eec-ebf5c09fba32\" (UID: \"a1b58e52-1dea-4216-9eec-ebf5c09fba32\") " Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.143350 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5acbd9d-4cd5-4704-97f9-6350c12021a3" containerName="registry-server" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.143453 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1b58e52-1dea-4216-9eec-ebf5c09fba32" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.144071 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564434-ps98x" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.155908 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.156249 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.159607 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1b58e52-1dea-4216-9eec-ebf5c09fba32-kube-api-access-jv69q" (OuterVolumeSpecName: "kube-api-access-jv69q") pod "a1b58e52-1dea-4216-9eec-ebf5c09fba32" (UID: "a1b58e52-1dea-4216-9eec-ebf5c09fba32"). InnerVolumeSpecName "kube-api-access-jv69q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.163846 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1b58e52-1dea-4216-9eec-ebf5c09fba32-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "a1b58e52-1dea-4216-9eec-ebf5c09fba32" (UID: "a1b58e52-1dea-4216-9eec-ebf5c09fba32"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.165130 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.167227 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564434-ps98x"] Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.167866 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1b58e52-1dea-4216-9eec-ebf5c09fba32-inventory" (OuterVolumeSpecName: "inventory") pod "a1b58e52-1dea-4216-9eec-ebf5c09fba32" (UID: "a1b58e52-1dea-4216-9eec-ebf5c09fba32"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.186479 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1b58e52-1dea-4216-9eec-ebf5c09fba32-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a1b58e52-1dea-4216-9eec-ebf5c09fba32" (UID: "a1b58e52-1dea-4216-9eec-ebf5c09fba32"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.245351 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8jrs\" (UniqueName: \"kubernetes.io/projected/e3599ae0-de48-40b8-979f-7f7125584a40-kube-api-access-t8jrs\") pod \"auto-csr-approver-29564434-ps98x\" (UID: \"e3599ae0-de48-40b8-979f-7f7125584a40\") " pod="openshift-infra/auto-csr-approver-29564434-ps98x" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.245714 4950 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b58e52-1dea-4216-9eec-ebf5c09fba32-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.245738 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1b58e52-1dea-4216-9eec-ebf5c09fba32-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.245752 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a1b58e52-1dea-4216-9eec-ebf5c09fba32-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.245765 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jv69q\" (UniqueName: \"kubernetes.io/projected/a1b58e52-1dea-4216-9eec-ebf5c09fba32-kube-api-access-jv69q\") on node \"crc\" DevicePath \"\"" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.350690 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8jrs\" (UniqueName: \"kubernetes.io/projected/e3599ae0-de48-40b8-979f-7f7125584a40-kube-api-access-t8jrs\") pod \"auto-csr-approver-29564434-ps98x\" (UID: \"e3599ae0-de48-40b8-979f-7f7125584a40\") " pod="openshift-infra/auto-csr-approver-29564434-ps98x" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.368629 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8jrs\" (UniqueName: \"kubernetes.io/projected/e3599ae0-de48-40b8-979f-7f7125584a40-kube-api-access-t8jrs\") pod \"auto-csr-approver-29564434-ps98x\" (UID: \"e3599ae0-de48-40b8-979f-7f7125584a40\") " pod="openshift-infra/auto-csr-approver-29564434-ps98x" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.545953 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564434-ps98x" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.574902 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh" event={"ID":"a1b58e52-1dea-4216-9eec-ebf5c09fba32","Type":"ContainerDied","Data":"565f6461e9ec1a067a2c3c2219485a95f19eae8ad4dee6a328e4cb0189098849"} Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.575383 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="565f6461e9ec1a067a2c3c2219485a95f19eae8ad4dee6a328e4cb0189098849" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.575087 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.675010 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vngd5"] Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.676550 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vngd5" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.679673 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.679908 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.680127 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.680315 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.692310 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vngd5"] Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.860502 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vngd5\" (UID: \"907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vngd5" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.860699 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hztx\" (UniqueName: \"kubernetes.io/projected/907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb-kube-api-access-2hztx\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vngd5\" (UID: \"907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vngd5" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.861241 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vngd5\" (UID: \"907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vngd5" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.962967 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vngd5\" (UID: \"907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vngd5" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.963698 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vngd5\" (UID: \"907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vngd5" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.963768 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hztx\" (UniqueName: \"kubernetes.io/projected/907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb-kube-api-access-2hztx\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vngd5\" (UID: \"907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vngd5" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.969758 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vngd5\" (UID: \"907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vngd5" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.970579 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vngd5\" (UID: \"907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vngd5" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.993998 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hztx\" (UniqueName: \"kubernetes.io/projected/907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb-kube-api-access-2hztx\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vngd5\" (UID: \"907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vngd5" Mar 18 20:34:00 crc kubenswrapper[4950]: I0318 20:34:00.997876 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vngd5" Mar 18 20:34:01 crc kubenswrapper[4950]: I0318 20:34:01.079899 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564434-ps98x"] Mar 18 20:34:01 crc kubenswrapper[4950]: W0318 20:34:01.110959 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3599ae0_de48_40b8_979f_7f7125584a40.slice/crio-e0ae5f337f096ca22c069537cc8fab039ea338964ccbbb8b9cc1af3b42b7e6a9 WatchSource:0}: Error finding container e0ae5f337f096ca22c069537cc8fab039ea338964ccbbb8b9cc1af3b42b7e6a9: Status 404 returned error can't find the container with id e0ae5f337f096ca22c069537cc8fab039ea338964ccbbb8b9cc1af3b42b7e6a9 Mar 18 20:34:01 crc kubenswrapper[4950]: I0318 20:34:01.497588 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vngd5"] Mar 18 20:34:01 crc kubenswrapper[4950]: W0318 20:34:01.500860 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod907a6ea9_e0ab_46d7_a2ca_c92f98ca0adb.slice/crio-66f6076c6b70fbda69bec40f361ad32d29e6154316665a9db97a64c9aeeaff1d WatchSource:0}: Error finding container 66f6076c6b70fbda69bec40f361ad32d29e6154316665a9db97a64c9aeeaff1d: Status 404 returned error can't find the container with id 66f6076c6b70fbda69bec40f361ad32d29e6154316665a9db97a64c9aeeaff1d Mar 18 20:34:01 crc kubenswrapper[4950]: I0318 20:34:01.587816 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vngd5" event={"ID":"907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb","Type":"ContainerStarted","Data":"66f6076c6b70fbda69bec40f361ad32d29e6154316665a9db97a64c9aeeaff1d"} Mar 18 20:34:01 crc kubenswrapper[4950]: I0318 20:34:01.589677 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564434-ps98x" event={"ID":"e3599ae0-de48-40b8-979f-7f7125584a40","Type":"ContainerStarted","Data":"e0ae5f337f096ca22c069537cc8fab039ea338964ccbbb8b9cc1af3b42b7e6a9"} Mar 18 20:34:02 crc kubenswrapper[4950]: I0318 20:34:02.602623 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vngd5" event={"ID":"907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb","Type":"ContainerStarted","Data":"768b95b44dadfe1e523ff513f3b2412fb537fab5f4e51653e0a1847d11c92a49"} Mar 18 20:34:02 crc kubenswrapper[4950]: I0318 20:34:02.607842 4950 generic.go:334] "Generic (PLEG): container finished" podID="e3599ae0-de48-40b8-979f-7f7125584a40" containerID="3bb623dccda8a730d1aa964444c4ce4a4a8f3bbbae5d40443ac00cf6665496ca" exitCode=0 Mar 18 20:34:02 crc kubenswrapper[4950]: I0318 20:34:02.608060 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564434-ps98x" event={"ID":"e3599ae0-de48-40b8-979f-7f7125584a40","Type":"ContainerDied","Data":"3bb623dccda8a730d1aa964444c4ce4a4a8f3bbbae5d40443ac00cf6665496ca"} Mar 18 20:34:02 crc kubenswrapper[4950]: I0318 20:34:02.631465 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vngd5" podStartSLOduration=2.1440155340000002 podStartE2EDuration="2.631448071s" podCreationTimestamp="2026-03-18 20:34:00 +0000 UTC" firstStartedPulling="2026-03-18 20:34:01.503342513 +0000 UTC m=+1654.744184381" lastFinishedPulling="2026-03-18 20:34:01.99077503 +0000 UTC m=+1655.231616918" observedRunningTime="2026-03-18 20:34:02.617585476 +0000 UTC m=+1655.858427354" watchObservedRunningTime="2026-03-18 20:34:02.631448071 +0000 UTC m=+1655.872289939" Mar 18 20:34:03 crc kubenswrapper[4950]: I0318 20:34:03.964927 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564434-ps98x" Mar 18 20:34:04 crc kubenswrapper[4950]: I0318 20:34:04.167910 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8jrs\" (UniqueName: \"kubernetes.io/projected/e3599ae0-de48-40b8-979f-7f7125584a40-kube-api-access-t8jrs\") pod \"e3599ae0-de48-40b8-979f-7f7125584a40\" (UID: \"e3599ae0-de48-40b8-979f-7f7125584a40\") " Mar 18 20:34:04 crc kubenswrapper[4950]: I0318 20:34:04.174186 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3599ae0-de48-40b8-979f-7f7125584a40-kube-api-access-t8jrs" (OuterVolumeSpecName: "kube-api-access-t8jrs") pod "e3599ae0-de48-40b8-979f-7f7125584a40" (UID: "e3599ae0-de48-40b8-979f-7f7125584a40"). InnerVolumeSpecName "kube-api-access-t8jrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:34:04 crc kubenswrapper[4950]: I0318 20:34:04.270627 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8jrs\" (UniqueName: \"kubernetes.io/projected/e3599ae0-de48-40b8-979f-7f7125584a40-kube-api-access-t8jrs\") on node \"crc\" DevicePath \"\"" Mar 18 20:34:04 crc kubenswrapper[4950]: I0318 20:34:04.634396 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564434-ps98x" event={"ID":"e3599ae0-de48-40b8-979f-7f7125584a40","Type":"ContainerDied","Data":"e0ae5f337f096ca22c069537cc8fab039ea338964ccbbb8b9cc1af3b42b7e6a9"} Mar 18 20:34:04 crc kubenswrapper[4950]: I0318 20:34:04.634484 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0ae5f337f096ca22c069537cc8fab039ea338964ccbbb8b9cc1af3b42b7e6a9" Mar 18 20:34:04 crc kubenswrapper[4950]: I0318 20:34:04.634500 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564434-ps98x" Mar 18 20:34:05 crc kubenswrapper[4950]: I0318 20:34:05.049287 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564428-7k4b9"] Mar 18 20:34:05 crc kubenswrapper[4950]: I0318 20:34:05.058669 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564428-7k4b9"] Mar 18 20:34:05 crc kubenswrapper[4950]: I0318 20:34:05.501180 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2548eb6-5b6d-4ffd-9655-a766eecf2128" path="/var/lib/kubelet/pods/d2548eb6-5b6d-4ffd-9655-a766eecf2128/volumes" Mar 18 20:34:46 crc kubenswrapper[4950]: I0318 20:34:46.693917 4950 scope.go:117] "RemoveContainer" containerID="e3dcff1971656e6b2a038d556bf18d138c9cfff265d2ca9e52f845ed6aa55b08" Mar 18 20:34:46 crc kubenswrapper[4950]: I0318 20:34:46.729326 4950 scope.go:117] "RemoveContainer" containerID="b646686c45c126d79b65dc5a0bf87aa8f95eac390bc0427a5b4e6f5ac1fc3993" Mar 18 20:34:49 crc kubenswrapper[4950]: I0318 20:34:49.072200 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-mz5gw"] Mar 18 20:34:49 crc kubenswrapper[4950]: I0318 20:34:49.081298 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-mz5gw"] Mar 18 20:34:49 crc kubenswrapper[4950]: I0318 20:34:49.489640 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01a1bd17-8d39-4435-abe7-15617b6566ec" path="/var/lib/kubelet/pods/01a1bd17-8d39-4435-abe7-15617b6566ec/volumes" Mar 18 20:34:50 crc kubenswrapper[4950]: I0318 20:34:50.037280 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-f5c3-account-create-update-m5jc8"] Mar 18 20:34:50 crc kubenswrapper[4950]: I0318 20:34:50.045329 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-cefd-account-create-update-ppvt2"] Mar 18 20:34:50 crc kubenswrapper[4950]: I0318 20:34:50.053373 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-n55x2"] Mar 18 20:34:50 crc kubenswrapper[4950]: I0318 20:34:50.060850 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-cefd-account-create-update-ppvt2"] Mar 18 20:34:50 crc kubenswrapper[4950]: I0318 20:34:50.070852 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-f5c3-account-create-update-m5jc8"] Mar 18 20:34:50 crc kubenswrapper[4950]: I0318 20:34:50.078464 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-n55x2"] Mar 18 20:34:51 crc kubenswrapper[4950]: I0318 20:34:51.492325 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44e46c00-4ccb-4739-89c2-df4872915c1f" path="/var/lib/kubelet/pods/44e46c00-4ccb-4739-89c2-df4872915c1f/volumes" Mar 18 20:34:51 crc kubenswrapper[4950]: I0318 20:34:51.495099 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58d9f568-de35-40a7-be22-4ab9e138791f" path="/var/lib/kubelet/pods/58d9f568-de35-40a7-be22-4ab9e138791f/volumes" Mar 18 20:34:51 crc kubenswrapper[4950]: I0318 20:34:51.496278 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8f7b4ae-81da-49ac-978c-d6e0ff3c839e" path="/var/lib/kubelet/pods/e8f7b4ae-81da-49ac-978c-d6e0ff3c839e/volumes" Mar 18 20:34:52 crc kubenswrapper[4950]: I0318 20:34:52.031037 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-aa88-account-create-update-78r4v"] Mar 18 20:34:52 crc kubenswrapper[4950]: I0318 20:34:52.039359 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-2kh85"] Mar 18 20:34:52 crc kubenswrapper[4950]: I0318 20:34:52.049536 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-2kh85"] Mar 18 20:34:52 crc kubenswrapper[4950]: I0318 20:34:52.059092 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-aa88-account-create-update-78r4v"] Mar 18 20:34:53 crc kubenswrapper[4950]: I0318 20:34:53.492124 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99f9366a-2413-4243-9196-95088cf66e87" path="/var/lib/kubelet/pods/99f9366a-2413-4243-9196-95088cf66e87/volumes" Mar 18 20:34:53 crc kubenswrapper[4950]: I0318 20:34:53.493247 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c89c3ec6-2109-4bd7-92e4-5d872dc95cca" path="/var/lib/kubelet/pods/c89c3ec6-2109-4bd7-92e4-5d872dc95cca/volumes" Mar 18 20:35:03 crc kubenswrapper[4950]: I0318 20:35:03.836204 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:35:03 crc kubenswrapper[4950]: I0318 20:35:03.836838 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:35:11 crc kubenswrapper[4950]: I0318 20:35:11.322634 4950 generic.go:334] "Generic (PLEG): container finished" podID="907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb" containerID="768b95b44dadfe1e523ff513f3b2412fb537fab5f4e51653e0a1847d11c92a49" exitCode=0 Mar 18 20:35:11 crc kubenswrapper[4950]: I0318 20:35:11.322715 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vngd5" event={"ID":"907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb","Type":"ContainerDied","Data":"768b95b44dadfe1e523ff513f3b2412fb537fab5f4e51653e0a1847d11c92a49"} Mar 18 20:35:12 crc kubenswrapper[4950]: I0318 20:35:12.038006 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-2n7qm"] Mar 18 20:35:12 crc kubenswrapper[4950]: I0318 20:35:12.052679 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-2n7qm"] Mar 18 20:35:12 crc kubenswrapper[4950]: I0318 20:35:12.720071 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vngd5" Mar 18 20:35:12 crc kubenswrapper[4950]: I0318 20:35:12.907057 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hztx\" (UniqueName: \"kubernetes.io/projected/907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb-kube-api-access-2hztx\") pod \"907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb\" (UID: \"907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb\") " Mar 18 20:35:12 crc kubenswrapper[4950]: I0318 20:35:12.907370 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb-ssh-key-openstack-edpm-ipam\") pod \"907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb\" (UID: \"907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb\") " Mar 18 20:35:12 crc kubenswrapper[4950]: I0318 20:35:12.907483 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb-inventory\") pod \"907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb\" (UID: \"907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb\") " Mar 18 20:35:12 crc kubenswrapper[4950]: I0318 20:35:12.915678 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb-kube-api-access-2hztx" (OuterVolumeSpecName: "kube-api-access-2hztx") pod "907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb" (UID: "907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb"). InnerVolumeSpecName "kube-api-access-2hztx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:35:12 crc kubenswrapper[4950]: I0318 20:35:12.931973 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb-inventory" (OuterVolumeSpecName: "inventory") pod "907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb" (UID: "907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:35:12 crc kubenswrapper[4950]: I0318 20:35:12.937536 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb" (UID: "907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.009624 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hztx\" (UniqueName: \"kubernetes.io/projected/907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb-kube-api-access-2hztx\") on node \"crc\" DevicePath \"\"" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.009661 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.009672 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.350161 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vngd5" event={"ID":"907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb","Type":"ContainerDied","Data":"66f6076c6b70fbda69bec40f361ad32d29e6154316665a9db97a64c9aeeaff1d"} Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.350466 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66f6076c6b70fbda69bec40f361ad32d29e6154316665a9db97a64c9aeeaff1d" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.350560 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vngd5" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.442515 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6gml"] Mar 18 20:35:13 crc kubenswrapper[4950]: E0318 20:35:13.442966 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3599ae0-de48-40b8-979f-7f7125584a40" containerName="oc" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.442985 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3599ae0-de48-40b8-979f-7f7125584a40" containerName="oc" Mar 18 20:35:13 crc kubenswrapper[4950]: E0318 20:35:13.443003 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.443013 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.443264 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.443285 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3599ae0-de48-40b8-979f-7f7125584a40" containerName="oc" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.444051 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6gml" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.445955 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.446291 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.446358 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.448723 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.450509 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6gml"] Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.489820 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="884d39cd-355d-491b-ac02-b47295da532f" path="/var/lib/kubelet/pods/884d39cd-355d-491b-ac02-b47295da532f/volumes" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.622098 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8b1fb431-7ff9-498a-a4b3-352822335933-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x6gml\" (UID: \"8b1fb431-7ff9-498a-a4b3-352822335933\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6gml" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.622545 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blx6r\" (UniqueName: \"kubernetes.io/projected/8b1fb431-7ff9-498a-a4b3-352822335933-kube-api-access-blx6r\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x6gml\" (UID: \"8b1fb431-7ff9-498a-a4b3-352822335933\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6gml" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.622668 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b1fb431-7ff9-498a-a4b3-352822335933-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x6gml\" (UID: \"8b1fb431-7ff9-498a-a4b3-352822335933\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6gml" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.724364 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8b1fb431-7ff9-498a-a4b3-352822335933-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x6gml\" (UID: \"8b1fb431-7ff9-498a-a4b3-352822335933\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6gml" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.724598 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blx6r\" (UniqueName: \"kubernetes.io/projected/8b1fb431-7ff9-498a-a4b3-352822335933-kube-api-access-blx6r\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x6gml\" (UID: \"8b1fb431-7ff9-498a-a4b3-352822335933\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6gml" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.724662 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b1fb431-7ff9-498a-a4b3-352822335933-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x6gml\" (UID: \"8b1fb431-7ff9-498a-a4b3-352822335933\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6gml" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.730200 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b1fb431-7ff9-498a-a4b3-352822335933-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x6gml\" (UID: \"8b1fb431-7ff9-498a-a4b3-352822335933\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6gml" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.731011 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8b1fb431-7ff9-498a-a4b3-352822335933-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x6gml\" (UID: \"8b1fb431-7ff9-498a-a4b3-352822335933\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6gml" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.742730 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blx6r\" (UniqueName: \"kubernetes.io/projected/8b1fb431-7ff9-498a-a4b3-352822335933-kube-api-access-blx6r\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x6gml\" (UID: \"8b1fb431-7ff9-498a-a4b3-352822335933\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6gml" Mar 18 20:35:13 crc kubenswrapper[4950]: I0318 20:35:13.763272 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6gml" Mar 18 20:35:14 crc kubenswrapper[4950]: I0318 20:35:14.345653 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6gml"] Mar 18 20:35:14 crc kubenswrapper[4950]: I0318 20:35:14.363582 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6gml" event={"ID":"8b1fb431-7ff9-498a-a4b3-352822335933","Type":"ContainerStarted","Data":"bbda7a2c4123b638e9f76decf0e0a16c5e106057dc19644384586ceecc2e3189"} Mar 18 20:35:15 crc kubenswrapper[4950]: I0318 20:35:15.381326 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6gml" event={"ID":"8b1fb431-7ff9-498a-a4b3-352822335933","Type":"ContainerStarted","Data":"5ecdc9e1d79118ed35463ec1bd847a70d5c6f6f6aa7cffa37d8cc10e9554ba4d"} Mar 18 20:35:15 crc kubenswrapper[4950]: I0318 20:35:15.410166 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6gml" podStartSLOduration=1.8834606630000001 podStartE2EDuration="2.410133323s" podCreationTimestamp="2026-03-18 20:35:13 +0000 UTC" firstStartedPulling="2026-03-18 20:35:14.352572352 +0000 UTC m=+1727.593414220" lastFinishedPulling="2026-03-18 20:35:14.879244972 +0000 UTC m=+1728.120086880" observedRunningTime="2026-03-18 20:35:15.398741284 +0000 UTC m=+1728.639583162" watchObservedRunningTime="2026-03-18 20:35:15.410133323 +0000 UTC m=+1728.650975211" Mar 18 20:35:20 crc kubenswrapper[4950]: I0318 20:35:20.040732 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-sr482"] Mar 18 20:35:20 crc kubenswrapper[4950]: I0318 20:35:20.048430 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-sr482"] Mar 18 20:35:21 crc kubenswrapper[4950]: I0318 20:35:21.451934 4950 generic.go:334] "Generic (PLEG): container finished" podID="8b1fb431-7ff9-498a-a4b3-352822335933" containerID="5ecdc9e1d79118ed35463ec1bd847a70d5c6f6f6aa7cffa37d8cc10e9554ba4d" exitCode=0 Mar 18 20:35:21 crc kubenswrapper[4950]: I0318 20:35:21.452022 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6gml" event={"ID":"8b1fb431-7ff9-498a-a4b3-352822335933","Type":"ContainerDied","Data":"5ecdc9e1d79118ed35463ec1bd847a70d5c6f6f6aa7cffa37d8cc10e9554ba4d"} Mar 18 20:35:21 crc kubenswrapper[4950]: I0318 20:35:21.488677 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32f194c3-aab7-4bf3-ae91-2a7051efc958" path="/var/lib/kubelet/pods/32f194c3-aab7-4bf3-ae91-2a7051efc958/volumes" Mar 18 20:35:22 crc kubenswrapper[4950]: I0318 20:35:22.882729 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6gml" Mar 18 20:35:22 crc kubenswrapper[4950]: I0318 20:35:22.996999 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8b1fb431-7ff9-498a-a4b3-352822335933-ssh-key-openstack-edpm-ipam\") pod \"8b1fb431-7ff9-498a-a4b3-352822335933\" (UID: \"8b1fb431-7ff9-498a-a4b3-352822335933\") " Mar 18 20:35:22 crc kubenswrapper[4950]: I0318 20:35:22.997094 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b1fb431-7ff9-498a-a4b3-352822335933-inventory\") pod \"8b1fb431-7ff9-498a-a4b3-352822335933\" (UID: \"8b1fb431-7ff9-498a-a4b3-352822335933\") " Mar 18 20:35:22 crc kubenswrapper[4950]: I0318 20:35:22.997293 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blx6r\" (UniqueName: \"kubernetes.io/projected/8b1fb431-7ff9-498a-a4b3-352822335933-kube-api-access-blx6r\") pod \"8b1fb431-7ff9-498a-a4b3-352822335933\" (UID: \"8b1fb431-7ff9-498a-a4b3-352822335933\") " Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.004191 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b1fb431-7ff9-498a-a4b3-352822335933-kube-api-access-blx6r" (OuterVolumeSpecName: "kube-api-access-blx6r") pod "8b1fb431-7ff9-498a-a4b3-352822335933" (UID: "8b1fb431-7ff9-498a-a4b3-352822335933"). InnerVolumeSpecName "kube-api-access-blx6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.028401 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b1fb431-7ff9-498a-a4b3-352822335933-inventory" (OuterVolumeSpecName: "inventory") pod "8b1fb431-7ff9-498a-a4b3-352822335933" (UID: "8b1fb431-7ff9-498a-a4b3-352822335933"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.030952 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b1fb431-7ff9-498a-a4b3-352822335933-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8b1fb431-7ff9-498a-a4b3-352822335933" (UID: "8b1fb431-7ff9-498a-a4b3-352822335933"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.099167 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blx6r\" (UniqueName: \"kubernetes.io/projected/8b1fb431-7ff9-498a-a4b3-352822335933-kube-api-access-blx6r\") on node \"crc\" DevicePath \"\"" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.099196 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8b1fb431-7ff9-498a-a4b3-352822335933-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.099206 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b1fb431-7ff9-498a-a4b3-352822335933-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.472405 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6gml" event={"ID":"8b1fb431-7ff9-498a-a4b3-352822335933","Type":"ContainerDied","Data":"bbda7a2c4123b638e9f76decf0e0a16c5e106057dc19644384586ceecc2e3189"} Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.472611 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbda7a2c4123b638e9f76decf0e0a16c5e106057dc19644384586ceecc2e3189" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.472490 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6gml" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.548253 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-xhfbh"] Mar 18 20:35:23 crc kubenswrapper[4950]: E0318 20:35:23.549314 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b1fb431-7ff9-498a-a4b3-352822335933" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.549458 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b1fb431-7ff9-498a-a4b3-352822335933" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.549901 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b1fb431-7ff9-498a-a4b3-352822335933" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.550738 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xhfbh" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.555898 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.557594 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.557720 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.557757 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.584819 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-xhfbh"] Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.713535 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/84bce393-6964-42d4-9126-076287a2dece-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xhfbh\" (UID: \"84bce393-6964-42d4-9126-076287a2dece\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xhfbh" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.714039 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kl2n\" (UniqueName: \"kubernetes.io/projected/84bce393-6964-42d4-9126-076287a2dece-kube-api-access-5kl2n\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xhfbh\" (UID: \"84bce393-6964-42d4-9126-076287a2dece\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xhfbh" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.714091 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84bce393-6964-42d4-9126-076287a2dece-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xhfbh\" (UID: \"84bce393-6964-42d4-9126-076287a2dece\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xhfbh" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.815808 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kl2n\" (UniqueName: \"kubernetes.io/projected/84bce393-6964-42d4-9126-076287a2dece-kube-api-access-5kl2n\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xhfbh\" (UID: \"84bce393-6964-42d4-9126-076287a2dece\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xhfbh" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.815867 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84bce393-6964-42d4-9126-076287a2dece-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xhfbh\" (UID: \"84bce393-6964-42d4-9126-076287a2dece\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xhfbh" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.815940 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/84bce393-6964-42d4-9126-076287a2dece-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xhfbh\" (UID: \"84bce393-6964-42d4-9126-076287a2dece\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xhfbh" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.823465 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/84bce393-6964-42d4-9126-076287a2dece-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xhfbh\" (UID: \"84bce393-6964-42d4-9126-076287a2dece\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xhfbh" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.823515 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84bce393-6964-42d4-9126-076287a2dece-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xhfbh\" (UID: \"84bce393-6964-42d4-9126-076287a2dece\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xhfbh" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.831165 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kl2n\" (UniqueName: \"kubernetes.io/projected/84bce393-6964-42d4-9126-076287a2dece-kube-api-access-5kl2n\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xhfbh\" (UID: \"84bce393-6964-42d4-9126-076287a2dece\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xhfbh" Mar 18 20:35:23 crc kubenswrapper[4950]: I0318 20:35:23.879805 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xhfbh" Mar 18 20:35:24 crc kubenswrapper[4950]: I0318 20:35:24.428505 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-xhfbh"] Mar 18 20:35:24 crc kubenswrapper[4950]: I0318 20:35:24.482154 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xhfbh" event={"ID":"84bce393-6964-42d4-9126-076287a2dece","Type":"ContainerStarted","Data":"f4ea30e682c3a55edd6d1bbffca651cde38aa545198728bd00dd96db5dcda913"} Mar 18 20:35:25 crc kubenswrapper[4950]: I0318 20:35:25.497908 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xhfbh" event={"ID":"84bce393-6964-42d4-9126-076287a2dece","Type":"ContainerStarted","Data":"68f707f726c227fe94bff58052b50aead3ac2b367f00a474c05bb6f61be620d7"} Mar 18 20:35:25 crc kubenswrapper[4950]: I0318 20:35:25.525562 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xhfbh" podStartSLOduration=2.000880988 podStartE2EDuration="2.525525605s" podCreationTimestamp="2026-03-18 20:35:23 +0000 UTC" firstStartedPulling="2026-03-18 20:35:24.436355591 +0000 UTC m=+1737.677197459" lastFinishedPulling="2026-03-18 20:35:24.961000208 +0000 UTC m=+1738.201842076" observedRunningTime="2026-03-18 20:35:25.521798177 +0000 UTC m=+1738.762640045" watchObservedRunningTime="2026-03-18 20:35:25.525525605 +0000 UTC m=+1738.766367473" Mar 18 20:35:33 crc kubenswrapper[4950]: I0318 20:35:33.836760 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:35:33 crc kubenswrapper[4950]: I0318 20:35:33.837383 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:35:36 crc kubenswrapper[4950]: I0318 20:35:36.041083 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-z2ffj"] Mar 18 20:35:36 crc kubenswrapper[4950]: I0318 20:35:36.057665 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-z2ffj"] Mar 18 20:35:37 crc kubenswrapper[4950]: I0318 20:35:37.489488 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24db2c06-d91e-4fc1-bcb5-9f29546368e2" path="/var/lib/kubelet/pods/24db2c06-d91e-4fc1-bcb5-9f29546368e2/volumes" Mar 18 20:35:40 crc kubenswrapper[4950]: I0318 20:35:40.038165 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-9524-account-create-update-fznw7"] Mar 18 20:35:40 crc kubenswrapper[4950]: I0318 20:35:40.056001 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-vnl9g"] Mar 18 20:35:40 crc kubenswrapper[4950]: I0318 20:35:40.063858 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-9524-account-create-update-fznw7"] Mar 18 20:35:40 crc kubenswrapper[4950]: I0318 20:35:40.075629 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-32f0-account-create-update-7fwpt"] Mar 18 20:35:40 crc kubenswrapper[4950]: I0318 20:35:40.086011 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-t4zg5"] Mar 18 20:35:40 crc kubenswrapper[4950]: I0318 20:35:40.096121 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-7fe1-account-create-update-9rdjc"] Mar 18 20:35:40 crc kubenswrapper[4950]: I0318 20:35:40.104593 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-t4zg5"] Mar 18 20:35:40 crc kubenswrapper[4950]: I0318 20:35:40.112991 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-vnl9g"] Mar 18 20:35:40 crc kubenswrapper[4950]: I0318 20:35:40.123349 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-32f0-account-create-update-7fwpt"] Mar 18 20:35:40 crc kubenswrapper[4950]: I0318 20:35:40.132306 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-7fe1-account-create-update-9rdjc"] Mar 18 20:35:41 crc kubenswrapper[4950]: I0318 20:35:41.491222 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0af1dd1d-4630-49a6-9f2d-a056fefecf78" path="/var/lib/kubelet/pods/0af1dd1d-4630-49a6-9f2d-a056fefecf78/volumes" Mar 18 20:35:41 crc kubenswrapper[4950]: I0318 20:35:41.491798 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3abd28b2-fbc4-4b22-8f6f-a34917f8cba3" path="/var/lib/kubelet/pods/3abd28b2-fbc4-4b22-8f6f-a34917f8cba3/volumes" Mar 18 20:35:41 crc kubenswrapper[4950]: I0318 20:35:41.492306 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="875a302d-40dd-4c0f-a729-cd6f3dcb4956" path="/var/lib/kubelet/pods/875a302d-40dd-4c0f-a729-cd6f3dcb4956/volumes" Mar 18 20:35:41 crc kubenswrapper[4950]: I0318 20:35:41.492846 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fffb916-a3fd-400a-afd9-997515bf6850" path="/var/lib/kubelet/pods/9fffb916-a3fd-400a-afd9-997515bf6850/volumes" Mar 18 20:35:41 crc kubenswrapper[4950]: I0318 20:35:41.493805 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5a72833-ca01-4da1-84e9-2b144ac64814" path="/var/lib/kubelet/pods/e5a72833-ca01-4da1-84e9-2b144ac64814/volumes" Mar 18 20:35:45 crc kubenswrapper[4950]: I0318 20:35:45.059220 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-jz448"] Mar 18 20:35:45 crc kubenswrapper[4950]: I0318 20:35:45.075195 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-jz448"] Mar 18 20:35:45 crc kubenswrapper[4950]: I0318 20:35:45.492343 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90aa20e7-e9c9-46c5-ad83-8004caa09547" path="/var/lib/kubelet/pods/90aa20e7-e9c9-46c5-ad83-8004caa09547/volumes" Mar 18 20:35:46 crc kubenswrapper[4950]: I0318 20:35:46.857915 4950 scope.go:117] "RemoveContainer" containerID="780cfeaa8a4e5f4288db988fa84b84f89bfc27c5e2b1268a00628c29be67c890" Mar 18 20:35:46 crc kubenswrapper[4950]: I0318 20:35:46.882712 4950 scope.go:117] "RemoveContainer" containerID="f75e0f84cb3b8ea848df445d53cb2dffeb6e6d3f41d31f462db1e9d5dc29c9e2" Mar 18 20:35:46 crc kubenswrapper[4950]: I0318 20:35:46.925776 4950 scope.go:117] "RemoveContainer" containerID="7b94ccfce68784ddcfbf806e3c9c79f6543d2028916875e5833b74b17696f3a3" Mar 18 20:35:46 crc kubenswrapper[4950]: I0318 20:35:46.961554 4950 scope.go:117] "RemoveContainer" containerID="a9c4f36c3ddf85aa65683808380cd3ce1c2813d7594b3d5caf33ff0deb12b3fe" Mar 18 20:35:46 crc kubenswrapper[4950]: I0318 20:35:46.998048 4950 scope.go:117] "RemoveContainer" containerID="a302c9bdde190529bc585b2f5e9e741d7c3462d6c1e115bfbc5162fcfaf814e9" Mar 18 20:35:47 crc kubenswrapper[4950]: I0318 20:35:47.038214 4950 scope.go:117] "RemoveContainer" containerID="3134bc10c0a641b70ab56d0f32141528cb9d14dc72b09ac5bed072548aee33b4" Mar 18 20:35:47 crc kubenswrapper[4950]: I0318 20:35:47.078870 4950 scope.go:117] "RemoveContainer" containerID="66e79885253efa65a0435682ed0a52ce59b093fa665af1665f584db420ccd084" Mar 18 20:35:47 crc kubenswrapper[4950]: I0318 20:35:47.101193 4950 scope.go:117] "RemoveContainer" containerID="2a537136389097a94c9a681b9048c1051b7469641cbea118bd175be37c57658e" Mar 18 20:35:47 crc kubenswrapper[4950]: I0318 20:35:47.123558 4950 scope.go:117] "RemoveContainer" containerID="735d0240236758f8873531be60822231e19073829893db212a7243f9d5cf82f2" Mar 18 20:35:47 crc kubenswrapper[4950]: I0318 20:35:47.165022 4950 scope.go:117] "RemoveContainer" containerID="2cb382e16134e5b58b9d66cfbefac3335e78588db4a5110f353dc456e45a0327" Mar 18 20:35:47 crc kubenswrapper[4950]: I0318 20:35:47.193952 4950 scope.go:117] "RemoveContainer" containerID="c80f2b4f8af07b87c741978f2d3efdb10ca356b077fe4071a9bb0750ebb7b4ad" Mar 18 20:35:47 crc kubenswrapper[4950]: I0318 20:35:47.224928 4950 scope.go:117] "RemoveContainer" containerID="5cf81e528e7dd2c3faddabd82b95a20fa458fcb198d8aa1978a9e8bbe30ec8cf" Mar 18 20:35:47 crc kubenswrapper[4950]: I0318 20:35:47.252756 4950 scope.go:117] "RemoveContainer" containerID="a520c3de603944698f7246c1400a6eb9e6213248cb08ebdf7d5e4507728ba307" Mar 18 20:35:47 crc kubenswrapper[4950]: I0318 20:35:47.281434 4950 scope.go:117] "RemoveContainer" containerID="5cd1f30806cdb193c9d3a294c6fe419e5156658820e1bb223ecc369c4dfc440e" Mar 18 20:35:47 crc kubenswrapper[4950]: I0318 20:35:47.303742 4950 scope.go:117] "RemoveContainer" containerID="2b08ed8575e0d50da93bca56a3da39611d55b13bb3028a048104c7cb5b26993d" Mar 18 20:36:00 crc kubenswrapper[4950]: I0318 20:36:00.147629 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564436-bzgsz"] Mar 18 20:36:00 crc kubenswrapper[4950]: I0318 20:36:00.149177 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564436-bzgsz" Mar 18 20:36:00 crc kubenswrapper[4950]: I0318 20:36:00.155776 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:36:00 crc kubenswrapper[4950]: I0318 20:36:00.155777 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:36:00 crc kubenswrapper[4950]: I0318 20:36:00.155778 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:36:00 crc kubenswrapper[4950]: I0318 20:36:00.172184 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564436-bzgsz"] Mar 18 20:36:00 crc kubenswrapper[4950]: I0318 20:36:00.334978 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4rx2\" (UniqueName: \"kubernetes.io/projected/720dd688-830b-4b1b-b73a-7f6f4938585f-kube-api-access-n4rx2\") pod \"auto-csr-approver-29564436-bzgsz\" (UID: \"720dd688-830b-4b1b-b73a-7f6f4938585f\") " pod="openshift-infra/auto-csr-approver-29564436-bzgsz" Mar 18 20:36:00 crc kubenswrapper[4950]: I0318 20:36:00.437061 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4rx2\" (UniqueName: \"kubernetes.io/projected/720dd688-830b-4b1b-b73a-7f6f4938585f-kube-api-access-n4rx2\") pod \"auto-csr-approver-29564436-bzgsz\" (UID: \"720dd688-830b-4b1b-b73a-7f6f4938585f\") " pod="openshift-infra/auto-csr-approver-29564436-bzgsz" Mar 18 20:36:00 crc kubenswrapper[4950]: I0318 20:36:00.460352 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4rx2\" (UniqueName: \"kubernetes.io/projected/720dd688-830b-4b1b-b73a-7f6f4938585f-kube-api-access-n4rx2\") pod \"auto-csr-approver-29564436-bzgsz\" (UID: \"720dd688-830b-4b1b-b73a-7f6f4938585f\") " pod="openshift-infra/auto-csr-approver-29564436-bzgsz" Mar 18 20:36:00 crc kubenswrapper[4950]: I0318 20:36:00.469749 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564436-bzgsz" Mar 18 20:36:00 crc kubenswrapper[4950]: I0318 20:36:00.776996 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564436-bzgsz"] Mar 18 20:36:00 crc kubenswrapper[4950]: I0318 20:36:00.898077 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564436-bzgsz" event={"ID":"720dd688-830b-4b1b-b73a-7f6f4938585f","Type":"ContainerStarted","Data":"40214ddcdf384a7e78df2e4571f4cee3a6cf46f0b875b900f2f6940222b36480"} Mar 18 20:36:02 crc kubenswrapper[4950]: I0318 20:36:02.922645 4950 generic.go:334] "Generic (PLEG): container finished" podID="720dd688-830b-4b1b-b73a-7f6f4938585f" containerID="c37cea83b16e5d03396ded63bd4f219f34c138a56c3186ed38c292261c5c63c8" exitCode=0 Mar 18 20:36:02 crc kubenswrapper[4950]: I0318 20:36:02.922759 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564436-bzgsz" event={"ID":"720dd688-830b-4b1b-b73a-7f6f4938585f","Type":"ContainerDied","Data":"c37cea83b16e5d03396ded63bd4f219f34c138a56c3186ed38c292261c5c63c8"} Mar 18 20:36:03 crc kubenswrapper[4950]: I0318 20:36:03.836010 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:36:03 crc kubenswrapper[4950]: I0318 20:36:03.836451 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:36:03 crc kubenswrapper[4950]: I0318 20:36:03.836553 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:36:03 crc kubenswrapper[4950]: I0318 20:36:03.837461 4950 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e"} pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 20:36:03 crc kubenswrapper[4950]: I0318 20:36:03.837554 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" containerID="cri-o://f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" gracePeriod=600 Mar 18 20:36:03 crc kubenswrapper[4950]: E0318 20:36:03.988547 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:36:04 crc kubenswrapper[4950]: I0318 20:36:04.280618 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564436-bzgsz" Mar 18 20:36:04 crc kubenswrapper[4950]: I0318 20:36:04.403958 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4rx2\" (UniqueName: \"kubernetes.io/projected/720dd688-830b-4b1b-b73a-7f6f4938585f-kube-api-access-n4rx2\") pod \"720dd688-830b-4b1b-b73a-7f6f4938585f\" (UID: \"720dd688-830b-4b1b-b73a-7f6f4938585f\") " Mar 18 20:36:04 crc kubenswrapper[4950]: I0318 20:36:04.462773 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/720dd688-830b-4b1b-b73a-7f6f4938585f-kube-api-access-n4rx2" (OuterVolumeSpecName: "kube-api-access-n4rx2") pod "720dd688-830b-4b1b-b73a-7f6f4938585f" (UID: "720dd688-830b-4b1b-b73a-7f6f4938585f"). InnerVolumeSpecName "kube-api-access-n4rx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:36:04 crc kubenswrapper[4950]: I0318 20:36:04.506009 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4rx2\" (UniqueName: \"kubernetes.io/projected/720dd688-830b-4b1b-b73a-7f6f4938585f-kube-api-access-n4rx2\") on node \"crc\" DevicePath \"\"" Mar 18 20:36:04 crc kubenswrapper[4950]: I0318 20:36:04.941367 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564436-bzgsz" event={"ID":"720dd688-830b-4b1b-b73a-7f6f4938585f","Type":"ContainerDied","Data":"40214ddcdf384a7e78df2e4571f4cee3a6cf46f0b875b900f2f6940222b36480"} Mar 18 20:36:04 crc kubenswrapper[4950]: I0318 20:36:04.941629 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40214ddcdf384a7e78df2e4571f4cee3a6cf46f0b875b900f2f6940222b36480" Mar 18 20:36:04 crc kubenswrapper[4950]: I0318 20:36:04.941385 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564436-bzgsz" Mar 18 20:36:04 crc kubenswrapper[4950]: I0318 20:36:04.944206 4950 generic.go:334] "Generic (PLEG): container finished" podID="4048b439-3266-46e7-9de0-22377efacc46" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" exitCode=0 Mar 18 20:36:04 crc kubenswrapper[4950]: I0318 20:36:04.944245 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerDied","Data":"f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e"} Mar 18 20:36:04 crc kubenswrapper[4950]: I0318 20:36:04.944278 4950 scope.go:117] "RemoveContainer" containerID="02a13f19effbd6df78e6bd43d75fe3a70d3a67fc27d7dba640df7ed8a4afc83f" Mar 18 20:36:04 crc kubenswrapper[4950]: I0318 20:36:04.946369 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:36:04 crc kubenswrapper[4950]: E0318 20:36:04.947307 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:36:05 crc kubenswrapper[4950]: I0318 20:36:05.347445 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564430-gxf7d"] Mar 18 20:36:05 crc kubenswrapper[4950]: I0318 20:36:05.359288 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564430-gxf7d"] Mar 18 20:36:05 crc kubenswrapper[4950]: I0318 20:36:05.492334 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d021afc1-77fb-4064-8b4c-a81f25c293eb" path="/var/lib/kubelet/pods/d021afc1-77fb-4064-8b4c-a81f25c293eb/volumes" Mar 18 20:36:06 crc kubenswrapper[4950]: I0318 20:36:06.970027 4950 generic.go:334] "Generic (PLEG): container finished" podID="84bce393-6964-42d4-9126-076287a2dece" containerID="68f707f726c227fe94bff58052b50aead3ac2b367f00a474c05bb6f61be620d7" exitCode=0 Mar 18 20:36:06 crc kubenswrapper[4950]: I0318 20:36:06.970268 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xhfbh" event={"ID":"84bce393-6964-42d4-9126-076287a2dece","Type":"ContainerDied","Data":"68f707f726c227fe94bff58052b50aead3ac2b367f00a474c05bb6f61be620d7"} Mar 18 20:36:08 crc kubenswrapper[4950]: I0318 20:36:08.400401 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xhfbh" Mar 18 20:36:08 crc kubenswrapper[4950]: I0318 20:36:08.486187 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kl2n\" (UniqueName: \"kubernetes.io/projected/84bce393-6964-42d4-9126-076287a2dece-kube-api-access-5kl2n\") pod \"84bce393-6964-42d4-9126-076287a2dece\" (UID: \"84bce393-6964-42d4-9126-076287a2dece\") " Mar 18 20:36:08 crc kubenswrapper[4950]: I0318 20:36:08.486382 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84bce393-6964-42d4-9126-076287a2dece-inventory\") pod \"84bce393-6964-42d4-9126-076287a2dece\" (UID: \"84bce393-6964-42d4-9126-076287a2dece\") " Mar 18 20:36:08 crc kubenswrapper[4950]: I0318 20:36:08.487199 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/84bce393-6964-42d4-9126-076287a2dece-ssh-key-openstack-edpm-ipam\") pod \"84bce393-6964-42d4-9126-076287a2dece\" (UID: \"84bce393-6964-42d4-9126-076287a2dece\") " Mar 18 20:36:08 crc kubenswrapper[4950]: I0318 20:36:08.492168 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84bce393-6964-42d4-9126-076287a2dece-kube-api-access-5kl2n" (OuterVolumeSpecName: "kube-api-access-5kl2n") pod "84bce393-6964-42d4-9126-076287a2dece" (UID: "84bce393-6964-42d4-9126-076287a2dece"). InnerVolumeSpecName "kube-api-access-5kl2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:36:08 crc kubenswrapper[4950]: I0318 20:36:08.515672 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84bce393-6964-42d4-9126-076287a2dece-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "84bce393-6964-42d4-9126-076287a2dece" (UID: "84bce393-6964-42d4-9126-076287a2dece"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:36:08 crc kubenswrapper[4950]: I0318 20:36:08.517161 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84bce393-6964-42d4-9126-076287a2dece-inventory" (OuterVolumeSpecName: "inventory") pod "84bce393-6964-42d4-9126-076287a2dece" (UID: "84bce393-6964-42d4-9126-076287a2dece"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:36:08 crc kubenswrapper[4950]: I0318 20:36:08.591071 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/84bce393-6964-42d4-9126-076287a2dece-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:36:08 crc kubenswrapper[4950]: I0318 20:36:08.591125 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kl2n\" (UniqueName: \"kubernetes.io/projected/84bce393-6964-42d4-9126-076287a2dece-kube-api-access-5kl2n\") on node \"crc\" DevicePath \"\"" Mar 18 20:36:08 crc kubenswrapper[4950]: I0318 20:36:08.591146 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84bce393-6964-42d4-9126-076287a2dece-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:36:08 crc kubenswrapper[4950]: I0318 20:36:08.991507 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xhfbh" event={"ID":"84bce393-6964-42d4-9126-076287a2dece","Type":"ContainerDied","Data":"f4ea30e682c3a55edd6d1bbffca651cde38aa545198728bd00dd96db5dcda913"} Mar 18 20:36:08 crc kubenswrapper[4950]: I0318 20:36:08.991905 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4ea30e682c3a55edd6d1bbffca651cde38aa545198728bd00dd96db5dcda913" Mar 18 20:36:08 crc kubenswrapper[4950]: I0318 20:36:08.991566 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xhfbh" Mar 18 20:36:09 crc kubenswrapper[4950]: I0318 20:36:09.109557 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6"] Mar 18 20:36:09 crc kubenswrapper[4950]: E0318 20:36:09.110056 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84bce393-6964-42d4-9126-076287a2dece" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:36:09 crc kubenswrapper[4950]: I0318 20:36:09.110079 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="84bce393-6964-42d4-9126-076287a2dece" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:36:09 crc kubenswrapper[4950]: E0318 20:36:09.110099 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="720dd688-830b-4b1b-b73a-7f6f4938585f" containerName="oc" Mar 18 20:36:09 crc kubenswrapper[4950]: I0318 20:36:09.110107 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="720dd688-830b-4b1b-b73a-7f6f4938585f" containerName="oc" Mar 18 20:36:09 crc kubenswrapper[4950]: I0318 20:36:09.110301 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="720dd688-830b-4b1b-b73a-7f6f4938585f" containerName="oc" Mar 18 20:36:09 crc kubenswrapper[4950]: I0318 20:36:09.110326 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="84bce393-6964-42d4-9126-076287a2dece" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:36:09 crc kubenswrapper[4950]: I0318 20:36:09.111001 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6" Mar 18 20:36:09 crc kubenswrapper[4950]: I0318 20:36:09.115513 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:36:09 crc kubenswrapper[4950]: I0318 20:36:09.115616 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:36:09 crc kubenswrapper[4950]: I0318 20:36:09.115775 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:36:09 crc kubenswrapper[4950]: I0318 20:36:09.117497 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:36:09 crc kubenswrapper[4950]: I0318 20:36:09.169156 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6"] Mar 18 20:36:09 crc kubenswrapper[4950]: I0318 20:36:09.200664 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz9bw\" (UniqueName: \"kubernetes.io/projected/5162a83c-d820-45d0-86c7-b4cb8a22a607-kube-api-access-gz9bw\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6\" (UID: \"5162a83c-d820-45d0-86c7-b4cb8a22a607\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6" Mar 18 20:36:09 crc kubenswrapper[4950]: I0318 20:36:09.200722 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5162a83c-d820-45d0-86c7-b4cb8a22a607-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6\" (UID: \"5162a83c-d820-45d0-86c7-b4cb8a22a607\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6" Mar 18 20:36:09 crc kubenswrapper[4950]: I0318 20:36:09.200776 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5162a83c-d820-45d0-86c7-b4cb8a22a607-ssh-key-openstack-edpm-ipam\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6\" (UID: \"5162a83c-d820-45d0-86c7-b4cb8a22a607\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6" Mar 18 20:36:09 crc kubenswrapper[4950]: I0318 20:36:09.302361 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5162a83c-d820-45d0-86c7-b4cb8a22a607-ssh-key-openstack-edpm-ipam\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6\" (UID: \"5162a83c-d820-45d0-86c7-b4cb8a22a607\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6" Mar 18 20:36:09 crc kubenswrapper[4950]: I0318 20:36:09.302543 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz9bw\" (UniqueName: \"kubernetes.io/projected/5162a83c-d820-45d0-86c7-b4cb8a22a607-kube-api-access-gz9bw\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6\" (UID: \"5162a83c-d820-45d0-86c7-b4cb8a22a607\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6" Mar 18 20:36:09 crc kubenswrapper[4950]: I0318 20:36:09.302581 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5162a83c-d820-45d0-86c7-b4cb8a22a607-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6\" (UID: \"5162a83c-d820-45d0-86c7-b4cb8a22a607\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6" Mar 18 20:36:09 crc kubenswrapper[4950]: I0318 20:36:09.313614 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5162a83c-d820-45d0-86c7-b4cb8a22a607-ssh-key-openstack-edpm-ipam\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6\" (UID: \"5162a83c-d820-45d0-86c7-b4cb8a22a607\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6" Mar 18 20:36:09 crc kubenswrapper[4950]: I0318 20:36:09.318911 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5162a83c-d820-45d0-86c7-b4cb8a22a607-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6\" (UID: \"5162a83c-d820-45d0-86c7-b4cb8a22a607\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6" Mar 18 20:36:09 crc kubenswrapper[4950]: I0318 20:36:09.320755 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz9bw\" (UniqueName: \"kubernetes.io/projected/5162a83c-d820-45d0-86c7-b4cb8a22a607-kube-api-access-gz9bw\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6\" (UID: \"5162a83c-d820-45d0-86c7-b4cb8a22a607\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6" Mar 18 20:36:09 crc kubenswrapper[4950]: I0318 20:36:09.466851 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6" Mar 18 20:36:10 crc kubenswrapper[4950]: W0318 20:36:10.045122 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5162a83c_d820_45d0_86c7_b4cb8a22a607.slice/crio-c8f9209c745af2db1d4cc0ca412a4b1e6e030bf915d75da9c1f9718d76fce8bf WatchSource:0}: Error finding container c8f9209c745af2db1d4cc0ca412a4b1e6e030bf915d75da9c1f9718d76fce8bf: Status 404 returned error can't find the container with id c8f9209c745af2db1d4cc0ca412a4b1e6e030bf915d75da9c1f9718d76fce8bf Mar 18 20:36:10 crc kubenswrapper[4950]: I0318 20:36:10.048819 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6"] Mar 18 20:36:11 crc kubenswrapper[4950]: I0318 20:36:11.009868 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6" event={"ID":"5162a83c-d820-45d0-86c7-b4cb8a22a607","Type":"ContainerStarted","Data":"57a9ca3c7f01904503552eba81817b6563a3fb9ee3278c21bf505930c5ba0aef"} Mar 18 20:36:11 crc kubenswrapper[4950]: I0318 20:36:11.010433 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6" event={"ID":"5162a83c-d820-45d0-86c7-b4cb8a22a607","Type":"ContainerStarted","Data":"c8f9209c745af2db1d4cc0ca412a4b1e6e030bf915d75da9c1f9718d76fce8bf"} Mar 18 20:36:11 crc kubenswrapper[4950]: I0318 20:36:11.035379 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6" podStartSLOduration=1.546473781 podStartE2EDuration="2.035358167s" podCreationTimestamp="2026-03-18 20:36:09 +0000 UTC" firstStartedPulling="2026-03-18 20:36:10.048828506 +0000 UTC m=+1783.289670374" lastFinishedPulling="2026-03-18 20:36:10.537712892 +0000 UTC m=+1783.778554760" observedRunningTime="2026-03-18 20:36:11.033747225 +0000 UTC m=+1784.274589093" watchObservedRunningTime="2026-03-18 20:36:11.035358167 +0000 UTC m=+1784.276200035" Mar 18 20:36:12 crc kubenswrapper[4950]: I0318 20:36:12.041190 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-j5m65"] Mar 18 20:36:12 crc kubenswrapper[4950]: I0318 20:36:12.050899 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-j5m65"] Mar 18 20:36:13 crc kubenswrapper[4950]: I0318 20:36:13.500645 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbf245a4-e400-4fc2-9d2e-3788c6baf25d" path="/var/lib/kubelet/pods/dbf245a4-e400-4fc2-9d2e-3788c6baf25d/volumes" Mar 18 20:36:15 crc kubenswrapper[4950]: I0318 20:36:15.043960 4950 generic.go:334] "Generic (PLEG): container finished" podID="5162a83c-d820-45d0-86c7-b4cb8a22a607" containerID="57a9ca3c7f01904503552eba81817b6563a3fb9ee3278c21bf505930c5ba0aef" exitCode=0 Mar 18 20:36:15 crc kubenswrapper[4950]: I0318 20:36:15.044273 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6" event={"ID":"5162a83c-d820-45d0-86c7-b4cb8a22a607","Type":"ContainerDied","Data":"57a9ca3c7f01904503552eba81817b6563a3fb9ee3278c21bf505930c5ba0aef"} Mar 18 20:36:16 crc kubenswrapper[4950]: I0318 20:36:16.043027 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-42zwb"] Mar 18 20:36:16 crc kubenswrapper[4950]: I0318 20:36:16.054083 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-42zwb"] Mar 18 20:36:16 crc kubenswrapper[4950]: I0318 20:36:16.474645 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6" Mar 18 20:36:16 crc kubenswrapper[4950]: I0318 20:36:16.479443 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:36:16 crc kubenswrapper[4950]: E0318 20:36:16.479683 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:36:16 crc kubenswrapper[4950]: I0318 20:36:16.635976 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gz9bw\" (UniqueName: \"kubernetes.io/projected/5162a83c-d820-45d0-86c7-b4cb8a22a607-kube-api-access-gz9bw\") pod \"5162a83c-d820-45d0-86c7-b4cb8a22a607\" (UID: \"5162a83c-d820-45d0-86c7-b4cb8a22a607\") " Mar 18 20:36:16 crc kubenswrapper[4950]: I0318 20:36:16.636290 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5162a83c-d820-45d0-86c7-b4cb8a22a607-ssh-key-openstack-edpm-ipam\") pod \"5162a83c-d820-45d0-86c7-b4cb8a22a607\" (UID: \"5162a83c-d820-45d0-86c7-b4cb8a22a607\") " Mar 18 20:36:16 crc kubenswrapper[4950]: I0318 20:36:16.636664 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5162a83c-d820-45d0-86c7-b4cb8a22a607-inventory\") pod \"5162a83c-d820-45d0-86c7-b4cb8a22a607\" (UID: \"5162a83c-d820-45d0-86c7-b4cb8a22a607\") " Mar 18 20:36:16 crc kubenswrapper[4950]: I0318 20:36:16.642123 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5162a83c-d820-45d0-86c7-b4cb8a22a607-kube-api-access-gz9bw" (OuterVolumeSpecName: "kube-api-access-gz9bw") pod "5162a83c-d820-45d0-86c7-b4cb8a22a607" (UID: "5162a83c-d820-45d0-86c7-b4cb8a22a607"). InnerVolumeSpecName "kube-api-access-gz9bw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:36:16 crc kubenswrapper[4950]: I0318 20:36:16.661822 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5162a83c-d820-45d0-86c7-b4cb8a22a607-inventory" (OuterVolumeSpecName: "inventory") pod "5162a83c-d820-45d0-86c7-b4cb8a22a607" (UID: "5162a83c-d820-45d0-86c7-b4cb8a22a607"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:36:16 crc kubenswrapper[4950]: I0318 20:36:16.673293 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5162a83c-d820-45d0-86c7-b4cb8a22a607-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5162a83c-d820-45d0-86c7-b4cb8a22a607" (UID: "5162a83c-d820-45d0-86c7-b4cb8a22a607"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:36:16 crc kubenswrapper[4950]: I0318 20:36:16.739499 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5162a83c-d820-45d0-86c7-b4cb8a22a607-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:36:16 crc kubenswrapper[4950]: I0318 20:36:16.739536 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5162a83c-d820-45d0-86c7-b4cb8a22a607-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:36:16 crc kubenswrapper[4950]: I0318 20:36:16.739550 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gz9bw\" (UniqueName: \"kubernetes.io/projected/5162a83c-d820-45d0-86c7-b4cb8a22a607-kube-api-access-gz9bw\") on node \"crc\" DevicePath \"\"" Mar 18 20:36:17 crc kubenswrapper[4950]: I0318 20:36:17.061392 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6" event={"ID":"5162a83c-d820-45d0-86c7-b4cb8a22a607","Type":"ContainerDied","Data":"c8f9209c745af2db1d4cc0ca412a4b1e6e030bf915d75da9c1f9718d76fce8bf"} Mar 18 20:36:17 crc kubenswrapper[4950]: I0318 20:36:17.061474 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6" Mar 18 20:36:17 crc kubenswrapper[4950]: I0318 20:36:17.061480 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8f9209c745af2db1d4cc0ca412a4b1e6e030bf915d75da9c1f9718d76fce8bf" Mar 18 20:36:17 crc kubenswrapper[4950]: I0318 20:36:17.152941 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r"] Mar 18 20:36:17 crc kubenswrapper[4950]: E0318 20:36:17.153348 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5162a83c-d820-45d0-86c7-b4cb8a22a607" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Mar 18 20:36:17 crc kubenswrapper[4950]: I0318 20:36:17.153370 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="5162a83c-d820-45d0-86c7-b4cb8a22a607" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Mar 18 20:36:17 crc kubenswrapper[4950]: I0318 20:36:17.153606 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="5162a83c-d820-45d0-86c7-b4cb8a22a607" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Mar 18 20:36:17 crc kubenswrapper[4950]: I0318 20:36:17.154301 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r" Mar 18 20:36:17 crc kubenswrapper[4950]: I0318 20:36:17.165816 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:36:17 crc kubenswrapper[4950]: I0318 20:36:17.166460 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:36:17 crc kubenswrapper[4950]: I0318 20:36:17.172217 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:36:17 crc kubenswrapper[4950]: I0318 20:36:17.173812 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r"] Mar 18 20:36:17 crc kubenswrapper[4950]: I0318 20:36:17.176200 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:36:17 crc kubenswrapper[4950]: I0318 20:36:17.348982 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5495b98f-7c92-4918-b4f7-d1e7e06dd465-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r\" (UID: \"5495b98f-7c92-4918-b4f7-d1e7e06dd465\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r" Mar 18 20:36:17 crc kubenswrapper[4950]: I0318 20:36:17.349048 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc452\" (UniqueName: \"kubernetes.io/projected/5495b98f-7c92-4918-b4f7-d1e7e06dd465-kube-api-access-qc452\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r\" (UID: \"5495b98f-7c92-4918-b4f7-d1e7e06dd465\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r" Mar 18 20:36:17 crc kubenswrapper[4950]: I0318 20:36:17.349792 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5495b98f-7c92-4918-b4f7-d1e7e06dd465-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r\" (UID: \"5495b98f-7c92-4918-b4f7-d1e7e06dd465\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r" Mar 18 20:36:17 crc kubenswrapper[4950]: I0318 20:36:17.452135 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5495b98f-7c92-4918-b4f7-d1e7e06dd465-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r\" (UID: \"5495b98f-7c92-4918-b4f7-d1e7e06dd465\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r" Mar 18 20:36:17 crc kubenswrapper[4950]: I0318 20:36:17.452201 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc452\" (UniqueName: \"kubernetes.io/projected/5495b98f-7c92-4918-b4f7-d1e7e06dd465-kube-api-access-qc452\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r\" (UID: \"5495b98f-7c92-4918-b4f7-d1e7e06dd465\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r" Mar 18 20:36:17 crc kubenswrapper[4950]: I0318 20:36:17.452274 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5495b98f-7c92-4918-b4f7-d1e7e06dd465-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r\" (UID: \"5495b98f-7c92-4918-b4f7-d1e7e06dd465\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r" Mar 18 20:36:17 crc kubenswrapper[4950]: I0318 20:36:17.459016 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5495b98f-7c92-4918-b4f7-d1e7e06dd465-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r\" (UID: \"5495b98f-7c92-4918-b4f7-d1e7e06dd465\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r" Mar 18 20:36:17 crc kubenswrapper[4950]: I0318 20:36:17.459190 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5495b98f-7c92-4918-b4f7-d1e7e06dd465-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r\" (UID: \"5495b98f-7c92-4918-b4f7-d1e7e06dd465\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r" Mar 18 20:36:17 crc kubenswrapper[4950]: I0318 20:36:17.479353 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc452\" (UniqueName: \"kubernetes.io/projected/5495b98f-7c92-4918-b4f7-d1e7e06dd465-kube-api-access-qc452\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r\" (UID: \"5495b98f-7c92-4918-b4f7-d1e7e06dd465\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r" Mar 18 20:36:17 crc kubenswrapper[4950]: I0318 20:36:17.498282 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5989cc6f-e4c7-4511-a148-293f683ede13" path="/var/lib/kubelet/pods/5989cc6f-e4c7-4511-a148-293f683ede13/volumes" Mar 18 20:36:17 crc kubenswrapper[4950]: I0318 20:36:17.770231 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r" Mar 18 20:36:18 crc kubenswrapper[4950]: I0318 20:36:18.393070 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r"] Mar 18 20:36:19 crc kubenswrapper[4950]: I0318 20:36:19.102765 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r" event={"ID":"5495b98f-7c92-4918-b4f7-d1e7e06dd465","Type":"ContainerStarted","Data":"13a50ad267968a229900496c928471e816a9331ac8843faecfe054509b9197c2"} Mar 18 20:36:20 crc kubenswrapper[4950]: I0318 20:36:20.112048 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r" event={"ID":"5495b98f-7c92-4918-b4f7-d1e7e06dd465","Type":"ContainerStarted","Data":"a4eacba1d449d76b272bfeca9562771e6505952fbfd2fa9a0ab115c1c228b6f8"} Mar 18 20:36:20 crc kubenswrapper[4950]: I0318 20:36:20.136370 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r" podStartSLOduration=2.621778461 podStartE2EDuration="3.136344573s" podCreationTimestamp="2026-03-18 20:36:17 +0000 UTC" firstStartedPulling="2026-03-18 20:36:18.408121692 +0000 UTC m=+1791.648963580" lastFinishedPulling="2026-03-18 20:36:18.922687784 +0000 UTC m=+1792.163529692" observedRunningTime="2026-03-18 20:36:20.135250804 +0000 UTC m=+1793.376092682" watchObservedRunningTime="2026-03-18 20:36:20.136344573 +0000 UTC m=+1793.377186461" Mar 18 20:36:22 crc kubenswrapper[4950]: I0318 20:36:22.040090 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-dfbp4"] Mar 18 20:36:22 crc kubenswrapper[4950]: I0318 20:36:22.051192 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-dfbp4"] Mar 18 20:36:23 crc kubenswrapper[4950]: I0318 20:36:23.490012 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8589447-6f54-4a2b-80b5-29fdbccb4971" path="/var/lib/kubelet/pods/d8589447-6f54-4a2b-80b5-29fdbccb4971/volumes" Mar 18 20:36:29 crc kubenswrapper[4950]: I0318 20:36:29.482055 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:36:29 crc kubenswrapper[4950]: E0318 20:36:29.482607 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:36:34 crc kubenswrapper[4950]: I0318 20:36:34.057031 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-brgnq"] Mar 18 20:36:34 crc kubenswrapper[4950]: I0318 20:36:34.065026 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-brgnq"] Mar 18 20:36:35 crc kubenswrapper[4950]: I0318 20:36:35.493317 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82a7c2a4-9f57-460e-8bac-419f14222502" path="/var/lib/kubelet/pods/82a7c2a4-9f57-460e-8bac-419f14222502/volumes" Mar 18 20:36:40 crc kubenswrapper[4950]: I0318 20:36:40.037698 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-tpdmj"] Mar 18 20:36:40 crc kubenswrapper[4950]: I0318 20:36:40.050263 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-tpdmj"] Mar 18 20:36:41 crc kubenswrapper[4950]: I0318 20:36:41.500570 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4bb9e1a-0694-4e00-8ba6-f2765f82b983" path="/var/lib/kubelet/pods/a4bb9e1a-0694-4e00-8ba6-f2765f82b983/volumes" Mar 18 20:36:44 crc kubenswrapper[4950]: I0318 20:36:44.481228 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:36:44 crc kubenswrapper[4950]: E0318 20:36:44.481871 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:36:47 crc kubenswrapper[4950]: I0318 20:36:47.656190 4950 scope.go:117] "RemoveContainer" containerID="08cf42bf1bb71302f9083c4aa0908a4cd8aed1c546a9aa614f9927b0b0ad5bf8" Mar 18 20:36:47 crc kubenswrapper[4950]: I0318 20:36:47.785691 4950 scope.go:117] "RemoveContainer" containerID="161b8640c2219b243d32be7dcefce1bd3ee90d650d799042fa629a4ca59da713" Mar 18 20:36:47 crc kubenswrapper[4950]: I0318 20:36:47.852052 4950 scope.go:117] "RemoveContainer" containerID="076697f27bc810727a97bc72cc9216b889409ae3d998e9af07d0b9f87a899622" Mar 18 20:36:47 crc kubenswrapper[4950]: I0318 20:36:47.879260 4950 scope.go:117] "RemoveContainer" containerID="de9ab05a4141d7f72e470b7a98a89d526812c15d9a4174be30b8a51cc0d0d4b3" Mar 18 20:36:47 crc kubenswrapper[4950]: I0318 20:36:47.921721 4950 scope.go:117] "RemoveContainer" containerID="ef83c6302a1ab85f29e24e83bef380a024d32fad1c7855a688e1adbe340d26bf" Mar 18 20:36:47 crc kubenswrapper[4950]: I0318 20:36:47.965397 4950 scope.go:117] "RemoveContainer" containerID="67ac8553eb5aa5ab32a8c742ca36189fb2548b29cc3cfaab3867d8931285085c" Mar 18 20:36:55 crc kubenswrapper[4950]: I0318 20:36:55.479835 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:36:55 crc kubenswrapper[4950]: E0318 20:36:55.480742 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:37:07 crc kubenswrapper[4950]: I0318 20:37:07.486848 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:37:07 crc kubenswrapper[4950]: E0318 20:37:07.487694 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:37:12 crc kubenswrapper[4950]: I0318 20:37:12.570607 4950 generic.go:334] "Generic (PLEG): container finished" podID="5495b98f-7c92-4918-b4f7-d1e7e06dd465" containerID="a4eacba1d449d76b272bfeca9562771e6505952fbfd2fa9a0ab115c1c228b6f8" exitCode=0 Mar 18 20:37:12 crc kubenswrapper[4950]: I0318 20:37:12.570665 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r" event={"ID":"5495b98f-7c92-4918-b4f7-d1e7e06dd465","Type":"ContainerDied","Data":"a4eacba1d449d76b272bfeca9562771e6505952fbfd2fa9a0ab115c1c228b6f8"} Mar 18 20:37:13 crc kubenswrapper[4950]: I0318 20:37:13.992160 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r" Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.084491 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5495b98f-7c92-4918-b4f7-d1e7e06dd465-inventory\") pod \"5495b98f-7c92-4918-b4f7-d1e7e06dd465\" (UID: \"5495b98f-7c92-4918-b4f7-d1e7e06dd465\") " Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.084552 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qc452\" (UniqueName: \"kubernetes.io/projected/5495b98f-7c92-4918-b4f7-d1e7e06dd465-kube-api-access-qc452\") pod \"5495b98f-7c92-4918-b4f7-d1e7e06dd465\" (UID: \"5495b98f-7c92-4918-b4f7-d1e7e06dd465\") " Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.094988 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5495b98f-7c92-4918-b4f7-d1e7e06dd465-kube-api-access-qc452" (OuterVolumeSpecName: "kube-api-access-qc452") pod "5495b98f-7c92-4918-b4f7-d1e7e06dd465" (UID: "5495b98f-7c92-4918-b4f7-d1e7e06dd465"). InnerVolumeSpecName "kube-api-access-qc452". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.115866 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5495b98f-7c92-4918-b4f7-d1e7e06dd465-inventory" (OuterVolumeSpecName: "inventory") pod "5495b98f-7c92-4918-b4f7-d1e7e06dd465" (UID: "5495b98f-7c92-4918-b4f7-d1e7e06dd465"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.185629 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5495b98f-7c92-4918-b4f7-d1e7e06dd465-ssh-key-openstack-edpm-ipam\") pod \"5495b98f-7c92-4918-b4f7-d1e7e06dd465\" (UID: \"5495b98f-7c92-4918-b4f7-d1e7e06dd465\") " Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.186311 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5495b98f-7c92-4918-b4f7-d1e7e06dd465-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.186328 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qc452\" (UniqueName: \"kubernetes.io/projected/5495b98f-7c92-4918-b4f7-d1e7e06dd465-kube-api-access-qc452\") on node \"crc\" DevicePath \"\"" Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.211539 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5495b98f-7c92-4918-b4f7-d1e7e06dd465-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5495b98f-7c92-4918-b4f7-d1e7e06dd465" (UID: "5495b98f-7c92-4918-b4f7-d1e7e06dd465"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.288563 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5495b98f-7c92-4918-b4f7-d1e7e06dd465-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.592880 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r" event={"ID":"5495b98f-7c92-4918-b4f7-d1e7e06dd465","Type":"ContainerDied","Data":"13a50ad267968a229900496c928471e816a9331ac8843faecfe054509b9197c2"} Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.592927 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13a50ad267968a229900496c928471e816a9331ac8843faecfe054509b9197c2" Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.593362 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r" Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.697743 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2vct7"] Mar 18 20:37:14 crc kubenswrapper[4950]: E0318 20:37:14.698165 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5495b98f-7c92-4918-b4f7-d1e7e06dd465" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.698186 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="5495b98f-7c92-4918-b4f7-d1e7e06dd465" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.698431 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="5495b98f-7c92-4918-b4f7-d1e7e06dd465" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.699127 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2vct7" Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.705466 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.705679 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.705942 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.706139 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.706530 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2vct7"] Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.901688 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/651a2596-f562-4012-8f2a-7a3e1c562a0a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-2vct7\" (UID: \"651a2596-f562-4012-8f2a-7a3e1c562a0a\") " pod="openstack/ssh-known-hosts-edpm-deployment-2vct7" Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.901735 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kj4z2\" (UniqueName: \"kubernetes.io/projected/651a2596-f562-4012-8f2a-7a3e1c562a0a-kube-api-access-kj4z2\") pod \"ssh-known-hosts-edpm-deployment-2vct7\" (UID: \"651a2596-f562-4012-8f2a-7a3e1c562a0a\") " pod="openstack/ssh-known-hosts-edpm-deployment-2vct7" Mar 18 20:37:14 crc kubenswrapper[4950]: I0318 20:37:14.901984 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/651a2596-f562-4012-8f2a-7a3e1c562a0a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-2vct7\" (UID: \"651a2596-f562-4012-8f2a-7a3e1c562a0a\") " pod="openstack/ssh-known-hosts-edpm-deployment-2vct7" Mar 18 20:37:15 crc kubenswrapper[4950]: I0318 20:37:15.004105 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/651a2596-f562-4012-8f2a-7a3e1c562a0a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-2vct7\" (UID: \"651a2596-f562-4012-8f2a-7a3e1c562a0a\") " pod="openstack/ssh-known-hosts-edpm-deployment-2vct7" Mar 18 20:37:15 crc kubenswrapper[4950]: I0318 20:37:15.004191 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/651a2596-f562-4012-8f2a-7a3e1c562a0a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-2vct7\" (UID: \"651a2596-f562-4012-8f2a-7a3e1c562a0a\") " pod="openstack/ssh-known-hosts-edpm-deployment-2vct7" Mar 18 20:37:15 crc kubenswrapper[4950]: I0318 20:37:15.004234 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kj4z2\" (UniqueName: \"kubernetes.io/projected/651a2596-f562-4012-8f2a-7a3e1c562a0a-kube-api-access-kj4z2\") pod \"ssh-known-hosts-edpm-deployment-2vct7\" (UID: \"651a2596-f562-4012-8f2a-7a3e1c562a0a\") " pod="openstack/ssh-known-hosts-edpm-deployment-2vct7" Mar 18 20:37:15 crc kubenswrapper[4950]: I0318 20:37:15.009973 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/651a2596-f562-4012-8f2a-7a3e1c562a0a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-2vct7\" (UID: \"651a2596-f562-4012-8f2a-7a3e1c562a0a\") " pod="openstack/ssh-known-hosts-edpm-deployment-2vct7" Mar 18 20:37:15 crc kubenswrapper[4950]: I0318 20:37:15.011815 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/651a2596-f562-4012-8f2a-7a3e1c562a0a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-2vct7\" (UID: \"651a2596-f562-4012-8f2a-7a3e1c562a0a\") " pod="openstack/ssh-known-hosts-edpm-deployment-2vct7" Mar 18 20:37:15 crc kubenswrapper[4950]: I0318 20:37:15.047796 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kj4z2\" (UniqueName: \"kubernetes.io/projected/651a2596-f562-4012-8f2a-7a3e1c562a0a-kube-api-access-kj4z2\") pod \"ssh-known-hosts-edpm-deployment-2vct7\" (UID: \"651a2596-f562-4012-8f2a-7a3e1c562a0a\") " pod="openstack/ssh-known-hosts-edpm-deployment-2vct7" Mar 18 20:37:15 crc kubenswrapper[4950]: I0318 20:37:15.052097 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2vct7" Mar 18 20:37:15 crc kubenswrapper[4950]: I0318 20:37:15.607155 4950 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 20:37:15 crc kubenswrapper[4950]: I0318 20:37:15.607157 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2vct7"] Mar 18 20:37:16 crc kubenswrapper[4950]: I0318 20:37:16.609835 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2vct7" event={"ID":"651a2596-f562-4012-8f2a-7a3e1c562a0a","Type":"ContainerStarted","Data":"d63d38a73140c397d468f5213d693b05c42ef6f917d197623927e165b2ab0e18"} Mar 18 20:37:16 crc kubenswrapper[4950]: I0318 20:37:16.610291 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2vct7" event={"ID":"651a2596-f562-4012-8f2a-7a3e1c562a0a","Type":"ContainerStarted","Data":"929a4e9d789bf3d9cdbd802065e1e1023c55a8249ce6caac22118bef9291e783"} Mar 18 20:37:16 crc kubenswrapper[4950]: I0318 20:37:16.629989 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-2vct7" podStartSLOduration=2.133586663 podStartE2EDuration="2.629965306s" podCreationTimestamp="2026-03-18 20:37:14 +0000 UTC" firstStartedPulling="2026-03-18 20:37:15.606970564 +0000 UTC m=+1848.847812422" lastFinishedPulling="2026-03-18 20:37:16.103349197 +0000 UTC m=+1849.344191065" observedRunningTime="2026-03-18 20:37:16.627279215 +0000 UTC m=+1849.868121123" watchObservedRunningTime="2026-03-18 20:37:16.629965306 +0000 UTC m=+1849.870807174" Mar 18 20:37:18 crc kubenswrapper[4950]: I0318 20:37:18.479706 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:37:18 crc kubenswrapper[4950]: E0318 20:37:18.480341 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:37:20 crc kubenswrapper[4950]: I0318 20:37:20.060595 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-hdfzb"] Mar 18 20:37:20 crc kubenswrapper[4950]: I0318 20:37:20.069575 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-1118-account-create-update-6d4gl"] Mar 18 20:37:20 crc kubenswrapper[4950]: I0318 20:37:20.107304 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-dc64-account-create-update-tdrsv"] Mar 18 20:37:20 crc kubenswrapper[4950]: I0318 20:37:20.118076 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-dc64-account-create-update-tdrsv"] Mar 18 20:37:20 crc kubenswrapper[4950]: I0318 20:37:20.128625 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-1118-account-create-update-6d4gl"] Mar 18 20:37:20 crc kubenswrapper[4950]: I0318 20:37:20.139167 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-hdfzb"] Mar 18 20:37:21 crc kubenswrapper[4950]: I0318 20:37:21.036084 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-swn7d"] Mar 18 20:37:21 crc kubenswrapper[4950]: I0318 20:37:21.046633 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-91fe-account-create-update-kfnc8"] Mar 18 20:37:21 crc kubenswrapper[4950]: I0318 20:37:21.054674 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-6t686"] Mar 18 20:37:21 crc kubenswrapper[4950]: I0318 20:37:21.062819 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-swn7d"] Mar 18 20:37:21 crc kubenswrapper[4950]: I0318 20:37:21.069680 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-91fe-account-create-update-kfnc8"] Mar 18 20:37:21 crc kubenswrapper[4950]: I0318 20:37:21.076743 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-6t686"] Mar 18 20:37:21 crc kubenswrapper[4950]: I0318 20:37:21.492332 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15a8a30f-2d5d-4cbc-8fd6-71ec32c73712" path="/var/lib/kubelet/pods/15a8a30f-2d5d-4cbc-8fd6-71ec32c73712/volumes" Mar 18 20:37:21 crc kubenswrapper[4950]: I0318 20:37:21.492892 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c3408fa-8790-4354-b564-1c0978f35665" path="/var/lib/kubelet/pods/1c3408fa-8790-4354-b564-1c0978f35665/volumes" Mar 18 20:37:21 crc kubenswrapper[4950]: I0318 20:37:21.493428 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c107bf1-4431-444a-bb5b-2a44f4864e7e" path="/var/lib/kubelet/pods/3c107bf1-4431-444a-bb5b-2a44f4864e7e/volumes" Mar 18 20:37:21 crc kubenswrapper[4950]: I0318 20:37:21.493942 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90c9ab22-c07c-48df-afec-0d5439f06f96" path="/var/lib/kubelet/pods/90c9ab22-c07c-48df-afec-0d5439f06f96/volumes" Mar 18 20:37:21 crc kubenswrapper[4950]: I0318 20:37:21.494931 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3" path="/var/lib/kubelet/pods/9a7e5209-9c77-40b4-acd8-a1b7e44ba3a3/volumes" Mar 18 20:37:21 crc kubenswrapper[4950]: I0318 20:37:21.495443 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c21582df-e249-45d2-82de-9f0ab2592388" path="/var/lib/kubelet/pods/c21582df-e249-45d2-82de-9f0ab2592388/volumes" Mar 18 20:37:23 crc kubenswrapper[4950]: I0318 20:37:23.675688 4950 generic.go:334] "Generic (PLEG): container finished" podID="651a2596-f562-4012-8f2a-7a3e1c562a0a" containerID="d63d38a73140c397d468f5213d693b05c42ef6f917d197623927e165b2ab0e18" exitCode=0 Mar 18 20:37:23 crc kubenswrapper[4950]: I0318 20:37:23.675730 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2vct7" event={"ID":"651a2596-f562-4012-8f2a-7a3e1c562a0a","Type":"ContainerDied","Data":"d63d38a73140c397d468f5213d693b05c42ef6f917d197623927e165b2ab0e18"} Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.123666 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2vct7" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.166031 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/651a2596-f562-4012-8f2a-7a3e1c562a0a-ssh-key-openstack-edpm-ipam\") pod \"651a2596-f562-4012-8f2a-7a3e1c562a0a\" (UID: \"651a2596-f562-4012-8f2a-7a3e1c562a0a\") " Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.166248 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/651a2596-f562-4012-8f2a-7a3e1c562a0a-inventory-0\") pod \"651a2596-f562-4012-8f2a-7a3e1c562a0a\" (UID: \"651a2596-f562-4012-8f2a-7a3e1c562a0a\") " Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.166313 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kj4z2\" (UniqueName: \"kubernetes.io/projected/651a2596-f562-4012-8f2a-7a3e1c562a0a-kube-api-access-kj4z2\") pod \"651a2596-f562-4012-8f2a-7a3e1c562a0a\" (UID: \"651a2596-f562-4012-8f2a-7a3e1c562a0a\") " Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.171657 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/651a2596-f562-4012-8f2a-7a3e1c562a0a-kube-api-access-kj4z2" (OuterVolumeSpecName: "kube-api-access-kj4z2") pod "651a2596-f562-4012-8f2a-7a3e1c562a0a" (UID: "651a2596-f562-4012-8f2a-7a3e1c562a0a"). InnerVolumeSpecName "kube-api-access-kj4z2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.201358 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/651a2596-f562-4012-8f2a-7a3e1c562a0a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "651a2596-f562-4012-8f2a-7a3e1c562a0a" (UID: "651a2596-f562-4012-8f2a-7a3e1c562a0a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.212997 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/651a2596-f562-4012-8f2a-7a3e1c562a0a-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "651a2596-f562-4012-8f2a-7a3e1c562a0a" (UID: "651a2596-f562-4012-8f2a-7a3e1c562a0a"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.268858 4950 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/651a2596-f562-4012-8f2a-7a3e1c562a0a-inventory-0\") on node \"crc\" DevicePath \"\"" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.268895 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kj4z2\" (UniqueName: \"kubernetes.io/projected/651a2596-f562-4012-8f2a-7a3e1c562a0a-kube-api-access-kj4z2\") on node \"crc\" DevicePath \"\"" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.268910 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/651a2596-f562-4012-8f2a-7a3e1c562a0a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.701751 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2vct7" event={"ID":"651a2596-f562-4012-8f2a-7a3e1c562a0a","Type":"ContainerDied","Data":"929a4e9d789bf3d9cdbd802065e1e1023c55a8249ce6caac22118bef9291e783"} Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.701811 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="929a4e9d789bf3d9cdbd802065e1e1023c55a8249ce6caac22118bef9291e783" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.701910 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2vct7" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.793892 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ww4zp"] Mar 18 20:37:25 crc kubenswrapper[4950]: E0318 20:37:25.794330 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="651a2596-f562-4012-8f2a-7a3e1c562a0a" containerName="ssh-known-hosts-edpm-deployment" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.794363 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="651a2596-f562-4012-8f2a-7a3e1c562a0a" containerName="ssh-known-hosts-edpm-deployment" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.794619 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="651a2596-f562-4012-8f2a-7a3e1c562a0a" containerName="ssh-known-hosts-edpm-deployment" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.795325 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ww4zp" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.798173 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.803810 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.805737 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.806442 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.815065 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ww4zp"] Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.878359 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2309d051-312e-45c6-8c31-28362aef9908-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ww4zp\" (UID: \"2309d051-312e-45c6-8c31-28362aef9908\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ww4zp" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.878446 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlb8b\" (UniqueName: \"kubernetes.io/projected/2309d051-312e-45c6-8c31-28362aef9908-kube-api-access-rlb8b\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ww4zp\" (UID: \"2309d051-312e-45c6-8c31-28362aef9908\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ww4zp" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.878577 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2309d051-312e-45c6-8c31-28362aef9908-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ww4zp\" (UID: \"2309d051-312e-45c6-8c31-28362aef9908\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ww4zp" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.980932 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2309d051-312e-45c6-8c31-28362aef9908-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ww4zp\" (UID: \"2309d051-312e-45c6-8c31-28362aef9908\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ww4zp" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.981025 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlb8b\" (UniqueName: \"kubernetes.io/projected/2309d051-312e-45c6-8c31-28362aef9908-kube-api-access-rlb8b\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ww4zp\" (UID: \"2309d051-312e-45c6-8c31-28362aef9908\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ww4zp" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.981048 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2309d051-312e-45c6-8c31-28362aef9908-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ww4zp\" (UID: \"2309d051-312e-45c6-8c31-28362aef9908\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ww4zp" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.986988 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2309d051-312e-45c6-8c31-28362aef9908-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ww4zp\" (UID: \"2309d051-312e-45c6-8c31-28362aef9908\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ww4zp" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.987155 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2309d051-312e-45c6-8c31-28362aef9908-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ww4zp\" (UID: \"2309d051-312e-45c6-8c31-28362aef9908\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ww4zp" Mar 18 20:37:25 crc kubenswrapper[4950]: I0318 20:37:25.997853 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlb8b\" (UniqueName: \"kubernetes.io/projected/2309d051-312e-45c6-8c31-28362aef9908-kube-api-access-rlb8b\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ww4zp\" (UID: \"2309d051-312e-45c6-8c31-28362aef9908\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ww4zp" Mar 18 20:37:26 crc kubenswrapper[4950]: I0318 20:37:26.111493 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ww4zp" Mar 18 20:37:26 crc kubenswrapper[4950]: I0318 20:37:26.639330 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ww4zp"] Mar 18 20:37:26 crc kubenswrapper[4950]: W0318 20:37:26.646588 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2309d051_312e_45c6_8c31_28362aef9908.slice/crio-f48f578b0596e1c59be9434be5252eeb328d4647003bc2338dc95e031daefb44 WatchSource:0}: Error finding container f48f578b0596e1c59be9434be5252eeb328d4647003bc2338dc95e031daefb44: Status 404 returned error can't find the container with id f48f578b0596e1c59be9434be5252eeb328d4647003bc2338dc95e031daefb44 Mar 18 20:37:26 crc kubenswrapper[4950]: I0318 20:37:26.710209 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ww4zp" event={"ID":"2309d051-312e-45c6-8c31-28362aef9908","Type":"ContainerStarted","Data":"f48f578b0596e1c59be9434be5252eeb328d4647003bc2338dc95e031daefb44"} Mar 18 20:37:27 crc kubenswrapper[4950]: I0318 20:37:27.722855 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ww4zp" event={"ID":"2309d051-312e-45c6-8c31-28362aef9908","Type":"ContainerStarted","Data":"8a2b008f71fb3e166d73cdac64b2766db32e331a36fa319f713b5a43296f1434"} Mar 18 20:37:27 crc kubenswrapper[4950]: I0318 20:37:27.762959 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ww4zp" podStartSLOduration=2.304572493 podStartE2EDuration="2.762910964s" podCreationTimestamp="2026-03-18 20:37:25 +0000 UTC" firstStartedPulling="2026-03-18 20:37:26.649203435 +0000 UTC m=+1859.890045343" lastFinishedPulling="2026-03-18 20:37:27.107541946 +0000 UTC m=+1860.348383814" observedRunningTime="2026-03-18 20:37:27.748319999 +0000 UTC m=+1860.989161877" watchObservedRunningTime="2026-03-18 20:37:27.762910964 +0000 UTC m=+1861.003752852" Mar 18 20:37:31 crc kubenswrapper[4950]: I0318 20:37:31.480618 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:37:31 crc kubenswrapper[4950]: E0318 20:37:31.481460 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:37:36 crc kubenswrapper[4950]: I0318 20:37:36.812334 4950 generic.go:334] "Generic (PLEG): container finished" podID="2309d051-312e-45c6-8c31-28362aef9908" containerID="8a2b008f71fb3e166d73cdac64b2766db32e331a36fa319f713b5a43296f1434" exitCode=0 Mar 18 20:37:36 crc kubenswrapper[4950]: I0318 20:37:36.813008 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ww4zp" event={"ID":"2309d051-312e-45c6-8c31-28362aef9908","Type":"ContainerDied","Data":"8a2b008f71fb3e166d73cdac64b2766db32e331a36fa319f713b5a43296f1434"} Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.191901 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ww4zp" Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.341474 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2309d051-312e-45c6-8c31-28362aef9908-inventory\") pod \"2309d051-312e-45c6-8c31-28362aef9908\" (UID: \"2309d051-312e-45c6-8c31-28362aef9908\") " Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.342595 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlb8b\" (UniqueName: \"kubernetes.io/projected/2309d051-312e-45c6-8c31-28362aef9908-kube-api-access-rlb8b\") pod \"2309d051-312e-45c6-8c31-28362aef9908\" (UID: \"2309d051-312e-45c6-8c31-28362aef9908\") " Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.342675 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2309d051-312e-45c6-8c31-28362aef9908-ssh-key-openstack-edpm-ipam\") pod \"2309d051-312e-45c6-8c31-28362aef9908\" (UID: \"2309d051-312e-45c6-8c31-28362aef9908\") " Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.347334 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2309d051-312e-45c6-8c31-28362aef9908-kube-api-access-rlb8b" (OuterVolumeSpecName: "kube-api-access-rlb8b") pod "2309d051-312e-45c6-8c31-28362aef9908" (UID: "2309d051-312e-45c6-8c31-28362aef9908"). InnerVolumeSpecName "kube-api-access-rlb8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.364899 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2309d051-312e-45c6-8c31-28362aef9908-inventory" (OuterVolumeSpecName: "inventory") pod "2309d051-312e-45c6-8c31-28362aef9908" (UID: "2309d051-312e-45c6-8c31-28362aef9908"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.377552 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2309d051-312e-45c6-8c31-28362aef9908-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2309d051-312e-45c6-8c31-28362aef9908" (UID: "2309d051-312e-45c6-8c31-28362aef9908"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.443843 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2309d051-312e-45c6-8c31-28362aef9908-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.443886 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlb8b\" (UniqueName: \"kubernetes.io/projected/2309d051-312e-45c6-8c31-28362aef9908-kube-api-access-rlb8b\") on node \"crc\" DevicePath \"\"" Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.443901 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2309d051-312e-45c6-8c31-28362aef9908-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.836530 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ww4zp" event={"ID":"2309d051-312e-45c6-8c31-28362aef9908","Type":"ContainerDied","Data":"f48f578b0596e1c59be9434be5252eeb328d4647003bc2338dc95e031daefb44"} Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.836603 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f48f578b0596e1c59be9434be5252eeb328d4647003bc2338dc95e031daefb44" Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.836635 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ww4zp" Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.926281 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj"] Mar 18 20:37:38 crc kubenswrapper[4950]: E0318 20:37:38.926924 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2309d051-312e-45c6-8c31-28362aef9908" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.926964 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="2309d051-312e-45c6-8c31-28362aef9908" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.927174 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="2309d051-312e-45c6-8c31-28362aef9908" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.927925 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj" Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.934218 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.934223 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.934592 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.936732 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.941356 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj"] Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.978178 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a75b3257-7409-45c9-acaa-d85cec7f9015-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj\" (UID: \"a75b3257-7409-45c9-acaa-d85cec7f9015\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj" Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.978267 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4d4lx\" (UniqueName: \"kubernetes.io/projected/a75b3257-7409-45c9-acaa-d85cec7f9015-kube-api-access-4d4lx\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj\" (UID: \"a75b3257-7409-45c9-acaa-d85cec7f9015\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj" Mar 18 20:37:38 crc kubenswrapper[4950]: I0318 20:37:38.978371 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a75b3257-7409-45c9-acaa-d85cec7f9015-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj\" (UID: \"a75b3257-7409-45c9-acaa-d85cec7f9015\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj" Mar 18 20:37:39 crc kubenswrapper[4950]: I0318 20:37:39.079519 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a75b3257-7409-45c9-acaa-d85cec7f9015-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj\" (UID: \"a75b3257-7409-45c9-acaa-d85cec7f9015\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj" Mar 18 20:37:39 crc kubenswrapper[4950]: I0318 20:37:39.079584 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4d4lx\" (UniqueName: \"kubernetes.io/projected/a75b3257-7409-45c9-acaa-d85cec7f9015-kube-api-access-4d4lx\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj\" (UID: \"a75b3257-7409-45c9-acaa-d85cec7f9015\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj" Mar 18 20:37:39 crc kubenswrapper[4950]: I0318 20:37:39.079614 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a75b3257-7409-45c9-acaa-d85cec7f9015-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj\" (UID: \"a75b3257-7409-45c9-acaa-d85cec7f9015\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj" Mar 18 20:37:39 crc kubenswrapper[4950]: I0318 20:37:39.092548 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a75b3257-7409-45c9-acaa-d85cec7f9015-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj\" (UID: \"a75b3257-7409-45c9-acaa-d85cec7f9015\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj" Mar 18 20:37:39 crc kubenswrapper[4950]: I0318 20:37:39.096865 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4d4lx\" (UniqueName: \"kubernetes.io/projected/a75b3257-7409-45c9-acaa-d85cec7f9015-kube-api-access-4d4lx\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj\" (UID: \"a75b3257-7409-45c9-acaa-d85cec7f9015\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj" Mar 18 20:37:39 crc kubenswrapper[4950]: I0318 20:37:39.100985 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a75b3257-7409-45c9-acaa-d85cec7f9015-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj\" (UID: \"a75b3257-7409-45c9-acaa-d85cec7f9015\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj" Mar 18 20:37:39 crc kubenswrapper[4950]: I0318 20:37:39.249711 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj" Mar 18 20:37:39 crc kubenswrapper[4950]: I0318 20:37:39.764768 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj"] Mar 18 20:37:39 crc kubenswrapper[4950]: I0318 20:37:39.848623 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj" event={"ID":"a75b3257-7409-45c9-acaa-d85cec7f9015","Type":"ContainerStarted","Data":"f253081c73209e47be062ae3c4987141deab5caa00ba2ff59ebd883c87ab6bd7"} Mar 18 20:37:40 crc kubenswrapper[4950]: I0318 20:37:40.858184 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj" event={"ID":"a75b3257-7409-45c9-acaa-d85cec7f9015","Type":"ContainerStarted","Data":"fce0212fcd35805d80c1c3dd6c623a0546a5e213b1ce61fa4658144f9b749eaa"} Mar 18 20:37:40 crc kubenswrapper[4950]: I0318 20:37:40.886524 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj" podStartSLOduration=2.422390946 podStartE2EDuration="2.886497569s" podCreationTimestamp="2026-03-18 20:37:38 +0000 UTC" firstStartedPulling="2026-03-18 20:37:39.782880176 +0000 UTC m=+1873.023722054" lastFinishedPulling="2026-03-18 20:37:40.246986809 +0000 UTC m=+1873.487828677" observedRunningTime="2026-03-18 20:37:40.87819823 +0000 UTC m=+1874.119040098" watchObservedRunningTime="2026-03-18 20:37:40.886497569 +0000 UTC m=+1874.127339437" Mar 18 20:37:42 crc kubenswrapper[4950]: I0318 20:37:42.480048 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:37:42 crc kubenswrapper[4950]: E0318 20:37:42.480626 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:37:48 crc kubenswrapper[4950]: I0318 20:37:48.043919 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5wz42"] Mar 18 20:37:48 crc kubenswrapper[4950]: I0318 20:37:48.050994 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5wz42"] Mar 18 20:37:48 crc kubenswrapper[4950]: I0318 20:37:48.097980 4950 scope.go:117] "RemoveContainer" containerID="17f18f1a155e9ef4b996c63aafaeda20d2259fcb48b022b2e9b7b4e3e7d1ff39" Mar 18 20:37:48 crc kubenswrapper[4950]: I0318 20:37:48.119822 4950 scope.go:117] "RemoveContainer" containerID="7fc7429f018879091bb018b4a873c962ed82b644be5c1aa9828b1bedb37c2e25" Mar 18 20:37:48 crc kubenswrapper[4950]: I0318 20:37:48.171533 4950 scope.go:117] "RemoveContainer" containerID="164cccb8565a81eb585d5aa123611b9f78b542e75607a9b63101fa870e906493" Mar 18 20:37:48 crc kubenswrapper[4950]: I0318 20:37:48.219138 4950 scope.go:117] "RemoveContainer" containerID="2a8c4760decf0fe38f5a95798f9b66c27c17fa33491aa829a44dd89f1950b2ee" Mar 18 20:37:48 crc kubenswrapper[4950]: I0318 20:37:48.258036 4950 scope.go:117] "RemoveContainer" containerID="45e8537d70c5d73022541ed9f24875d9a618a2231a495244009b73ccefb10bd2" Mar 18 20:37:48 crc kubenswrapper[4950]: I0318 20:37:48.295265 4950 scope.go:117] "RemoveContainer" containerID="eb35c59a9465151717c7e151ffbe1e5b8992d800790f92dd7e3a481dac5f00c0" Mar 18 20:37:49 crc kubenswrapper[4950]: I0318 20:37:49.491351 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4958284-3b28-4eff-bf68-16e2160710d1" path="/var/lib/kubelet/pods/d4958284-3b28-4eff-bf68-16e2160710d1/volumes" Mar 18 20:37:50 crc kubenswrapper[4950]: I0318 20:37:50.957533 4950 generic.go:334] "Generic (PLEG): container finished" podID="a75b3257-7409-45c9-acaa-d85cec7f9015" containerID="fce0212fcd35805d80c1c3dd6c623a0546a5e213b1ce61fa4658144f9b749eaa" exitCode=0 Mar 18 20:37:50 crc kubenswrapper[4950]: I0318 20:37:50.957670 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj" event={"ID":"a75b3257-7409-45c9-acaa-d85cec7f9015","Type":"ContainerDied","Data":"fce0212fcd35805d80c1c3dd6c623a0546a5e213b1ce61fa4658144f9b749eaa"} Mar 18 20:37:52 crc kubenswrapper[4950]: I0318 20:37:52.510495 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj" Mar 18 20:37:52 crc kubenswrapper[4950]: I0318 20:37:52.683729 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a75b3257-7409-45c9-acaa-d85cec7f9015-inventory\") pod \"a75b3257-7409-45c9-acaa-d85cec7f9015\" (UID: \"a75b3257-7409-45c9-acaa-d85cec7f9015\") " Mar 18 20:37:52 crc kubenswrapper[4950]: I0318 20:37:52.683834 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a75b3257-7409-45c9-acaa-d85cec7f9015-ssh-key-openstack-edpm-ipam\") pod \"a75b3257-7409-45c9-acaa-d85cec7f9015\" (UID: \"a75b3257-7409-45c9-acaa-d85cec7f9015\") " Mar 18 20:37:52 crc kubenswrapper[4950]: I0318 20:37:52.683959 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4lx\" (UniqueName: \"kubernetes.io/projected/a75b3257-7409-45c9-acaa-d85cec7f9015-kube-api-access-4d4lx\") pod \"a75b3257-7409-45c9-acaa-d85cec7f9015\" (UID: \"a75b3257-7409-45c9-acaa-d85cec7f9015\") " Mar 18 20:37:52 crc kubenswrapper[4950]: I0318 20:37:52.689977 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a75b3257-7409-45c9-acaa-d85cec7f9015-kube-api-access-4d4lx" (OuterVolumeSpecName: "kube-api-access-4d4lx") pod "a75b3257-7409-45c9-acaa-d85cec7f9015" (UID: "a75b3257-7409-45c9-acaa-d85cec7f9015"). InnerVolumeSpecName "kube-api-access-4d4lx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:37:52 crc kubenswrapper[4950]: I0318 20:37:52.708581 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a75b3257-7409-45c9-acaa-d85cec7f9015-inventory" (OuterVolumeSpecName: "inventory") pod "a75b3257-7409-45c9-acaa-d85cec7f9015" (UID: "a75b3257-7409-45c9-acaa-d85cec7f9015"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:37:52 crc kubenswrapper[4950]: I0318 20:37:52.730576 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a75b3257-7409-45c9-acaa-d85cec7f9015-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a75b3257-7409-45c9-acaa-d85cec7f9015" (UID: "a75b3257-7409-45c9-acaa-d85cec7f9015"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:37:52 crc kubenswrapper[4950]: I0318 20:37:52.787230 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a75b3257-7409-45c9-acaa-d85cec7f9015-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:37:52 crc kubenswrapper[4950]: I0318 20:37:52.787263 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a75b3257-7409-45c9-acaa-d85cec7f9015-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:37:52 crc kubenswrapper[4950]: I0318 20:37:52.787274 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4lx\" (UniqueName: \"kubernetes.io/projected/a75b3257-7409-45c9-acaa-d85cec7f9015-kube-api-access-4d4lx\") on node \"crc\" DevicePath \"\"" Mar 18 20:37:52 crc kubenswrapper[4950]: I0318 20:37:52.983366 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj" event={"ID":"a75b3257-7409-45c9-acaa-d85cec7f9015","Type":"ContainerDied","Data":"f253081c73209e47be062ae3c4987141deab5caa00ba2ff59ebd883c87ab6bd7"} Mar 18 20:37:52 crc kubenswrapper[4950]: I0318 20:37:52.983403 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f253081c73209e47be062ae3c4987141deab5caa00ba2ff59ebd883c87ab6bd7" Mar 18 20:37:52 crc kubenswrapper[4950]: I0318 20:37:52.983467 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj" Mar 18 20:37:53 crc kubenswrapper[4950]: I0318 20:37:53.480643 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:37:53 crc kubenswrapper[4950]: E0318 20:37:53.481014 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:38:00 crc kubenswrapper[4950]: I0318 20:38:00.177673 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564438-6lsrq"] Mar 18 20:38:00 crc kubenswrapper[4950]: E0318 20:38:00.179198 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a75b3257-7409-45c9-acaa-d85cec7f9015" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:38:00 crc kubenswrapper[4950]: I0318 20:38:00.179232 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="a75b3257-7409-45c9-acaa-d85cec7f9015" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:38:00 crc kubenswrapper[4950]: I0318 20:38:00.179695 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="a75b3257-7409-45c9-acaa-d85cec7f9015" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:38:00 crc kubenswrapper[4950]: I0318 20:38:00.180971 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564438-6lsrq" Mar 18 20:38:00 crc kubenswrapper[4950]: I0318 20:38:00.184111 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:38:00 crc kubenswrapper[4950]: I0318 20:38:00.184524 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:38:00 crc kubenswrapper[4950]: I0318 20:38:00.184816 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:38:00 crc kubenswrapper[4950]: I0318 20:38:00.199093 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564438-6lsrq"] Mar 18 20:38:00 crc kubenswrapper[4950]: I0318 20:38:00.339378 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gh5bm\" (UniqueName: \"kubernetes.io/projected/2c77a603-7308-4467-a365-a47380f4954f-kube-api-access-gh5bm\") pod \"auto-csr-approver-29564438-6lsrq\" (UID: \"2c77a603-7308-4467-a365-a47380f4954f\") " pod="openshift-infra/auto-csr-approver-29564438-6lsrq" Mar 18 20:38:00 crc kubenswrapper[4950]: I0318 20:38:00.441962 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gh5bm\" (UniqueName: \"kubernetes.io/projected/2c77a603-7308-4467-a365-a47380f4954f-kube-api-access-gh5bm\") pod \"auto-csr-approver-29564438-6lsrq\" (UID: \"2c77a603-7308-4467-a365-a47380f4954f\") " pod="openshift-infra/auto-csr-approver-29564438-6lsrq" Mar 18 20:38:00 crc kubenswrapper[4950]: I0318 20:38:00.462201 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gh5bm\" (UniqueName: \"kubernetes.io/projected/2c77a603-7308-4467-a365-a47380f4954f-kube-api-access-gh5bm\") pod \"auto-csr-approver-29564438-6lsrq\" (UID: \"2c77a603-7308-4467-a365-a47380f4954f\") " pod="openshift-infra/auto-csr-approver-29564438-6lsrq" Mar 18 20:38:00 crc kubenswrapper[4950]: I0318 20:38:00.514562 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564438-6lsrq" Mar 18 20:38:00 crc kubenswrapper[4950]: I0318 20:38:00.983272 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564438-6lsrq"] Mar 18 20:38:00 crc kubenswrapper[4950]: W0318 20:38:00.994806 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c77a603_7308_4467_a365_a47380f4954f.slice/crio-4551103a3844c073dd64c5d924f816102b27b89aa074822955575b1a4ef2bbe6 WatchSource:0}: Error finding container 4551103a3844c073dd64c5d924f816102b27b89aa074822955575b1a4ef2bbe6: Status 404 returned error can't find the container with id 4551103a3844c073dd64c5d924f816102b27b89aa074822955575b1a4ef2bbe6 Mar 18 20:38:01 crc kubenswrapper[4950]: I0318 20:38:01.074223 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564438-6lsrq" event={"ID":"2c77a603-7308-4467-a365-a47380f4954f","Type":"ContainerStarted","Data":"4551103a3844c073dd64c5d924f816102b27b89aa074822955575b1a4ef2bbe6"} Mar 18 20:38:03 crc kubenswrapper[4950]: I0318 20:38:03.097057 4950 generic.go:334] "Generic (PLEG): container finished" podID="2c77a603-7308-4467-a365-a47380f4954f" containerID="bfb5f2081a3871a172aa70452dcf9a517756cd507b27dae22e2b7c5c719cfa38" exitCode=0 Mar 18 20:38:03 crc kubenswrapper[4950]: I0318 20:38:03.097475 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564438-6lsrq" event={"ID":"2c77a603-7308-4467-a365-a47380f4954f","Type":"ContainerDied","Data":"bfb5f2081a3871a172aa70452dcf9a517756cd507b27dae22e2b7c5c719cfa38"} Mar 18 20:38:05 crc kubenswrapper[4950]: I0318 20:38:05.241160 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564438-6lsrq" Mar 18 20:38:05 crc kubenswrapper[4950]: I0318 20:38:05.426891 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gh5bm\" (UniqueName: \"kubernetes.io/projected/2c77a603-7308-4467-a365-a47380f4954f-kube-api-access-gh5bm\") pod \"2c77a603-7308-4467-a365-a47380f4954f\" (UID: \"2c77a603-7308-4467-a365-a47380f4954f\") " Mar 18 20:38:05 crc kubenswrapper[4950]: I0318 20:38:05.440564 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c77a603-7308-4467-a365-a47380f4954f-kube-api-access-gh5bm" (OuterVolumeSpecName: "kube-api-access-gh5bm") pod "2c77a603-7308-4467-a365-a47380f4954f" (UID: "2c77a603-7308-4467-a365-a47380f4954f"). InnerVolumeSpecName "kube-api-access-gh5bm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:38:05 crc kubenswrapper[4950]: I0318 20:38:05.529286 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gh5bm\" (UniqueName: \"kubernetes.io/projected/2c77a603-7308-4467-a365-a47380f4954f-kube-api-access-gh5bm\") on node \"crc\" DevicePath \"\"" Mar 18 20:38:06 crc kubenswrapper[4950]: I0318 20:38:06.145390 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564438-6lsrq" event={"ID":"2c77a603-7308-4467-a365-a47380f4954f","Type":"ContainerDied","Data":"4551103a3844c073dd64c5d924f816102b27b89aa074822955575b1a4ef2bbe6"} Mar 18 20:38:06 crc kubenswrapper[4950]: I0318 20:38:06.145806 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4551103a3844c073dd64c5d924f816102b27b89aa074822955575b1a4ef2bbe6" Mar 18 20:38:06 crc kubenswrapper[4950]: I0318 20:38:06.145526 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564438-6lsrq" Mar 18 20:38:06 crc kubenswrapper[4950]: I0318 20:38:06.325036 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564432-qs5nb"] Mar 18 20:38:06 crc kubenswrapper[4950]: I0318 20:38:06.333819 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564432-qs5nb"] Mar 18 20:38:06 crc kubenswrapper[4950]: I0318 20:38:06.479680 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:38:06 crc kubenswrapper[4950]: E0318 20:38:06.480235 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:38:07 crc kubenswrapper[4950]: I0318 20:38:07.044562 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-n6btl"] Mar 18 20:38:07 crc kubenswrapper[4950]: I0318 20:38:07.058077 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-n6btl"] Mar 18 20:38:07 crc kubenswrapper[4950]: I0318 20:38:07.489719 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d3a6361-b1e7-4192-b620-71bb92c2844a" path="/var/lib/kubelet/pods/3d3a6361-b1e7-4192-b620-71bb92c2844a/volumes" Mar 18 20:38:07 crc kubenswrapper[4950]: I0318 20:38:07.490561 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c71dbcb-5a73-4c85-8a1f-fa389a0e284f" path="/var/lib/kubelet/pods/9c71dbcb-5a73-4c85-8a1f-fa389a0e284f/volumes" Mar 18 20:38:11 crc kubenswrapper[4950]: I0318 20:38:11.040532 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lzx4p"] Mar 18 20:38:11 crc kubenswrapper[4950]: I0318 20:38:11.051967 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lzx4p"] Mar 18 20:38:11 crc kubenswrapper[4950]: I0318 20:38:11.491608 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff62857a-136b-4b75-abf8-655bab5c8984" path="/var/lib/kubelet/pods/ff62857a-136b-4b75-abf8-655bab5c8984/volumes" Mar 18 20:38:11 crc kubenswrapper[4950]: I0318 20:38:11.881476 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vsm85"] Mar 18 20:38:11 crc kubenswrapper[4950]: E0318 20:38:11.882326 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c77a603-7308-4467-a365-a47380f4954f" containerName="oc" Mar 18 20:38:11 crc kubenswrapper[4950]: I0318 20:38:11.882442 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c77a603-7308-4467-a365-a47380f4954f" containerName="oc" Mar 18 20:38:11 crc kubenswrapper[4950]: I0318 20:38:11.882738 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c77a603-7308-4467-a365-a47380f4954f" containerName="oc" Mar 18 20:38:11 crc kubenswrapper[4950]: I0318 20:38:11.884942 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vsm85" Mar 18 20:38:11 crc kubenswrapper[4950]: I0318 20:38:11.903775 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vsm85"] Mar 18 20:38:11 crc kubenswrapper[4950]: I0318 20:38:11.951386 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48729719-ff85-4882-bc36-773a924f9987-catalog-content\") pod \"redhat-operators-vsm85\" (UID: \"48729719-ff85-4882-bc36-773a924f9987\") " pod="openshift-marketplace/redhat-operators-vsm85" Mar 18 20:38:11 crc kubenswrapper[4950]: I0318 20:38:11.951670 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48729719-ff85-4882-bc36-773a924f9987-utilities\") pod \"redhat-operators-vsm85\" (UID: \"48729719-ff85-4882-bc36-773a924f9987\") " pod="openshift-marketplace/redhat-operators-vsm85" Mar 18 20:38:11 crc kubenswrapper[4950]: I0318 20:38:11.951763 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsd9k\" (UniqueName: \"kubernetes.io/projected/48729719-ff85-4882-bc36-773a924f9987-kube-api-access-dsd9k\") pod \"redhat-operators-vsm85\" (UID: \"48729719-ff85-4882-bc36-773a924f9987\") " pod="openshift-marketplace/redhat-operators-vsm85" Mar 18 20:38:12 crc kubenswrapper[4950]: I0318 20:38:12.054401 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48729719-ff85-4882-bc36-773a924f9987-catalog-content\") pod \"redhat-operators-vsm85\" (UID: \"48729719-ff85-4882-bc36-773a924f9987\") " pod="openshift-marketplace/redhat-operators-vsm85" Mar 18 20:38:12 crc kubenswrapper[4950]: I0318 20:38:12.055267 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48729719-ff85-4882-bc36-773a924f9987-catalog-content\") pod \"redhat-operators-vsm85\" (UID: \"48729719-ff85-4882-bc36-773a924f9987\") " pod="openshift-marketplace/redhat-operators-vsm85" Mar 18 20:38:12 crc kubenswrapper[4950]: I0318 20:38:12.055653 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48729719-ff85-4882-bc36-773a924f9987-utilities\") pod \"redhat-operators-vsm85\" (UID: \"48729719-ff85-4882-bc36-773a924f9987\") " pod="openshift-marketplace/redhat-operators-vsm85" Mar 18 20:38:12 crc kubenswrapper[4950]: I0318 20:38:12.055906 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsd9k\" (UniqueName: \"kubernetes.io/projected/48729719-ff85-4882-bc36-773a924f9987-kube-api-access-dsd9k\") pod \"redhat-operators-vsm85\" (UID: \"48729719-ff85-4882-bc36-773a924f9987\") " pod="openshift-marketplace/redhat-operators-vsm85" Mar 18 20:38:12 crc kubenswrapper[4950]: I0318 20:38:12.056286 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48729719-ff85-4882-bc36-773a924f9987-utilities\") pod \"redhat-operators-vsm85\" (UID: \"48729719-ff85-4882-bc36-773a924f9987\") " pod="openshift-marketplace/redhat-operators-vsm85" Mar 18 20:38:12 crc kubenswrapper[4950]: I0318 20:38:12.078705 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsd9k\" (UniqueName: \"kubernetes.io/projected/48729719-ff85-4882-bc36-773a924f9987-kube-api-access-dsd9k\") pod \"redhat-operators-vsm85\" (UID: \"48729719-ff85-4882-bc36-773a924f9987\") " pod="openshift-marketplace/redhat-operators-vsm85" Mar 18 20:38:12 crc kubenswrapper[4950]: I0318 20:38:12.213602 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vsm85" Mar 18 20:38:12 crc kubenswrapper[4950]: I0318 20:38:12.724574 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vsm85"] Mar 18 20:38:13 crc kubenswrapper[4950]: I0318 20:38:13.216241 4950 generic.go:334] "Generic (PLEG): container finished" podID="48729719-ff85-4882-bc36-773a924f9987" containerID="9fa67b7296a28540f7934eebcfdf5c8ca921e242c0d82d87443f5f7b6f75b689" exitCode=0 Mar 18 20:38:13 crc kubenswrapper[4950]: I0318 20:38:13.216338 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vsm85" event={"ID":"48729719-ff85-4882-bc36-773a924f9987","Type":"ContainerDied","Data":"9fa67b7296a28540f7934eebcfdf5c8ca921e242c0d82d87443f5f7b6f75b689"} Mar 18 20:38:13 crc kubenswrapper[4950]: I0318 20:38:13.217525 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vsm85" event={"ID":"48729719-ff85-4882-bc36-773a924f9987","Type":"ContainerStarted","Data":"da0628457c7b83e208b89cce58099c66982f84d52058d647fd311f601cb6543b"} Mar 18 20:38:14 crc kubenswrapper[4950]: I0318 20:38:14.228301 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vsm85" event={"ID":"48729719-ff85-4882-bc36-773a924f9987","Type":"ContainerStarted","Data":"857326e8d5c1f67e00adb9c24b5658a4041ac30b00037e6c6d6a63be7fcd59b2"} Mar 18 20:38:17 crc kubenswrapper[4950]: I0318 20:38:17.498974 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:38:17 crc kubenswrapper[4950]: E0318 20:38:17.500275 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:38:20 crc kubenswrapper[4950]: I0318 20:38:20.285799 4950 generic.go:334] "Generic (PLEG): container finished" podID="48729719-ff85-4882-bc36-773a924f9987" containerID="857326e8d5c1f67e00adb9c24b5658a4041ac30b00037e6c6d6a63be7fcd59b2" exitCode=0 Mar 18 20:38:20 crc kubenswrapper[4950]: I0318 20:38:20.285872 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vsm85" event={"ID":"48729719-ff85-4882-bc36-773a924f9987","Type":"ContainerDied","Data":"857326e8d5c1f67e00adb9c24b5658a4041ac30b00037e6c6d6a63be7fcd59b2"} Mar 18 20:38:21 crc kubenswrapper[4950]: I0318 20:38:21.297388 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vsm85" event={"ID":"48729719-ff85-4882-bc36-773a924f9987","Type":"ContainerStarted","Data":"a17ad7cd407068977cefa47281c6cfbaa06d152426bfa04a4baef91c9c573ec3"} Mar 18 20:38:21 crc kubenswrapper[4950]: I0318 20:38:21.322133 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vsm85" podStartSLOduration=2.625047371 podStartE2EDuration="10.322118602s" podCreationTimestamp="2026-03-18 20:38:11 +0000 UTC" firstStartedPulling="2026-03-18 20:38:13.21811568 +0000 UTC m=+1906.458957558" lastFinishedPulling="2026-03-18 20:38:20.915186921 +0000 UTC m=+1914.156028789" observedRunningTime="2026-03-18 20:38:21.316070192 +0000 UTC m=+1914.556912060" watchObservedRunningTime="2026-03-18 20:38:21.322118602 +0000 UTC m=+1914.562960470" Mar 18 20:38:22 crc kubenswrapper[4950]: I0318 20:38:22.214148 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vsm85" Mar 18 20:38:22 crc kubenswrapper[4950]: I0318 20:38:22.215109 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vsm85" Mar 18 20:38:23 crc kubenswrapper[4950]: I0318 20:38:23.266348 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vsm85" podUID="48729719-ff85-4882-bc36-773a924f9987" containerName="registry-server" probeResult="failure" output=< Mar 18 20:38:23 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 20:38:23 crc kubenswrapper[4950]: > Mar 18 20:38:30 crc kubenswrapper[4950]: I0318 20:38:30.484109 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:38:30 crc kubenswrapper[4950]: E0318 20:38:30.484785 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:38:32 crc kubenswrapper[4950]: I0318 20:38:32.264148 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vsm85" Mar 18 20:38:32 crc kubenswrapper[4950]: I0318 20:38:32.322385 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vsm85" Mar 18 20:38:32 crc kubenswrapper[4950]: I0318 20:38:32.498169 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vsm85"] Mar 18 20:38:33 crc kubenswrapper[4950]: I0318 20:38:33.393759 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vsm85" podUID="48729719-ff85-4882-bc36-773a924f9987" containerName="registry-server" containerID="cri-o://a17ad7cd407068977cefa47281c6cfbaa06d152426bfa04a4baef91c9c573ec3" gracePeriod=2 Mar 18 20:38:33 crc kubenswrapper[4950]: I0318 20:38:33.801534 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vsm85" Mar 18 20:38:33 crc kubenswrapper[4950]: I0318 20:38:33.992099 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsd9k\" (UniqueName: \"kubernetes.io/projected/48729719-ff85-4882-bc36-773a924f9987-kube-api-access-dsd9k\") pod \"48729719-ff85-4882-bc36-773a924f9987\" (UID: \"48729719-ff85-4882-bc36-773a924f9987\") " Mar 18 20:38:33 crc kubenswrapper[4950]: I0318 20:38:33.992572 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48729719-ff85-4882-bc36-773a924f9987-utilities\") pod \"48729719-ff85-4882-bc36-773a924f9987\" (UID: \"48729719-ff85-4882-bc36-773a924f9987\") " Mar 18 20:38:33 crc kubenswrapper[4950]: I0318 20:38:33.992701 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48729719-ff85-4882-bc36-773a924f9987-catalog-content\") pod \"48729719-ff85-4882-bc36-773a924f9987\" (UID: \"48729719-ff85-4882-bc36-773a924f9987\") " Mar 18 20:38:33 crc kubenswrapper[4950]: I0318 20:38:33.993072 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48729719-ff85-4882-bc36-773a924f9987-utilities" (OuterVolumeSpecName: "utilities") pod "48729719-ff85-4882-bc36-773a924f9987" (UID: "48729719-ff85-4882-bc36-773a924f9987"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:38:33 crc kubenswrapper[4950]: I0318 20:38:33.993191 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48729719-ff85-4882-bc36-773a924f9987-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:38:34 crc kubenswrapper[4950]: I0318 20:38:34.006690 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48729719-ff85-4882-bc36-773a924f9987-kube-api-access-dsd9k" (OuterVolumeSpecName: "kube-api-access-dsd9k") pod "48729719-ff85-4882-bc36-773a924f9987" (UID: "48729719-ff85-4882-bc36-773a924f9987"). InnerVolumeSpecName "kube-api-access-dsd9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:38:34 crc kubenswrapper[4950]: I0318 20:38:34.095050 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsd9k\" (UniqueName: \"kubernetes.io/projected/48729719-ff85-4882-bc36-773a924f9987-kube-api-access-dsd9k\") on node \"crc\" DevicePath \"\"" Mar 18 20:38:34 crc kubenswrapper[4950]: I0318 20:38:34.166039 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48729719-ff85-4882-bc36-773a924f9987-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "48729719-ff85-4882-bc36-773a924f9987" (UID: "48729719-ff85-4882-bc36-773a924f9987"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:38:34 crc kubenswrapper[4950]: I0318 20:38:34.196240 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48729719-ff85-4882-bc36-773a924f9987-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:38:34 crc kubenswrapper[4950]: I0318 20:38:34.404872 4950 generic.go:334] "Generic (PLEG): container finished" podID="48729719-ff85-4882-bc36-773a924f9987" containerID="a17ad7cd407068977cefa47281c6cfbaa06d152426bfa04a4baef91c9c573ec3" exitCode=0 Mar 18 20:38:34 crc kubenswrapper[4950]: I0318 20:38:34.404918 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vsm85" event={"ID":"48729719-ff85-4882-bc36-773a924f9987","Type":"ContainerDied","Data":"a17ad7cd407068977cefa47281c6cfbaa06d152426bfa04a4baef91c9c573ec3"} Mar 18 20:38:34 crc kubenswrapper[4950]: I0318 20:38:34.404943 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vsm85" Mar 18 20:38:34 crc kubenswrapper[4950]: I0318 20:38:34.404964 4950 scope.go:117] "RemoveContainer" containerID="a17ad7cd407068977cefa47281c6cfbaa06d152426bfa04a4baef91c9c573ec3" Mar 18 20:38:34 crc kubenswrapper[4950]: I0318 20:38:34.404950 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vsm85" event={"ID":"48729719-ff85-4882-bc36-773a924f9987","Type":"ContainerDied","Data":"da0628457c7b83e208b89cce58099c66982f84d52058d647fd311f601cb6543b"} Mar 18 20:38:34 crc kubenswrapper[4950]: I0318 20:38:34.446112 4950 scope.go:117] "RemoveContainer" containerID="857326e8d5c1f67e00adb9c24b5658a4041ac30b00037e6c6d6a63be7fcd59b2" Mar 18 20:38:34 crc kubenswrapper[4950]: I0318 20:38:34.449069 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vsm85"] Mar 18 20:38:34 crc kubenswrapper[4950]: I0318 20:38:34.461465 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vsm85"] Mar 18 20:38:34 crc kubenswrapper[4950]: I0318 20:38:34.477201 4950 scope.go:117] "RemoveContainer" containerID="9fa67b7296a28540f7934eebcfdf5c8ca921e242c0d82d87443f5f7b6f75b689" Mar 18 20:38:34 crc kubenswrapper[4950]: I0318 20:38:34.506206 4950 scope.go:117] "RemoveContainer" containerID="a17ad7cd407068977cefa47281c6cfbaa06d152426bfa04a4baef91c9c573ec3" Mar 18 20:38:34 crc kubenswrapper[4950]: E0318 20:38:34.507075 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a17ad7cd407068977cefa47281c6cfbaa06d152426bfa04a4baef91c9c573ec3\": container with ID starting with a17ad7cd407068977cefa47281c6cfbaa06d152426bfa04a4baef91c9c573ec3 not found: ID does not exist" containerID="a17ad7cd407068977cefa47281c6cfbaa06d152426bfa04a4baef91c9c573ec3" Mar 18 20:38:34 crc kubenswrapper[4950]: I0318 20:38:34.507190 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a17ad7cd407068977cefa47281c6cfbaa06d152426bfa04a4baef91c9c573ec3"} err="failed to get container status \"a17ad7cd407068977cefa47281c6cfbaa06d152426bfa04a4baef91c9c573ec3\": rpc error: code = NotFound desc = could not find container \"a17ad7cd407068977cefa47281c6cfbaa06d152426bfa04a4baef91c9c573ec3\": container with ID starting with a17ad7cd407068977cefa47281c6cfbaa06d152426bfa04a4baef91c9c573ec3 not found: ID does not exist" Mar 18 20:38:34 crc kubenswrapper[4950]: I0318 20:38:34.507219 4950 scope.go:117] "RemoveContainer" containerID="857326e8d5c1f67e00adb9c24b5658a4041ac30b00037e6c6d6a63be7fcd59b2" Mar 18 20:38:34 crc kubenswrapper[4950]: E0318 20:38:34.507820 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"857326e8d5c1f67e00adb9c24b5658a4041ac30b00037e6c6d6a63be7fcd59b2\": container with ID starting with 857326e8d5c1f67e00adb9c24b5658a4041ac30b00037e6c6d6a63be7fcd59b2 not found: ID does not exist" containerID="857326e8d5c1f67e00adb9c24b5658a4041ac30b00037e6c6d6a63be7fcd59b2" Mar 18 20:38:34 crc kubenswrapper[4950]: I0318 20:38:34.507888 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"857326e8d5c1f67e00adb9c24b5658a4041ac30b00037e6c6d6a63be7fcd59b2"} err="failed to get container status \"857326e8d5c1f67e00adb9c24b5658a4041ac30b00037e6c6d6a63be7fcd59b2\": rpc error: code = NotFound desc = could not find container \"857326e8d5c1f67e00adb9c24b5658a4041ac30b00037e6c6d6a63be7fcd59b2\": container with ID starting with 857326e8d5c1f67e00adb9c24b5658a4041ac30b00037e6c6d6a63be7fcd59b2 not found: ID does not exist" Mar 18 20:38:34 crc kubenswrapper[4950]: I0318 20:38:34.507917 4950 scope.go:117] "RemoveContainer" containerID="9fa67b7296a28540f7934eebcfdf5c8ca921e242c0d82d87443f5f7b6f75b689" Mar 18 20:38:34 crc kubenswrapper[4950]: E0318 20:38:34.508502 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fa67b7296a28540f7934eebcfdf5c8ca921e242c0d82d87443f5f7b6f75b689\": container with ID starting with 9fa67b7296a28540f7934eebcfdf5c8ca921e242c0d82d87443f5f7b6f75b689 not found: ID does not exist" containerID="9fa67b7296a28540f7934eebcfdf5c8ca921e242c0d82d87443f5f7b6f75b689" Mar 18 20:38:34 crc kubenswrapper[4950]: I0318 20:38:34.508532 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fa67b7296a28540f7934eebcfdf5c8ca921e242c0d82d87443f5f7b6f75b689"} err="failed to get container status \"9fa67b7296a28540f7934eebcfdf5c8ca921e242c0d82d87443f5f7b6f75b689\": rpc error: code = NotFound desc = could not find container \"9fa67b7296a28540f7934eebcfdf5c8ca921e242c0d82d87443f5f7b6f75b689\": container with ID starting with 9fa67b7296a28540f7934eebcfdf5c8ca921e242c0d82d87443f5f7b6f75b689 not found: ID does not exist" Mar 18 20:38:35 crc kubenswrapper[4950]: I0318 20:38:35.489364 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48729719-ff85-4882-bc36-773a924f9987" path="/var/lib/kubelet/pods/48729719-ff85-4882-bc36-773a924f9987/volumes" Mar 18 20:38:43 crc kubenswrapper[4950]: I0318 20:38:43.479985 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:38:43 crc kubenswrapper[4950]: E0318 20:38:43.480820 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:38:48 crc kubenswrapper[4950]: I0318 20:38:48.425696 4950 scope.go:117] "RemoveContainer" containerID="1c4256450e87d1329579e44017780b5ac367ecc5a866b74fcd756d459d685bd9" Mar 18 20:38:48 crc kubenswrapper[4950]: I0318 20:38:48.488842 4950 scope.go:117] "RemoveContainer" containerID="9b8cd39ebaa272935fa8e4cd424226e975751123cef5c91404d6c296fc9242b4" Mar 18 20:38:48 crc kubenswrapper[4950]: I0318 20:38:48.550362 4950 scope.go:117] "RemoveContainer" containerID="0e584954e3a8b88f9cc604d88832e1d8a7009b70353dc58272712cf61225203c" Mar 18 20:38:48 crc kubenswrapper[4950]: I0318 20:38:48.597323 4950 scope.go:117] "RemoveContainer" containerID="a96279bdd3cc64e412ca56c34ddd8aad698c5e309382d4ce2300837520cebaab" Mar 18 20:38:53 crc kubenswrapper[4950]: I0318 20:38:53.062044 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-lj42n"] Mar 18 20:38:53 crc kubenswrapper[4950]: I0318 20:38:53.073091 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-lj42n"] Mar 18 20:38:53 crc kubenswrapper[4950]: I0318 20:38:53.497829 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f58dabee-7195-40e9-952e-9dfeefc1ee08" path="/var/lib/kubelet/pods/f58dabee-7195-40e9-952e-9dfeefc1ee08/volumes" Mar 18 20:38:58 crc kubenswrapper[4950]: I0318 20:38:58.480053 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:38:58 crc kubenswrapper[4950]: E0318 20:38:58.481062 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:39:10 crc kubenswrapper[4950]: I0318 20:39:10.481332 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:39:10 crc kubenswrapper[4950]: E0318 20:39:10.482328 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:39:21 crc kubenswrapper[4950]: I0318 20:39:21.480522 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:39:21 crc kubenswrapper[4950]: E0318 20:39:21.481387 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:39:33 crc kubenswrapper[4950]: I0318 20:39:33.482318 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:39:33 crc kubenswrapper[4950]: E0318 20:39:33.482968 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:39:45 crc kubenswrapper[4950]: I0318 20:39:45.481629 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:39:45 crc kubenswrapper[4950]: E0318 20:39:45.482828 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:39:48 crc kubenswrapper[4950]: I0318 20:39:48.742408 4950 scope.go:117] "RemoveContainer" containerID="656f6974b9820f48c5b34d3c8007059ae8bf8ad5ac44bf59d04c6ad0ee0e58f1" Mar 18 20:39:58 crc kubenswrapper[4950]: I0318 20:39:58.480367 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:39:58 crc kubenswrapper[4950]: E0318 20:39:58.481721 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:40:00 crc kubenswrapper[4950]: I0318 20:40:00.169461 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564440-jqvbk"] Mar 18 20:40:00 crc kubenswrapper[4950]: E0318 20:40:00.170293 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48729719-ff85-4882-bc36-773a924f9987" containerName="extract-utilities" Mar 18 20:40:00 crc kubenswrapper[4950]: I0318 20:40:00.170317 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="48729719-ff85-4882-bc36-773a924f9987" containerName="extract-utilities" Mar 18 20:40:00 crc kubenswrapper[4950]: E0318 20:40:00.170333 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48729719-ff85-4882-bc36-773a924f9987" containerName="registry-server" Mar 18 20:40:00 crc kubenswrapper[4950]: I0318 20:40:00.170343 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="48729719-ff85-4882-bc36-773a924f9987" containerName="registry-server" Mar 18 20:40:00 crc kubenswrapper[4950]: E0318 20:40:00.170370 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48729719-ff85-4882-bc36-773a924f9987" containerName="extract-content" Mar 18 20:40:00 crc kubenswrapper[4950]: I0318 20:40:00.170381 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="48729719-ff85-4882-bc36-773a924f9987" containerName="extract-content" Mar 18 20:40:00 crc kubenswrapper[4950]: I0318 20:40:00.170717 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="48729719-ff85-4882-bc36-773a924f9987" containerName="registry-server" Mar 18 20:40:00 crc kubenswrapper[4950]: I0318 20:40:00.171631 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564440-jqvbk" Mar 18 20:40:00 crc kubenswrapper[4950]: I0318 20:40:00.174561 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:40:00 crc kubenswrapper[4950]: I0318 20:40:00.175968 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:40:00 crc kubenswrapper[4950]: I0318 20:40:00.176530 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:40:00 crc kubenswrapper[4950]: I0318 20:40:00.187815 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564440-jqvbk"] Mar 18 20:40:00 crc kubenswrapper[4950]: I0318 20:40:00.255027 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vk8d\" (UniqueName: \"kubernetes.io/projected/8149017b-bc82-441c-8ede-ec04b1ab0614-kube-api-access-7vk8d\") pod \"auto-csr-approver-29564440-jqvbk\" (UID: \"8149017b-bc82-441c-8ede-ec04b1ab0614\") " pod="openshift-infra/auto-csr-approver-29564440-jqvbk" Mar 18 20:40:00 crc kubenswrapper[4950]: I0318 20:40:00.356649 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vk8d\" (UniqueName: \"kubernetes.io/projected/8149017b-bc82-441c-8ede-ec04b1ab0614-kube-api-access-7vk8d\") pod \"auto-csr-approver-29564440-jqvbk\" (UID: \"8149017b-bc82-441c-8ede-ec04b1ab0614\") " pod="openshift-infra/auto-csr-approver-29564440-jqvbk" Mar 18 20:40:00 crc kubenswrapper[4950]: I0318 20:40:00.378522 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vk8d\" (UniqueName: \"kubernetes.io/projected/8149017b-bc82-441c-8ede-ec04b1ab0614-kube-api-access-7vk8d\") pod \"auto-csr-approver-29564440-jqvbk\" (UID: \"8149017b-bc82-441c-8ede-ec04b1ab0614\") " pod="openshift-infra/auto-csr-approver-29564440-jqvbk" Mar 18 20:40:00 crc kubenswrapper[4950]: I0318 20:40:00.490351 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564440-jqvbk" Mar 18 20:40:00 crc kubenswrapper[4950]: I0318 20:40:00.981043 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564440-jqvbk"] Mar 18 20:40:01 crc kubenswrapper[4950]: I0318 20:40:01.286075 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564440-jqvbk" event={"ID":"8149017b-bc82-441c-8ede-ec04b1ab0614","Type":"ContainerStarted","Data":"e20ace086cd6975dd1328d6e551de1294ef1a1f07d7c0f61b7fb5427985d0b27"} Mar 18 20:40:03 crc kubenswrapper[4950]: I0318 20:40:03.307999 4950 generic.go:334] "Generic (PLEG): container finished" podID="8149017b-bc82-441c-8ede-ec04b1ab0614" containerID="4c36c400239f08c0485c4f59a33cd5f31b2e05e90ada501a1492b3d74e3af080" exitCode=0 Mar 18 20:40:03 crc kubenswrapper[4950]: I0318 20:40:03.308124 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564440-jqvbk" event={"ID":"8149017b-bc82-441c-8ede-ec04b1ab0614","Type":"ContainerDied","Data":"4c36c400239f08c0485c4f59a33cd5f31b2e05e90ada501a1492b3d74e3af080"} Mar 18 20:40:04 crc kubenswrapper[4950]: I0318 20:40:04.970710 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564440-jqvbk" Mar 18 20:40:05 crc kubenswrapper[4950]: I0318 20:40:05.051532 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vk8d\" (UniqueName: \"kubernetes.io/projected/8149017b-bc82-441c-8ede-ec04b1ab0614-kube-api-access-7vk8d\") pod \"8149017b-bc82-441c-8ede-ec04b1ab0614\" (UID: \"8149017b-bc82-441c-8ede-ec04b1ab0614\") " Mar 18 20:40:05 crc kubenswrapper[4950]: I0318 20:40:05.090875 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8149017b-bc82-441c-8ede-ec04b1ab0614-kube-api-access-7vk8d" (OuterVolumeSpecName: "kube-api-access-7vk8d") pod "8149017b-bc82-441c-8ede-ec04b1ab0614" (UID: "8149017b-bc82-441c-8ede-ec04b1ab0614"). InnerVolumeSpecName "kube-api-access-7vk8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:40:05 crc kubenswrapper[4950]: I0318 20:40:05.153510 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vk8d\" (UniqueName: \"kubernetes.io/projected/8149017b-bc82-441c-8ede-ec04b1ab0614-kube-api-access-7vk8d\") on node \"crc\" DevicePath \"\"" Mar 18 20:40:05 crc kubenswrapper[4950]: I0318 20:40:05.334482 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564440-jqvbk" event={"ID":"8149017b-bc82-441c-8ede-ec04b1ab0614","Type":"ContainerDied","Data":"e20ace086cd6975dd1328d6e551de1294ef1a1f07d7c0f61b7fb5427985d0b27"} Mar 18 20:40:05 crc kubenswrapper[4950]: I0318 20:40:05.334541 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e20ace086cd6975dd1328d6e551de1294ef1a1f07d7c0f61b7fb5427985d0b27" Mar 18 20:40:05 crc kubenswrapper[4950]: I0318 20:40:05.334570 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564440-jqvbk" Mar 18 20:40:06 crc kubenswrapper[4950]: I0318 20:40:06.075126 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564434-ps98x"] Mar 18 20:40:06 crc kubenswrapper[4950]: I0318 20:40:06.091205 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564434-ps98x"] Mar 18 20:40:07 crc kubenswrapper[4950]: I0318 20:40:07.506610 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3599ae0-de48-40b8-979f-7f7125584a40" path="/var/lib/kubelet/pods/e3599ae0-de48-40b8-979f-7f7125584a40/volumes" Mar 18 20:40:12 crc kubenswrapper[4950]: I0318 20:40:12.480816 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:40:12 crc kubenswrapper[4950]: E0318 20:40:12.481813 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:40:27 crc kubenswrapper[4950]: I0318 20:40:27.486912 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:40:27 crc kubenswrapper[4950]: E0318 20:40:27.487637 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:40:42 crc kubenswrapper[4950]: I0318 20:40:42.479948 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:40:42 crc kubenswrapper[4950]: E0318 20:40:42.480942 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:40:48 crc kubenswrapper[4950]: I0318 20:40:48.848146 4950 scope.go:117] "RemoveContainer" containerID="3bb623dccda8a730d1aa964444c4ce4a4a8f3bbbae5d40443ac00cf6665496ca" Mar 18 20:40:54 crc kubenswrapper[4950]: I0318 20:40:54.480331 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:40:54 crc kubenswrapper[4950]: E0318 20:40:54.480946 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:41:05 crc kubenswrapper[4950]: I0318 20:41:05.479925 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:41:05 crc kubenswrapper[4950]: I0318 20:41:05.925803 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerStarted","Data":"0eb938fd8ae6f1d971cd0e6b1a12f847fe46b0c1dd5c1077e4c34d45028879ca"} Mar 18 20:42:00 crc kubenswrapper[4950]: I0318 20:42:00.159983 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564442-k9z45"] Mar 18 20:42:00 crc kubenswrapper[4950]: E0318 20:42:00.161177 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8149017b-bc82-441c-8ede-ec04b1ab0614" containerName="oc" Mar 18 20:42:00 crc kubenswrapper[4950]: I0318 20:42:00.161197 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="8149017b-bc82-441c-8ede-ec04b1ab0614" containerName="oc" Mar 18 20:42:00 crc kubenswrapper[4950]: I0318 20:42:00.161549 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="8149017b-bc82-441c-8ede-ec04b1ab0614" containerName="oc" Mar 18 20:42:00 crc kubenswrapper[4950]: I0318 20:42:00.162442 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564442-k9z45" Mar 18 20:42:00 crc kubenswrapper[4950]: I0318 20:42:00.175874 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564442-k9z45"] Mar 18 20:42:00 crc kubenswrapper[4950]: I0318 20:42:00.177179 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:42:00 crc kubenswrapper[4950]: I0318 20:42:00.177923 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:42:00 crc kubenswrapper[4950]: I0318 20:42:00.178819 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:42:00 crc kubenswrapper[4950]: I0318 20:42:00.263520 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9hvf\" (UniqueName: \"kubernetes.io/projected/fe68a5f4-7e8b-42a7-bcab-fbd81aead72e-kube-api-access-z9hvf\") pod \"auto-csr-approver-29564442-k9z45\" (UID: \"fe68a5f4-7e8b-42a7-bcab-fbd81aead72e\") " pod="openshift-infra/auto-csr-approver-29564442-k9z45" Mar 18 20:42:00 crc kubenswrapper[4950]: I0318 20:42:00.364890 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9hvf\" (UniqueName: \"kubernetes.io/projected/fe68a5f4-7e8b-42a7-bcab-fbd81aead72e-kube-api-access-z9hvf\") pod \"auto-csr-approver-29564442-k9z45\" (UID: \"fe68a5f4-7e8b-42a7-bcab-fbd81aead72e\") " pod="openshift-infra/auto-csr-approver-29564442-k9z45" Mar 18 20:42:00 crc kubenswrapper[4950]: I0318 20:42:00.389288 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9hvf\" (UniqueName: \"kubernetes.io/projected/fe68a5f4-7e8b-42a7-bcab-fbd81aead72e-kube-api-access-z9hvf\") pod \"auto-csr-approver-29564442-k9z45\" (UID: \"fe68a5f4-7e8b-42a7-bcab-fbd81aead72e\") " pod="openshift-infra/auto-csr-approver-29564442-k9z45" Mar 18 20:42:00 crc kubenswrapper[4950]: I0318 20:42:00.492623 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564442-k9z45" Mar 18 20:42:00 crc kubenswrapper[4950]: I0318 20:42:00.994267 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564442-k9z45"] Mar 18 20:42:01 crc kubenswrapper[4950]: I0318 20:42:01.451569 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564442-k9z45" event={"ID":"fe68a5f4-7e8b-42a7-bcab-fbd81aead72e","Type":"ContainerStarted","Data":"7ee8868e6b12c6141a1c7a9de344ccf226b76447f0be57650486956a4bcd93e5"} Mar 18 20:42:03 crc kubenswrapper[4950]: I0318 20:42:03.482971 4950 generic.go:334] "Generic (PLEG): container finished" podID="fe68a5f4-7e8b-42a7-bcab-fbd81aead72e" containerID="4476e98b688feef0f2d052155367dfc73a22f97ece6566f6363dc47cd0ebb8ec" exitCode=0 Mar 18 20:42:03 crc kubenswrapper[4950]: I0318 20:42:03.488279 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564442-k9z45" event={"ID":"fe68a5f4-7e8b-42a7-bcab-fbd81aead72e","Type":"ContainerDied","Data":"4476e98b688feef0f2d052155367dfc73a22f97ece6566f6363dc47cd0ebb8ec"} Mar 18 20:42:04 crc kubenswrapper[4950]: I0318 20:42:04.922980 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564442-k9z45" Mar 18 20:42:05 crc kubenswrapper[4950]: I0318 20:42:05.053856 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9hvf\" (UniqueName: \"kubernetes.io/projected/fe68a5f4-7e8b-42a7-bcab-fbd81aead72e-kube-api-access-z9hvf\") pod \"fe68a5f4-7e8b-42a7-bcab-fbd81aead72e\" (UID: \"fe68a5f4-7e8b-42a7-bcab-fbd81aead72e\") " Mar 18 20:42:05 crc kubenswrapper[4950]: I0318 20:42:05.073588 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe68a5f4-7e8b-42a7-bcab-fbd81aead72e-kube-api-access-z9hvf" (OuterVolumeSpecName: "kube-api-access-z9hvf") pod "fe68a5f4-7e8b-42a7-bcab-fbd81aead72e" (UID: "fe68a5f4-7e8b-42a7-bcab-fbd81aead72e"). InnerVolumeSpecName "kube-api-access-z9hvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:42:05 crc kubenswrapper[4950]: I0318 20:42:05.156103 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9hvf\" (UniqueName: \"kubernetes.io/projected/fe68a5f4-7e8b-42a7-bcab-fbd81aead72e-kube-api-access-z9hvf\") on node \"crc\" DevicePath \"\"" Mar 18 20:42:05 crc kubenswrapper[4950]: I0318 20:42:05.510694 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564442-k9z45" event={"ID":"fe68a5f4-7e8b-42a7-bcab-fbd81aead72e","Type":"ContainerDied","Data":"7ee8868e6b12c6141a1c7a9de344ccf226b76447f0be57650486956a4bcd93e5"} Mar 18 20:42:05 crc kubenswrapper[4950]: I0318 20:42:05.510741 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ee8868e6b12c6141a1c7a9de344ccf226b76447f0be57650486956a4bcd93e5" Mar 18 20:42:05 crc kubenswrapper[4950]: I0318 20:42:05.510808 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564442-k9z45" Mar 18 20:42:06 crc kubenswrapper[4950]: I0318 20:42:06.038475 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564436-bzgsz"] Mar 18 20:42:06 crc kubenswrapper[4950]: I0318 20:42:06.051615 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564436-bzgsz"] Mar 18 20:42:07 crc kubenswrapper[4950]: I0318 20:42:07.494597 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="720dd688-830b-4b1b-b73a-7f6f4938585f" path="/var/lib/kubelet/pods/720dd688-830b-4b1b-b73a-7f6f4938585f/volumes" Mar 18 20:42:43 crc kubenswrapper[4950]: E0318 20:42:43.087486 4950 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.163:39324->38.102.83.163:37995: write tcp 38.102.83.163:39324->38.102.83.163:37995: write: broken pipe Mar 18 20:42:48 crc kubenswrapper[4950]: I0318 20:42:48.940339 4950 scope.go:117] "RemoveContainer" containerID="c37cea83b16e5d03396ded63bd4f219f34c138a56c3186ed38c292261c5c63c8" Mar 18 20:43:12 crc kubenswrapper[4950]: I0318 20:43:12.913547 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj"] Mar 18 20:43:12 crc kubenswrapper[4950]: I0318 20:43:12.922845 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh"] Mar 18 20:43:12 crc kubenswrapper[4950]: I0318 20:43:12.934285 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r"] Mar 18 20:43:12 crc kubenswrapper[4950]: I0318 20:43:12.950211 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv"] Mar 18 20:43:12 crc kubenswrapper[4950]: I0318 20:43:12.963560 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2vct7"] Mar 18 20:43:12 crc kubenswrapper[4950]: I0318 20:43:12.975595 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6"] Mar 18 20:43:12 crc kubenswrapper[4950]: I0318 20:43:12.986781 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-b6nlh"] Mar 18 20:43:13 crc kubenswrapper[4950]: I0318 20:43:13.000017 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2vct7"] Mar 18 20:43:13 crc kubenswrapper[4950]: I0318 20:43:13.009326 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6gml"] Mar 18 20:43:13 crc kubenswrapper[4950]: I0318 20:43:13.016280 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9xb2r"] Mar 18 20:43:13 crc kubenswrapper[4950]: I0318 20:43:13.023181 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-xhfbh"] Mar 18 20:43:13 crc kubenswrapper[4950]: I0318 20:43:13.032054 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-bm9mv"] Mar 18 20:43:13 crc kubenswrapper[4950]: I0318 20:43:13.038847 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ww4zp"] Mar 18 20:43:13 crc kubenswrapper[4950]: I0318 20:43:13.044706 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9q6zj"] Mar 18 20:43:13 crc kubenswrapper[4950]: I0318 20:43:13.050039 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-xhfbh"] Mar 18 20:43:13 crc kubenswrapper[4950]: I0318 20:43:13.056762 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x6gml"] Mar 18 20:43:13 crc kubenswrapper[4950]: I0318 20:43:13.064406 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-z79h6"] Mar 18 20:43:13 crc kubenswrapper[4950]: I0318 20:43:13.071554 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ww4zp"] Mar 18 20:43:13 crc kubenswrapper[4950]: I0318 20:43:13.077159 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vngd5"] Mar 18 20:43:13 crc kubenswrapper[4950]: I0318 20:43:13.082258 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vngd5"] Mar 18 20:43:13 crc kubenswrapper[4950]: I0318 20:43:13.491795 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2309d051-312e-45c6-8c31-28362aef9908" path="/var/lib/kubelet/pods/2309d051-312e-45c6-8c31-28362aef9908/volumes" Mar 18 20:43:13 crc kubenswrapper[4950]: I0318 20:43:13.492971 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5162a83c-d820-45d0-86c7-b4cb8a22a607" path="/var/lib/kubelet/pods/5162a83c-d820-45d0-86c7-b4cb8a22a607/volumes" Mar 18 20:43:13 crc kubenswrapper[4950]: I0318 20:43:13.494070 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5495b98f-7c92-4918-b4f7-d1e7e06dd465" path="/var/lib/kubelet/pods/5495b98f-7c92-4918-b4f7-d1e7e06dd465/volumes" Mar 18 20:43:13 crc kubenswrapper[4950]: I0318 20:43:13.495182 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="651a2596-f562-4012-8f2a-7a3e1c562a0a" path="/var/lib/kubelet/pods/651a2596-f562-4012-8f2a-7a3e1c562a0a/volumes" Mar 18 20:43:13 crc kubenswrapper[4950]: I0318 20:43:13.497174 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0" path="/var/lib/kubelet/pods/7e103e2b-1ce3-4fa7-9ee4-aaa3eb5f87b0/volumes" Mar 18 20:43:13 crc kubenswrapper[4950]: I0318 20:43:13.498312 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84bce393-6964-42d4-9126-076287a2dece" path="/var/lib/kubelet/pods/84bce393-6964-42d4-9126-076287a2dece/volumes" Mar 18 20:43:13 crc kubenswrapper[4950]: I0318 20:43:13.499674 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b1fb431-7ff9-498a-a4b3-352822335933" path="/var/lib/kubelet/pods/8b1fb431-7ff9-498a-a4b3-352822335933/volumes" Mar 18 20:43:13 crc kubenswrapper[4950]: I0318 20:43:13.501969 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb" path="/var/lib/kubelet/pods/907a6ea9-e0ab-46d7-a2ca-c92f98ca0adb/volumes" Mar 18 20:43:13 crc kubenswrapper[4950]: I0318 20:43:13.503190 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1b58e52-1dea-4216-9eec-ebf5c09fba32" path="/var/lib/kubelet/pods/a1b58e52-1dea-4216-9eec-ebf5c09fba32/volumes" Mar 18 20:43:13 crc kubenswrapper[4950]: I0318 20:43:13.504441 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a75b3257-7409-45c9-acaa-d85cec7f9015" path="/var/lib/kubelet/pods/a75b3257-7409-45c9-acaa-d85cec7f9015/volumes" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.684373 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq"] Mar 18 20:43:26 crc kubenswrapper[4950]: E0318 20:43:26.685310 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe68a5f4-7e8b-42a7-bcab-fbd81aead72e" containerName="oc" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.685325 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe68a5f4-7e8b-42a7-bcab-fbd81aead72e" containerName="oc" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.685539 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe68a5f4-7e8b-42a7-bcab-fbd81aead72e" containerName="oc" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.686125 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.687953 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.687959 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.687959 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.689388 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.694671 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.699977 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq"] Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.777270 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq\" (UID: \"11355e20-3f52-4010-8c35-0318c88ad131\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.777350 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq\" (UID: \"11355e20-3f52-4010-8c35-0318c88ad131\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.777483 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq\" (UID: \"11355e20-3f52-4010-8c35-0318c88ad131\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.777573 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq\" (UID: \"11355e20-3f52-4010-8c35-0318c88ad131\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.777698 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fshfp\" (UniqueName: \"kubernetes.io/projected/11355e20-3f52-4010-8c35-0318c88ad131-kube-api-access-fshfp\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq\" (UID: \"11355e20-3f52-4010-8c35-0318c88ad131\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.879890 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq\" (UID: \"11355e20-3f52-4010-8c35-0318c88ad131\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.880020 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq\" (UID: \"11355e20-3f52-4010-8c35-0318c88ad131\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.880060 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fshfp\" (UniqueName: \"kubernetes.io/projected/11355e20-3f52-4010-8c35-0318c88ad131-kube-api-access-fshfp\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq\" (UID: \"11355e20-3f52-4010-8c35-0318c88ad131\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.880121 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq\" (UID: \"11355e20-3f52-4010-8c35-0318c88ad131\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.880172 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq\" (UID: \"11355e20-3f52-4010-8c35-0318c88ad131\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.887047 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq\" (UID: \"11355e20-3f52-4010-8c35-0318c88ad131\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.887469 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq\" (UID: \"11355e20-3f52-4010-8c35-0318c88ad131\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.890483 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq\" (UID: \"11355e20-3f52-4010-8c35-0318c88ad131\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.890660 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq\" (UID: \"11355e20-3f52-4010-8c35-0318c88ad131\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" Mar 18 20:43:26 crc kubenswrapper[4950]: I0318 20:43:26.901133 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fshfp\" (UniqueName: \"kubernetes.io/projected/11355e20-3f52-4010-8c35-0318c88ad131-kube-api-access-fshfp\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq\" (UID: \"11355e20-3f52-4010-8c35-0318c88ad131\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" Mar 18 20:43:27 crc kubenswrapper[4950]: I0318 20:43:27.005053 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" Mar 18 20:43:27 crc kubenswrapper[4950]: I0318 20:43:27.604654 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq"] Mar 18 20:43:27 crc kubenswrapper[4950]: W0318 20:43:27.616157 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11355e20_3f52_4010_8c35_0318c88ad131.slice/crio-f8a4477646ed89d148021ee7c0564a7f6260f2e7f7e2ec8e00cffac6aa689f64 WatchSource:0}: Error finding container f8a4477646ed89d148021ee7c0564a7f6260f2e7f7e2ec8e00cffac6aa689f64: Status 404 returned error can't find the container with id f8a4477646ed89d148021ee7c0564a7f6260f2e7f7e2ec8e00cffac6aa689f64 Mar 18 20:43:27 crc kubenswrapper[4950]: I0318 20:43:27.620056 4950 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 20:43:28 crc kubenswrapper[4950]: I0318 20:43:28.167316 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:43:28 crc kubenswrapper[4950]: I0318 20:43:28.206836 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" event={"ID":"11355e20-3f52-4010-8c35-0318c88ad131","Type":"ContainerStarted","Data":"f8a4477646ed89d148021ee7c0564a7f6260f2e7f7e2ec8e00cffac6aa689f64"} Mar 18 20:43:29 crc kubenswrapper[4950]: I0318 20:43:29.225011 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" event={"ID":"11355e20-3f52-4010-8c35-0318c88ad131","Type":"ContainerStarted","Data":"4a069588b0d33374ef9224127932d00f939ff2b273f5d40a03963fdcfbdc9104"} Mar 18 20:43:29 crc kubenswrapper[4950]: I0318 20:43:29.259870 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" podStartSLOduration=2.715081422 podStartE2EDuration="3.259851743s" podCreationTimestamp="2026-03-18 20:43:26 +0000 UTC" firstStartedPulling="2026-03-18 20:43:27.619787171 +0000 UTC m=+2220.860629049" lastFinishedPulling="2026-03-18 20:43:28.164557502 +0000 UTC m=+2221.405399370" observedRunningTime="2026-03-18 20:43:29.253462536 +0000 UTC m=+2222.494304424" watchObservedRunningTime="2026-03-18 20:43:29.259851743 +0000 UTC m=+2222.500693621" Mar 18 20:43:33 crc kubenswrapper[4950]: I0318 20:43:33.836636 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:43:33 crc kubenswrapper[4950]: I0318 20:43:33.837371 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:43:39 crc kubenswrapper[4950]: I0318 20:43:39.314725 4950 generic.go:334] "Generic (PLEG): container finished" podID="11355e20-3f52-4010-8c35-0318c88ad131" containerID="4a069588b0d33374ef9224127932d00f939ff2b273f5d40a03963fdcfbdc9104" exitCode=0 Mar 18 20:43:39 crc kubenswrapper[4950]: I0318 20:43:39.314819 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" event={"ID":"11355e20-3f52-4010-8c35-0318c88ad131","Type":"ContainerDied","Data":"4a069588b0d33374ef9224127932d00f939ff2b273f5d40a03963fdcfbdc9104"} Mar 18 20:43:40 crc kubenswrapper[4950]: I0318 20:43:40.749355 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" Mar 18 20:43:40 crc kubenswrapper[4950]: I0318 20:43:40.844362 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-ceph\") pod \"11355e20-3f52-4010-8c35-0318c88ad131\" (UID: \"11355e20-3f52-4010-8c35-0318c88ad131\") " Mar 18 20:43:40 crc kubenswrapper[4950]: I0318 20:43:40.844454 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-ssh-key-openstack-edpm-ipam\") pod \"11355e20-3f52-4010-8c35-0318c88ad131\" (UID: \"11355e20-3f52-4010-8c35-0318c88ad131\") " Mar 18 20:43:40 crc kubenswrapper[4950]: I0318 20:43:40.844480 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-repo-setup-combined-ca-bundle\") pod \"11355e20-3f52-4010-8c35-0318c88ad131\" (UID: \"11355e20-3f52-4010-8c35-0318c88ad131\") " Mar 18 20:43:40 crc kubenswrapper[4950]: I0318 20:43:40.844525 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fshfp\" (UniqueName: \"kubernetes.io/projected/11355e20-3f52-4010-8c35-0318c88ad131-kube-api-access-fshfp\") pod \"11355e20-3f52-4010-8c35-0318c88ad131\" (UID: \"11355e20-3f52-4010-8c35-0318c88ad131\") " Mar 18 20:43:40 crc kubenswrapper[4950]: I0318 20:43:40.844564 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-inventory\") pod \"11355e20-3f52-4010-8c35-0318c88ad131\" (UID: \"11355e20-3f52-4010-8c35-0318c88ad131\") " Mar 18 20:43:40 crc kubenswrapper[4950]: I0318 20:43:40.861593 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-ceph" (OuterVolumeSpecName: "ceph") pod "11355e20-3f52-4010-8c35-0318c88ad131" (UID: "11355e20-3f52-4010-8c35-0318c88ad131"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:43:40 crc kubenswrapper[4950]: I0318 20:43:40.861863 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "11355e20-3f52-4010-8c35-0318c88ad131" (UID: "11355e20-3f52-4010-8c35-0318c88ad131"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:43:40 crc kubenswrapper[4950]: I0318 20:43:40.866150 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11355e20-3f52-4010-8c35-0318c88ad131-kube-api-access-fshfp" (OuterVolumeSpecName: "kube-api-access-fshfp") pod "11355e20-3f52-4010-8c35-0318c88ad131" (UID: "11355e20-3f52-4010-8c35-0318c88ad131"). InnerVolumeSpecName "kube-api-access-fshfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:43:40 crc kubenswrapper[4950]: I0318 20:43:40.877555 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-inventory" (OuterVolumeSpecName: "inventory") pod "11355e20-3f52-4010-8c35-0318c88ad131" (UID: "11355e20-3f52-4010-8c35-0318c88ad131"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:43:40 crc kubenswrapper[4950]: I0318 20:43:40.879806 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "11355e20-3f52-4010-8c35-0318c88ad131" (UID: "11355e20-3f52-4010-8c35-0318c88ad131"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:43:40 crc kubenswrapper[4950]: I0318 20:43:40.946837 4950 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-ceph\") on node \"crc\" DevicePath \"\"" Mar 18 20:43:40 crc kubenswrapper[4950]: I0318 20:43:40.946874 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:43:40 crc kubenswrapper[4950]: I0318 20:43:40.946892 4950 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:43:40 crc kubenswrapper[4950]: I0318 20:43:40.946903 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fshfp\" (UniqueName: \"kubernetes.io/projected/11355e20-3f52-4010-8c35-0318c88ad131-kube-api-access-fshfp\") on node \"crc\" DevicePath \"\"" Mar 18 20:43:40 crc kubenswrapper[4950]: I0318 20:43:40.946916 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/11355e20-3f52-4010-8c35-0318c88ad131-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.341285 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" event={"ID":"11355e20-3f52-4010-8c35-0318c88ad131","Type":"ContainerDied","Data":"f8a4477646ed89d148021ee7c0564a7f6260f2e7f7e2ec8e00cffac6aa689f64"} Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.341335 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8a4477646ed89d148021ee7c0564a7f6260f2e7f7e2ec8e00cffac6aa689f64" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.341338 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.416397 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf"] Mar 18 20:43:41 crc kubenswrapper[4950]: E0318 20:43:41.416821 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11355e20-3f52-4010-8c35-0318c88ad131" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.416835 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="11355e20-3f52-4010-8c35-0318c88ad131" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.422625 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="11355e20-3f52-4010-8c35-0318c88ad131" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.423887 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.426728 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.426996 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.427076 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.427369 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.435284 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.460496 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf\" (UID: \"8230645b-e426-4c0e-a28b-565ba98d18a1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.460536 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf\" (UID: \"8230645b-e426-4c0e-a28b-565ba98d18a1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.460599 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf\" (UID: \"8230645b-e426-4c0e-a28b-565ba98d18a1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.460657 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf\" (UID: \"8230645b-e426-4c0e-a28b-565ba98d18a1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.460680 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdjkr\" (UniqueName: \"kubernetes.io/projected/8230645b-e426-4c0e-a28b-565ba98d18a1-kube-api-access-sdjkr\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf\" (UID: \"8230645b-e426-4c0e-a28b-565ba98d18a1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.467244 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf"] Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.563840 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf\" (UID: \"8230645b-e426-4c0e-a28b-565ba98d18a1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.564521 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf\" (UID: \"8230645b-e426-4c0e-a28b-565ba98d18a1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.564761 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf\" (UID: \"8230645b-e426-4c0e-a28b-565ba98d18a1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.564880 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf\" (UID: \"8230645b-e426-4c0e-a28b-565ba98d18a1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.564935 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdjkr\" (UniqueName: \"kubernetes.io/projected/8230645b-e426-4c0e-a28b-565ba98d18a1-kube-api-access-sdjkr\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf\" (UID: \"8230645b-e426-4c0e-a28b-565ba98d18a1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.569391 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf\" (UID: \"8230645b-e426-4c0e-a28b-565ba98d18a1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.570657 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf\" (UID: \"8230645b-e426-4c0e-a28b-565ba98d18a1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.574733 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf\" (UID: \"8230645b-e426-4c0e-a28b-565ba98d18a1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.577977 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf\" (UID: \"8230645b-e426-4c0e-a28b-565ba98d18a1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.586079 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdjkr\" (UniqueName: \"kubernetes.io/projected/8230645b-e426-4c0e-a28b-565ba98d18a1-kube-api-access-sdjkr\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf\" (UID: \"8230645b-e426-4c0e-a28b-565ba98d18a1\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" Mar 18 20:43:41 crc kubenswrapper[4950]: I0318 20:43:41.762368 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" Mar 18 20:43:42 crc kubenswrapper[4950]: I0318 20:43:42.290630 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf"] Mar 18 20:43:42 crc kubenswrapper[4950]: I0318 20:43:42.349476 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" event={"ID":"8230645b-e426-4c0e-a28b-565ba98d18a1","Type":"ContainerStarted","Data":"5d6f999c3707f06a8af2371ead063e1219e2953d72f2dbfda2236a1ff8637bc1"} Mar 18 20:43:43 crc kubenswrapper[4950]: I0318 20:43:43.364144 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" event={"ID":"8230645b-e426-4c0e-a28b-565ba98d18a1","Type":"ContainerStarted","Data":"55881ed90e1103f82e4b8dd127583137efe45648c960f1fba288c8cec343e9d4"} Mar 18 20:43:43 crc kubenswrapper[4950]: I0318 20:43:43.388044 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" podStartSLOduration=1.939962116 podStartE2EDuration="2.388026247s" podCreationTimestamp="2026-03-18 20:43:41 +0000 UTC" firstStartedPulling="2026-03-18 20:43:42.294536312 +0000 UTC m=+2235.535378180" lastFinishedPulling="2026-03-18 20:43:42.742600443 +0000 UTC m=+2235.983442311" observedRunningTime="2026-03-18 20:43:43.380707645 +0000 UTC m=+2236.621549523" watchObservedRunningTime="2026-03-18 20:43:43.388026247 +0000 UTC m=+2236.628868115" Mar 18 20:43:49 crc kubenswrapper[4950]: I0318 20:43:49.056966 4950 scope.go:117] "RemoveContainer" containerID="d63d38a73140c397d468f5213d693b05c42ef6f917d197623927e165b2ab0e18" Mar 18 20:43:49 crc kubenswrapper[4950]: I0318 20:43:49.105390 4950 scope.go:117] "RemoveContainer" containerID="177a725128594a13952a18229398d0c915fcbbb80992adf5d9a78770c3435cb3" Mar 18 20:43:49 crc kubenswrapper[4950]: I0318 20:43:49.151965 4950 scope.go:117] "RemoveContainer" containerID="5ecdc9e1d79118ed35463ec1bd847a70d5c6f6f6aa7cffa37d8cc10e9554ba4d" Mar 18 20:43:49 crc kubenswrapper[4950]: I0318 20:43:49.237216 4950 scope.go:117] "RemoveContainer" containerID="8a2b008f71fb3e166d73cdac64b2766db32e331a36fa319f713b5a43296f1434" Mar 18 20:43:49 crc kubenswrapper[4950]: I0318 20:43:49.270385 4950 scope.go:117] "RemoveContainer" containerID="68f707f726c227fe94bff58052b50aead3ac2b367f00a474c05bb6f61be620d7" Mar 18 20:43:49 crc kubenswrapper[4950]: I0318 20:43:49.362329 4950 scope.go:117] "RemoveContainer" containerID="57a9ca3c7f01904503552eba81817b6563a3fb9ee3278c21bf505930c5ba0aef" Mar 18 20:43:49 crc kubenswrapper[4950]: I0318 20:43:49.395256 4950 scope.go:117] "RemoveContainer" containerID="a4eacba1d449d76b272bfeca9562771e6505952fbfd2fa9a0ab115c1c228b6f8" Mar 18 20:43:49 crc kubenswrapper[4950]: I0318 20:43:49.459906 4950 scope.go:117] "RemoveContainer" containerID="74242ce84c406b044ae39cf55d1dea785c786d4e82138ad9b68e2eedf3cc8061" Mar 18 20:43:49 crc kubenswrapper[4950]: I0318 20:43:49.547717 4950 scope.go:117] "RemoveContainer" containerID="fce0212fcd35805d80c1c3dd6c623a0546a5e213b1ce61fa4658144f9b749eaa" Mar 18 20:43:49 crc kubenswrapper[4950]: I0318 20:43:49.582282 4950 scope.go:117] "RemoveContainer" containerID="768b95b44dadfe1e523ff513f3b2412fb537fab5f4e51653e0a1847d11c92a49" Mar 18 20:43:59 crc kubenswrapper[4950]: I0318 20:43:59.017014 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dbksg"] Mar 18 20:43:59 crc kubenswrapper[4950]: I0318 20:43:59.020188 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbksg" Mar 18 20:43:59 crc kubenswrapper[4950]: I0318 20:43:59.032547 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dbksg"] Mar 18 20:43:59 crc kubenswrapper[4950]: I0318 20:43:59.164062 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hglbc\" (UniqueName: \"kubernetes.io/projected/3de14edf-8d87-4da6-96bf-62b014bf3c35-kube-api-access-hglbc\") pod \"certified-operators-dbksg\" (UID: \"3de14edf-8d87-4da6-96bf-62b014bf3c35\") " pod="openshift-marketplace/certified-operators-dbksg" Mar 18 20:43:59 crc kubenswrapper[4950]: I0318 20:43:59.164449 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3de14edf-8d87-4da6-96bf-62b014bf3c35-utilities\") pod \"certified-operators-dbksg\" (UID: \"3de14edf-8d87-4da6-96bf-62b014bf3c35\") " pod="openshift-marketplace/certified-operators-dbksg" Mar 18 20:43:59 crc kubenswrapper[4950]: I0318 20:43:59.164479 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3de14edf-8d87-4da6-96bf-62b014bf3c35-catalog-content\") pod \"certified-operators-dbksg\" (UID: \"3de14edf-8d87-4da6-96bf-62b014bf3c35\") " pod="openshift-marketplace/certified-operators-dbksg" Mar 18 20:43:59 crc kubenswrapper[4950]: I0318 20:43:59.266184 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hglbc\" (UniqueName: \"kubernetes.io/projected/3de14edf-8d87-4da6-96bf-62b014bf3c35-kube-api-access-hglbc\") pod \"certified-operators-dbksg\" (UID: \"3de14edf-8d87-4da6-96bf-62b014bf3c35\") " pod="openshift-marketplace/certified-operators-dbksg" Mar 18 20:43:59 crc kubenswrapper[4950]: I0318 20:43:59.266256 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3de14edf-8d87-4da6-96bf-62b014bf3c35-utilities\") pod \"certified-operators-dbksg\" (UID: \"3de14edf-8d87-4da6-96bf-62b014bf3c35\") " pod="openshift-marketplace/certified-operators-dbksg" Mar 18 20:43:59 crc kubenswrapper[4950]: I0318 20:43:59.266297 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3de14edf-8d87-4da6-96bf-62b014bf3c35-catalog-content\") pod \"certified-operators-dbksg\" (UID: \"3de14edf-8d87-4da6-96bf-62b014bf3c35\") " pod="openshift-marketplace/certified-operators-dbksg" Mar 18 20:43:59 crc kubenswrapper[4950]: I0318 20:43:59.266819 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3de14edf-8d87-4da6-96bf-62b014bf3c35-utilities\") pod \"certified-operators-dbksg\" (UID: \"3de14edf-8d87-4da6-96bf-62b014bf3c35\") " pod="openshift-marketplace/certified-operators-dbksg" Mar 18 20:43:59 crc kubenswrapper[4950]: I0318 20:43:59.266859 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3de14edf-8d87-4da6-96bf-62b014bf3c35-catalog-content\") pod \"certified-operators-dbksg\" (UID: \"3de14edf-8d87-4da6-96bf-62b014bf3c35\") " pod="openshift-marketplace/certified-operators-dbksg" Mar 18 20:43:59 crc kubenswrapper[4950]: I0318 20:43:59.303553 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hglbc\" (UniqueName: \"kubernetes.io/projected/3de14edf-8d87-4da6-96bf-62b014bf3c35-kube-api-access-hglbc\") pod \"certified-operators-dbksg\" (UID: \"3de14edf-8d87-4da6-96bf-62b014bf3c35\") " pod="openshift-marketplace/certified-operators-dbksg" Mar 18 20:43:59 crc kubenswrapper[4950]: I0318 20:43:59.342638 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbksg" Mar 18 20:43:59 crc kubenswrapper[4950]: I0318 20:43:59.958888 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dbksg"] Mar 18 20:44:00 crc kubenswrapper[4950]: I0318 20:44:00.146879 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564444-7vvp5"] Mar 18 20:44:00 crc kubenswrapper[4950]: I0318 20:44:00.149382 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564444-7vvp5" Mar 18 20:44:00 crc kubenswrapper[4950]: I0318 20:44:00.154127 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:44:00 crc kubenswrapper[4950]: I0318 20:44:00.154755 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:44:00 crc kubenswrapper[4950]: I0318 20:44:00.154943 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:44:00 crc kubenswrapper[4950]: I0318 20:44:00.161703 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564444-7vvp5"] Mar 18 20:44:00 crc kubenswrapper[4950]: I0318 20:44:00.294676 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgmxl\" (UniqueName: \"kubernetes.io/projected/eb7b1442-7b35-43f0-aa0c-84c5a8e6d2c9-kube-api-access-tgmxl\") pod \"auto-csr-approver-29564444-7vvp5\" (UID: \"eb7b1442-7b35-43f0-aa0c-84c5a8e6d2c9\") " pod="openshift-infra/auto-csr-approver-29564444-7vvp5" Mar 18 20:44:00 crc kubenswrapper[4950]: I0318 20:44:00.395979 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgmxl\" (UniqueName: \"kubernetes.io/projected/eb7b1442-7b35-43f0-aa0c-84c5a8e6d2c9-kube-api-access-tgmxl\") pod \"auto-csr-approver-29564444-7vvp5\" (UID: \"eb7b1442-7b35-43f0-aa0c-84c5a8e6d2c9\") " pod="openshift-infra/auto-csr-approver-29564444-7vvp5" Mar 18 20:44:00 crc kubenswrapper[4950]: I0318 20:44:00.427688 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgmxl\" (UniqueName: \"kubernetes.io/projected/eb7b1442-7b35-43f0-aa0c-84c5a8e6d2c9-kube-api-access-tgmxl\") pod \"auto-csr-approver-29564444-7vvp5\" (UID: \"eb7b1442-7b35-43f0-aa0c-84c5a8e6d2c9\") " pod="openshift-infra/auto-csr-approver-29564444-7vvp5" Mar 18 20:44:00 crc kubenswrapper[4950]: I0318 20:44:00.512115 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564444-7vvp5" Mar 18 20:44:00 crc kubenswrapper[4950]: I0318 20:44:00.529711 4950 generic.go:334] "Generic (PLEG): container finished" podID="3de14edf-8d87-4da6-96bf-62b014bf3c35" containerID="f8b5d98d08deaa7ca43e48abcf2274e7f3f86ab416044cb42ab8c3bba4920f14" exitCode=0 Mar 18 20:44:00 crc kubenswrapper[4950]: I0318 20:44:00.529754 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbksg" event={"ID":"3de14edf-8d87-4da6-96bf-62b014bf3c35","Type":"ContainerDied","Data":"f8b5d98d08deaa7ca43e48abcf2274e7f3f86ab416044cb42ab8c3bba4920f14"} Mar 18 20:44:00 crc kubenswrapper[4950]: I0318 20:44:00.529777 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbksg" event={"ID":"3de14edf-8d87-4da6-96bf-62b014bf3c35","Type":"ContainerStarted","Data":"3ecd3c24b2ff2e996f2e24fccd65bfa94518c719840f80758df71f43cbbd281c"} Mar 18 20:44:01 crc kubenswrapper[4950]: I0318 20:44:01.056676 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564444-7vvp5"] Mar 18 20:44:01 crc kubenswrapper[4950]: I0318 20:44:01.539510 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564444-7vvp5" event={"ID":"eb7b1442-7b35-43f0-aa0c-84c5a8e6d2c9","Type":"ContainerStarted","Data":"e76c92436d5862de0f93e5dc074b684f2ccd4ec77af5c8feac930e21a0a00daa"} Mar 18 20:44:01 crc kubenswrapper[4950]: I0318 20:44:01.544037 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbksg" event={"ID":"3de14edf-8d87-4da6-96bf-62b014bf3c35","Type":"ContainerStarted","Data":"85d775d68fa91f9c40066554de5c93093a6cb4decf5faafa56f499840e25ecc8"} Mar 18 20:44:02 crc kubenswrapper[4950]: I0318 20:44:02.585060 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564444-7vvp5" event={"ID":"eb7b1442-7b35-43f0-aa0c-84c5a8e6d2c9","Type":"ContainerStarted","Data":"feb84a2195d48fdbead49f67c0f75b4e48bb65fe9e84653976a79583b17cbcc3"} Mar 18 20:44:02 crc kubenswrapper[4950]: I0318 20:44:02.604549 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29564444-7vvp5" podStartSLOduration=1.424895837 podStartE2EDuration="2.604527914s" podCreationTimestamp="2026-03-18 20:44:00 +0000 UTC" firstStartedPulling="2026-03-18 20:44:01.066109382 +0000 UTC m=+2254.306951270" lastFinishedPulling="2026-03-18 20:44:02.245741479 +0000 UTC m=+2255.486583347" observedRunningTime="2026-03-18 20:44:02.602331807 +0000 UTC m=+2255.843173695" watchObservedRunningTime="2026-03-18 20:44:02.604527914 +0000 UTC m=+2255.845369792" Mar 18 20:44:03 crc kubenswrapper[4950]: I0318 20:44:03.596455 4950 generic.go:334] "Generic (PLEG): container finished" podID="3de14edf-8d87-4da6-96bf-62b014bf3c35" containerID="85d775d68fa91f9c40066554de5c93093a6cb4decf5faafa56f499840e25ecc8" exitCode=0 Mar 18 20:44:03 crc kubenswrapper[4950]: I0318 20:44:03.596543 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbksg" event={"ID":"3de14edf-8d87-4da6-96bf-62b014bf3c35","Type":"ContainerDied","Data":"85d775d68fa91f9c40066554de5c93093a6cb4decf5faafa56f499840e25ecc8"} Mar 18 20:44:03 crc kubenswrapper[4950]: I0318 20:44:03.599087 4950 generic.go:334] "Generic (PLEG): container finished" podID="eb7b1442-7b35-43f0-aa0c-84c5a8e6d2c9" containerID="feb84a2195d48fdbead49f67c0f75b4e48bb65fe9e84653976a79583b17cbcc3" exitCode=0 Mar 18 20:44:03 crc kubenswrapper[4950]: I0318 20:44:03.599123 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564444-7vvp5" event={"ID":"eb7b1442-7b35-43f0-aa0c-84c5a8e6d2c9","Type":"ContainerDied","Data":"feb84a2195d48fdbead49f67c0f75b4e48bb65fe9e84653976a79583b17cbcc3"} Mar 18 20:44:03 crc kubenswrapper[4950]: I0318 20:44:03.836201 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:44:03 crc kubenswrapper[4950]: I0318 20:44:03.836757 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:44:04 crc kubenswrapper[4950]: I0318 20:44:04.608903 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbksg" event={"ID":"3de14edf-8d87-4da6-96bf-62b014bf3c35","Type":"ContainerStarted","Data":"6ab26df1a48dec1e448316b39644384bc4f9cbfb1d735c8ea19dd4c0157b9fff"} Mar 18 20:44:04 crc kubenswrapper[4950]: I0318 20:44:04.635892 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dbksg" podStartSLOduration=2.962236193 podStartE2EDuration="6.635870504s" podCreationTimestamp="2026-03-18 20:43:58 +0000 UTC" firstStartedPulling="2026-03-18 20:44:00.531342394 +0000 UTC m=+2253.772184262" lastFinishedPulling="2026-03-18 20:44:04.204976705 +0000 UTC m=+2257.445818573" observedRunningTime="2026-03-18 20:44:04.631599541 +0000 UTC m=+2257.872441409" watchObservedRunningTime="2026-03-18 20:44:04.635870504 +0000 UTC m=+2257.876712382" Mar 18 20:44:04 crc kubenswrapper[4950]: I0318 20:44:04.971219 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564444-7vvp5" Mar 18 20:44:04 crc kubenswrapper[4950]: I0318 20:44:04.981822 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgmxl\" (UniqueName: \"kubernetes.io/projected/eb7b1442-7b35-43f0-aa0c-84c5a8e6d2c9-kube-api-access-tgmxl\") pod \"eb7b1442-7b35-43f0-aa0c-84c5a8e6d2c9\" (UID: \"eb7b1442-7b35-43f0-aa0c-84c5a8e6d2c9\") " Mar 18 20:44:04 crc kubenswrapper[4950]: I0318 20:44:04.987135 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb7b1442-7b35-43f0-aa0c-84c5a8e6d2c9-kube-api-access-tgmxl" (OuterVolumeSpecName: "kube-api-access-tgmxl") pod "eb7b1442-7b35-43f0-aa0c-84c5a8e6d2c9" (UID: "eb7b1442-7b35-43f0-aa0c-84c5a8e6d2c9"). InnerVolumeSpecName "kube-api-access-tgmxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:44:05 crc kubenswrapper[4950]: I0318 20:44:05.083593 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgmxl\" (UniqueName: \"kubernetes.io/projected/eb7b1442-7b35-43f0-aa0c-84c5a8e6d2c9-kube-api-access-tgmxl\") on node \"crc\" DevicePath \"\"" Mar 18 20:44:05 crc kubenswrapper[4950]: I0318 20:44:05.620040 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564444-7vvp5" event={"ID":"eb7b1442-7b35-43f0-aa0c-84c5a8e6d2c9","Type":"ContainerDied","Data":"e76c92436d5862de0f93e5dc074b684f2ccd4ec77af5c8feac930e21a0a00daa"} Mar 18 20:44:05 crc kubenswrapper[4950]: I0318 20:44:05.620400 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e76c92436d5862de0f93e5dc074b684f2ccd4ec77af5c8feac930e21a0a00daa" Mar 18 20:44:05 crc kubenswrapper[4950]: I0318 20:44:05.620121 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564444-7vvp5" Mar 18 20:44:05 crc kubenswrapper[4950]: I0318 20:44:05.679663 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564438-6lsrq"] Mar 18 20:44:05 crc kubenswrapper[4950]: I0318 20:44:05.685463 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564438-6lsrq"] Mar 18 20:44:07 crc kubenswrapper[4950]: I0318 20:44:07.494858 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c77a603-7308-4467-a365-a47380f4954f" path="/var/lib/kubelet/pods/2c77a603-7308-4467-a365-a47380f4954f/volumes" Mar 18 20:44:07 crc kubenswrapper[4950]: I0318 20:44:07.987043 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jvg9t"] Mar 18 20:44:07 crc kubenswrapper[4950]: E0318 20:44:07.987543 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb7b1442-7b35-43f0-aa0c-84c5a8e6d2c9" containerName="oc" Mar 18 20:44:07 crc kubenswrapper[4950]: I0318 20:44:07.987563 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb7b1442-7b35-43f0-aa0c-84c5a8e6d2c9" containerName="oc" Mar 18 20:44:07 crc kubenswrapper[4950]: I0318 20:44:07.987744 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb7b1442-7b35-43f0-aa0c-84c5a8e6d2c9" containerName="oc" Mar 18 20:44:07 crc kubenswrapper[4950]: I0318 20:44:07.989237 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jvg9t" Mar 18 20:44:08 crc kubenswrapper[4950]: I0318 20:44:08.012002 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jvg9t"] Mar 18 20:44:08 crc kubenswrapper[4950]: I0318 20:44:08.135066 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxl2l\" (UniqueName: \"kubernetes.io/projected/030ed361-be9d-4dd0-88f3-f24f71716916-kube-api-access-zxl2l\") pod \"redhat-marketplace-jvg9t\" (UID: \"030ed361-be9d-4dd0-88f3-f24f71716916\") " pod="openshift-marketplace/redhat-marketplace-jvg9t" Mar 18 20:44:08 crc kubenswrapper[4950]: I0318 20:44:08.135121 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/030ed361-be9d-4dd0-88f3-f24f71716916-utilities\") pod \"redhat-marketplace-jvg9t\" (UID: \"030ed361-be9d-4dd0-88f3-f24f71716916\") " pod="openshift-marketplace/redhat-marketplace-jvg9t" Mar 18 20:44:08 crc kubenswrapper[4950]: I0318 20:44:08.135143 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/030ed361-be9d-4dd0-88f3-f24f71716916-catalog-content\") pod \"redhat-marketplace-jvg9t\" (UID: \"030ed361-be9d-4dd0-88f3-f24f71716916\") " pod="openshift-marketplace/redhat-marketplace-jvg9t" Mar 18 20:44:08 crc kubenswrapper[4950]: I0318 20:44:08.237299 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxl2l\" (UniqueName: \"kubernetes.io/projected/030ed361-be9d-4dd0-88f3-f24f71716916-kube-api-access-zxl2l\") pod \"redhat-marketplace-jvg9t\" (UID: \"030ed361-be9d-4dd0-88f3-f24f71716916\") " pod="openshift-marketplace/redhat-marketplace-jvg9t" Mar 18 20:44:08 crc kubenswrapper[4950]: I0318 20:44:08.237351 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/030ed361-be9d-4dd0-88f3-f24f71716916-utilities\") pod \"redhat-marketplace-jvg9t\" (UID: \"030ed361-be9d-4dd0-88f3-f24f71716916\") " pod="openshift-marketplace/redhat-marketplace-jvg9t" Mar 18 20:44:08 crc kubenswrapper[4950]: I0318 20:44:08.237375 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/030ed361-be9d-4dd0-88f3-f24f71716916-catalog-content\") pod \"redhat-marketplace-jvg9t\" (UID: \"030ed361-be9d-4dd0-88f3-f24f71716916\") " pod="openshift-marketplace/redhat-marketplace-jvg9t" Mar 18 20:44:08 crc kubenswrapper[4950]: I0318 20:44:08.237880 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/030ed361-be9d-4dd0-88f3-f24f71716916-catalog-content\") pod \"redhat-marketplace-jvg9t\" (UID: \"030ed361-be9d-4dd0-88f3-f24f71716916\") " pod="openshift-marketplace/redhat-marketplace-jvg9t" Mar 18 20:44:08 crc kubenswrapper[4950]: I0318 20:44:08.238038 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/030ed361-be9d-4dd0-88f3-f24f71716916-utilities\") pod \"redhat-marketplace-jvg9t\" (UID: \"030ed361-be9d-4dd0-88f3-f24f71716916\") " pod="openshift-marketplace/redhat-marketplace-jvg9t" Mar 18 20:44:08 crc kubenswrapper[4950]: I0318 20:44:08.256182 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxl2l\" (UniqueName: \"kubernetes.io/projected/030ed361-be9d-4dd0-88f3-f24f71716916-kube-api-access-zxl2l\") pod \"redhat-marketplace-jvg9t\" (UID: \"030ed361-be9d-4dd0-88f3-f24f71716916\") " pod="openshift-marketplace/redhat-marketplace-jvg9t" Mar 18 20:44:08 crc kubenswrapper[4950]: I0318 20:44:08.309950 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jvg9t" Mar 18 20:44:08 crc kubenswrapper[4950]: I0318 20:44:08.818838 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jvg9t"] Mar 18 20:44:09 crc kubenswrapper[4950]: E0318 20:44:09.154551 4950 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod030ed361_be9d_4dd0_88f3_f24f71716916.slice/crio-conmon-5efef99e3a644e1da9f175cb7afa70a6f18bdb74dd6854421b696c3da253736f.scope\": RecentStats: unable to find data in memory cache]" Mar 18 20:44:09 crc kubenswrapper[4950]: I0318 20:44:09.342891 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dbksg" Mar 18 20:44:09 crc kubenswrapper[4950]: I0318 20:44:09.342953 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dbksg" Mar 18 20:44:09 crc kubenswrapper[4950]: I0318 20:44:09.408605 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dbksg" Mar 18 20:44:09 crc kubenswrapper[4950]: I0318 20:44:09.654868 4950 generic.go:334] "Generic (PLEG): container finished" podID="030ed361-be9d-4dd0-88f3-f24f71716916" containerID="5efef99e3a644e1da9f175cb7afa70a6f18bdb74dd6854421b696c3da253736f" exitCode=0 Mar 18 20:44:09 crc kubenswrapper[4950]: I0318 20:44:09.654955 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvg9t" event={"ID":"030ed361-be9d-4dd0-88f3-f24f71716916","Type":"ContainerDied","Data":"5efef99e3a644e1da9f175cb7afa70a6f18bdb74dd6854421b696c3da253736f"} Mar 18 20:44:09 crc kubenswrapper[4950]: I0318 20:44:09.655006 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvg9t" event={"ID":"030ed361-be9d-4dd0-88f3-f24f71716916","Type":"ContainerStarted","Data":"2b031d442fd333620264c76424af08fb1dda57b75dfbe5fffe5aad5dbefc7c27"} Mar 18 20:44:09 crc kubenswrapper[4950]: I0318 20:44:09.714213 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dbksg" Mar 18 20:44:10 crc kubenswrapper[4950]: I0318 20:44:10.666069 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvg9t" event={"ID":"030ed361-be9d-4dd0-88f3-f24f71716916","Type":"ContainerStarted","Data":"13c621d8f148e6775f41927c66d035e32acfa835e16b51700f8e6d93c227b22f"} Mar 18 20:44:11 crc kubenswrapper[4950]: I0318 20:44:11.676933 4950 generic.go:334] "Generic (PLEG): container finished" podID="030ed361-be9d-4dd0-88f3-f24f71716916" containerID="13c621d8f148e6775f41927c66d035e32acfa835e16b51700f8e6d93c227b22f" exitCode=0 Mar 18 20:44:11 crc kubenswrapper[4950]: I0318 20:44:11.676987 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvg9t" event={"ID":"030ed361-be9d-4dd0-88f3-f24f71716916","Type":"ContainerDied","Data":"13c621d8f148e6775f41927c66d035e32acfa835e16b51700f8e6d93c227b22f"} Mar 18 20:44:11 crc kubenswrapper[4950]: I0318 20:44:11.760542 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dbksg"] Mar 18 20:44:11 crc kubenswrapper[4950]: I0318 20:44:11.760759 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dbksg" podUID="3de14edf-8d87-4da6-96bf-62b014bf3c35" containerName="registry-server" containerID="cri-o://6ab26df1a48dec1e448316b39644384bc4f9cbfb1d735c8ea19dd4c0157b9fff" gracePeriod=2 Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.253362 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbksg" Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.324608 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3de14edf-8d87-4da6-96bf-62b014bf3c35-catalog-content\") pod \"3de14edf-8d87-4da6-96bf-62b014bf3c35\" (UID: \"3de14edf-8d87-4da6-96bf-62b014bf3c35\") " Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.324849 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3de14edf-8d87-4da6-96bf-62b014bf3c35-utilities\") pod \"3de14edf-8d87-4da6-96bf-62b014bf3c35\" (UID: \"3de14edf-8d87-4da6-96bf-62b014bf3c35\") " Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.325022 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hglbc\" (UniqueName: \"kubernetes.io/projected/3de14edf-8d87-4da6-96bf-62b014bf3c35-kube-api-access-hglbc\") pod \"3de14edf-8d87-4da6-96bf-62b014bf3c35\" (UID: \"3de14edf-8d87-4da6-96bf-62b014bf3c35\") " Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.325651 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3de14edf-8d87-4da6-96bf-62b014bf3c35-utilities" (OuterVolumeSpecName: "utilities") pod "3de14edf-8d87-4da6-96bf-62b014bf3c35" (UID: "3de14edf-8d87-4da6-96bf-62b014bf3c35"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.334296 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3de14edf-8d87-4da6-96bf-62b014bf3c35-kube-api-access-hglbc" (OuterVolumeSpecName: "kube-api-access-hglbc") pod "3de14edf-8d87-4da6-96bf-62b014bf3c35" (UID: "3de14edf-8d87-4da6-96bf-62b014bf3c35"). InnerVolumeSpecName "kube-api-access-hglbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.384339 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3de14edf-8d87-4da6-96bf-62b014bf3c35-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3de14edf-8d87-4da6-96bf-62b014bf3c35" (UID: "3de14edf-8d87-4da6-96bf-62b014bf3c35"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.427660 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3de14edf-8d87-4da6-96bf-62b014bf3c35-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.427708 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hglbc\" (UniqueName: \"kubernetes.io/projected/3de14edf-8d87-4da6-96bf-62b014bf3c35-kube-api-access-hglbc\") on node \"crc\" DevicePath \"\"" Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.427721 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3de14edf-8d87-4da6-96bf-62b014bf3c35-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.686103 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvg9t" event={"ID":"030ed361-be9d-4dd0-88f3-f24f71716916","Type":"ContainerStarted","Data":"795a062a912929b083e90772f199efd2e3c04cdf098eada2226907293a4cf87e"} Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.689138 4950 generic.go:334] "Generic (PLEG): container finished" podID="3de14edf-8d87-4da6-96bf-62b014bf3c35" containerID="6ab26df1a48dec1e448316b39644384bc4f9cbfb1d735c8ea19dd4c0157b9fff" exitCode=0 Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.689165 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbksg" event={"ID":"3de14edf-8d87-4da6-96bf-62b014bf3c35","Type":"ContainerDied","Data":"6ab26df1a48dec1e448316b39644384bc4f9cbfb1d735c8ea19dd4c0157b9fff"} Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.689198 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dbksg" event={"ID":"3de14edf-8d87-4da6-96bf-62b014bf3c35","Type":"ContainerDied","Data":"3ecd3c24b2ff2e996f2e24fccd65bfa94518c719840f80758df71f43cbbd281c"} Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.689209 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dbksg" Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.689212 4950 scope.go:117] "RemoveContainer" containerID="6ab26df1a48dec1e448316b39644384bc4f9cbfb1d735c8ea19dd4c0157b9fff" Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.707855 4950 scope.go:117] "RemoveContainer" containerID="85d775d68fa91f9c40066554de5c93093a6cb4decf5faafa56f499840e25ecc8" Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.710552 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jvg9t" podStartSLOduration=3.146648544 podStartE2EDuration="5.710536952s" podCreationTimestamp="2026-03-18 20:44:07 +0000 UTC" firstStartedPulling="2026-03-18 20:44:09.660618686 +0000 UTC m=+2262.901460554" lastFinishedPulling="2026-03-18 20:44:12.224507094 +0000 UTC m=+2265.465348962" observedRunningTime="2026-03-18 20:44:12.708476218 +0000 UTC m=+2265.949318086" watchObservedRunningTime="2026-03-18 20:44:12.710536952 +0000 UTC m=+2265.951378820" Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.727224 4950 scope.go:117] "RemoveContainer" containerID="f8b5d98d08deaa7ca43e48abcf2274e7f3f86ab416044cb42ab8c3bba4920f14" Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.735515 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dbksg"] Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.744508 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dbksg"] Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.746364 4950 scope.go:117] "RemoveContainer" containerID="6ab26df1a48dec1e448316b39644384bc4f9cbfb1d735c8ea19dd4c0157b9fff" Mar 18 20:44:12 crc kubenswrapper[4950]: E0318 20:44:12.748581 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ab26df1a48dec1e448316b39644384bc4f9cbfb1d735c8ea19dd4c0157b9fff\": container with ID starting with 6ab26df1a48dec1e448316b39644384bc4f9cbfb1d735c8ea19dd4c0157b9fff not found: ID does not exist" containerID="6ab26df1a48dec1e448316b39644384bc4f9cbfb1d735c8ea19dd4c0157b9fff" Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.748623 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ab26df1a48dec1e448316b39644384bc4f9cbfb1d735c8ea19dd4c0157b9fff"} err="failed to get container status \"6ab26df1a48dec1e448316b39644384bc4f9cbfb1d735c8ea19dd4c0157b9fff\": rpc error: code = NotFound desc = could not find container \"6ab26df1a48dec1e448316b39644384bc4f9cbfb1d735c8ea19dd4c0157b9fff\": container with ID starting with 6ab26df1a48dec1e448316b39644384bc4f9cbfb1d735c8ea19dd4c0157b9fff not found: ID does not exist" Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.748651 4950 scope.go:117] "RemoveContainer" containerID="85d775d68fa91f9c40066554de5c93093a6cb4decf5faafa56f499840e25ecc8" Mar 18 20:44:12 crc kubenswrapper[4950]: E0318 20:44:12.749009 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85d775d68fa91f9c40066554de5c93093a6cb4decf5faafa56f499840e25ecc8\": container with ID starting with 85d775d68fa91f9c40066554de5c93093a6cb4decf5faafa56f499840e25ecc8 not found: ID does not exist" containerID="85d775d68fa91f9c40066554de5c93093a6cb4decf5faafa56f499840e25ecc8" Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.749027 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85d775d68fa91f9c40066554de5c93093a6cb4decf5faafa56f499840e25ecc8"} err="failed to get container status \"85d775d68fa91f9c40066554de5c93093a6cb4decf5faafa56f499840e25ecc8\": rpc error: code = NotFound desc = could not find container \"85d775d68fa91f9c40066554de5c93093a6cb4decf5faafa56f499840e25ecc8\": container with ID starting with 85d775d68fa91f9c40066554de5c93093a6cb4decf5faafa56f499840e25ecc8 not found: ID does not exist" Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.749039 4950 scope.go:117] "RemoveContainer" containerID="f8b5d98d08deaa7ca43e48abcf2274e7f3f86ab416044cb42ab8c3bba4920f14" Mar 18 20:44:12 crc kubenswrapper[4950]: E0318 20:44:12.749203 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8b5d98d08deaa7ca43e48abcf2274e7f3f86ab416044cb42ab8c3bba4920f14\": container with ID starting with f8b5d98d08deaa7ca43e48abcf2274e7f3f86ab416044cb42ab8c3bba4920f14 not found: ID does not exist" containerID="f8b5d98d08deaa7ca43e48abcf2274e7f3f86ab416044cb42ab8c3bba4920f14" Mar 18 20:44:12 crc kubenswrapper[4950]: I0318 20:44:12.749221 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8b5d98d08deaa7ca43e48abcf2274e7f3f86ab416044cb42ab8c3bba4920f14"} err="failed to get container status \"f8b5d98d08deaa7ca43e48abcf2274e7f3f86ab416044cb42ab8c3bba4920f14\": rpc error: code = NotFound desc = could not find container \"f8b5d98d08deaa7ca43e48abcf2274e7f3f86ab416044cb42ab8c3bba4920f14\": container with ID starting with f8b5d98d08deaa7ca43e48abcf2274e7f3f86ab416044cb42ab8c3bba4920f14 not found: ID does not exist" Mar 18 20:44:13 crc kubenswrapper[4950]: I0318 20:44:13.493180 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3de14edf-8d87-4da6-96bf-62b014bf3c35" path="/var/lib/kubelet/pods/3de14edf-8d87-4da6-96bf-62b014bf3c35/volumes" Mar 18 20:44:16 crc kubenswrapper[4950]: I0318 20:44:16.189466 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wdrqx"] Mar 18 20:44:16 crc kubenswrapper[4950]: E0318 20:44:16.190681 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3de14edf-8d87-4da6-96bf-62b014bf3c35" containerName="registry-server" Mar 18 20:44:16 crc kubenswrapper[4950]: I0318 20:44:16.190716 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="3de14edf-8d87-4da6-96bf-62b014bf3c35" containerName="registry-server" Mar 18 20:44:16 crc kubenswrapper[4950]: E0318 20:44:16.190740 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3de14edf-8d87-4da6-96bf-62b014bf3c35" containerName="extract-utilities" Mar 18 20:44:16 crc kubenswrapper[4950]: I0318 20:44:16.190757 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="3de14edf-8d87-4da6-96bf-62b014bf3c35" containerName="extract-utilities" Mar 18 20:44:16 crc kubenswrapper[4950]: E0318 20:44:16.190802 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3de14edf-8d87-4da6-96bf-62b014bf3c35" containerName="extract-content" Mar 18 20:44:16 crc kubenswrapper[4950]: I0318 20:44:16.190821 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="3de14edf-8d87-4da6-96bf-62b014bf3c35" containerName="extract-content" Mar 18 20:44:16 crc kubenswrapper[4950]: I0318 20:44:16.191272 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="3de14edf-8d87-4da6-96bf-62b014bf3c35" containerName="registry-server" Mar 18 20:44:16 crc kubenswrapper[4950]: I0318 20:44:16.193997 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wdrqx" Mar 18 20:44:16 crc kubenswrapper[4950]: I0318 20:44:16.205742 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wdrqx"] Mar 18 20:44:16 crc kubenswrapper[4950]: I0318 20:44:16.302070 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/436804c7-70e5-42ef-b7e5-14c02f4fd1ac-catalog-content\") pod \"community-operators-wdrqx\" (UID: \"436804c7-70e5-42ef-b7e5-14c02f4fd1ac\") " pod="openshift-marketplace/community-operators-wdrqx" Mar 18 20:44:16 crc kubenswrapper[4950]: I0318 20:44:16.302169 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rks7c\" (UniqueName: \"kubernetes.io/projected/436804c7-70e5-42ef-b7e5-14c02f4fd1ac-kube-api-access-rks7c\") pod \"community-operators-wdrqx\" (UID: \"436804c7-70e5-42ef-b7e5-14c02f4fd1ac\") " pod="openshift-marketplace/community-operators-wdrqx" Mar 18 20:44:16 crc kubenswrapper[4950]: I0318 20:44:16.302225 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/436804c7-70e5-42ef-b7e5-14c02f4fd1ac-utilities\") pod \"community-operators-wdrqx\" (UID: \"436804c7-70e5-42ef-b7e5-14c02f4fd1ac\") " pod="openshift-marketplace/community-operators-wdrqx" Mar 18 20:44:16 crc kubenswrapper[4950]: I0318 20:44:16.404380 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/436804c7-70e5-42ef-b7e5-14c02f4fd1ac-catalog-content\") pod \"community-operators-wdrqx\" (UID: \"436804c7-70e5-42ef-b7e5-14c02f4fd1ac\") " pod="openshift-marketplace/community-operators-wdrqx" Mar 18 20:44:16 crc kubenswrapper[4950]: I0318 20:44:16.404516 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rks7c\" (UniqueName: \"kubernetes.io/projected/436804c7-70e5-42ef-b7e5-14c02f4fd1ac-kube-api-access-rks7c\") pod \"community-operators-wdrqx\" (UID: \"436804c7-70e5-42ef-b7e5-14c02f4fd1ac\") " pod="openshift-marketplace/community-operators-wdrqx" Mar 18 20:44:16 crc kubenswrapper[4950]: I0318 20:44:16.404577 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/436804c7-70e5-42ef-b7e5-14c02f4fd1ac-utilities\") pod \"community-operators-wdrqx\" (UID: \"436804c7-70e5-42ef-b7e5-14c02f4fd1ac\") " pod="openshift-marketplace/community-operators-wdrqx" Mar 18 20:44:16 crc kubenswrapper[4950]: I0318 20:44:16.405051 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/436804c7-70e5-42ef-b7e5-14c02f4fd1ac-utilities\") pod \"community-operators-wdrqx\" (UID: \"436804c7-70e5-42ef-b7e5-14c02f4fd1ac\") " pod="openshift-marketplace/community-operators-wdrqx" Mar 18 20:44:16 crc kubenswrapper[4950]: I0318 20:44:16.405047 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/436804c7-70e5-42ef-b7e5-14c02f4fd1ac-catalog-content\") pod \"community-operators-wdrqx\" (UID: \"436804c7-70e5-42ef-b7e5-14c02f4fd1ac\") " pod="openshift-marketplace/community-operators-wdrqx" Mar 18 20:44:16 crc kubenswrapper[4950]: I0318 20:44:16.427699 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rks7c\" (UniqueName: \"kubernetes.io/projected/436804c7-70e5-42ef-b7e5-14c02f4fd1ac-kube-api-access-rks7c\") pod \"community-operators-wdrqx\" (UID: \"436804c7-70e5-42ef-b7e5-14c02f4fd1ac\") " pod="openshift-marketplace/community-operators-wdrqx" Mar 18 20:44:16 crc kubenswrapper[4950]: I0318 20:44:16.543868 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wdrqx" Mar 18 20:44:17 crc kubenswrapper[4950]: I0318 20:44:17.116714 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wdrqx"] Mar 18 20:44:17 crc kubenswrapper[4950]: I0318 20:44:17.761432 4950 generic.go:334] "Generic (PLEG): container finished" podID="436804c7-70e5-42ef-b7e5-14c02f4fd1ac" containerID="1bd43c64411fa94ec2dc328de5e59998cdaa3ffbb5e4f0103dbb6c78fe62370a" exitCode=0 Mar 18 20:44:17 crc kubenswrapper[4950]: I0318 20:44:17.761472 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wdrqx" event={"ID":"436804c7-70e5-42ef-b7e5-14c02f4fd1ac","Type":"ContainerDied","Data":"1bd43c64411fa94ec2dc328de5e59998cdaa3ffbb5e4f0103dbb6c78fe62370a"} Mar 18 20:44:17 crc kubenswrapper[4950]: I0318 20:44:17.761507 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wdrqx" event={"ID":"436804c7-70e5-42ef-b7e5-14c02f4fd1ac","Type":"ContainerStarted","Data":"83d1dfd0fa736078bc70ed22a11c5013450a33bb00625090e23bc64be942b4e7"} Mar 18 20:44:18 crc kubenswrapper[4950]: I0318 20:44:18.310529 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jvg9t" Mar 18 20:44:18 crc kubenswrapper[4950]: I0318 20:44:18.312442 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jvg9t" Mar 18 20:44:18 crc kubenswrapper[4950]: I0318 20:44:18.390861 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jvg9t" Mar 18 20:44:18 crc kubenswrapper[4950]: I0318 20:44:18.771762 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wdrqx" event={"ID":"436804c7-70e5-42ef-b7e5-14c02f4fd1ac","Type":"ContainerStarted","Data":"bab42cbb6e92a3432afc62ff577fe7eadd846e6776626efa78fc4f57742d877a"} Mar 18 20:44:18 crc kubenswrapper[4950]: I0318 20:44:18.824130 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jvg9t" Mar 18 20:44:20 crc kubenswrapper[4950]: I0318 20:44:20.771484 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jvg9t"] Mar 18 20:44:20 crc kubenswrapper[4950]: I0318 20:44:20.821552 4950 generic.go:334] "Generic (PLEG): container finished" podID="436804c7-70e5-42ef-b7e5-14c02f4fd1ac" containerID="bab42cbb6e92a3432afc62ff577fe7eadd846e6776626efa78fc4f57742d877a" exitCode=0 Mar 18 20:44:20 crc kubenswrapper[4950]: I0318 20:44:20.821666 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wdrqx" event={"ID":"436804c7-70e5-42ef-b7e5-14c02f4fd1ac","Type":"ContainerDied","Data":"bab42cbb6e92a3432afc62ff577fe7eadd846e6776626efa78fc4f57742d877a"} Mar 18 20:44:20 crc kubenswrapper[4950]: I0318 20:44:20.822165 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jvg9t" podUID="030ed361-be9d-4dd0-88f3-f24f71716916" containerName="registry-server" containerID="cri-o://795a062a912929b083e90772f199efd2e3c04cdf098eada2226907293a4cf87e" gracePeriod=2 Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.331029 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jvg9t" Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.401670 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxl2l\" (UniqueName: \"kubernetes.io/projected/030ed361-be9d-4dd0-88f3-f24f71716916-kube-api-access-zxl2l\") pod \"030ed361-be9d-4dd0-88f3-f24f71716916\" (UID: \"030ed361-be9d-4dd0-88f3-f24f71716916\") " Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.401712 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/030ed361-be9d-4dd0-88f3-f24f71716916-utilities\") pod \"030ed361-be9d-4dd0-88f3-f24f71716916\" (UID: \"030ed361-be9d-4dd0-88f3-f24f71716916\") " Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.401805 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/030ed361-be9d-4dd0-88f3-f24f71716916-catalog-content\") pod \"030ed361-be9d-4dd0-88f3-f24f71716916\" (UID: \"030ed361-be9d-4dd0-88f3-f24f71716916\") " Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.402695 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/030ed361-be9d-4dd0-88f3-f24f71716916-utilities" (OuterVolumeSpecName: "utilities") pod "030ed361-be9d-4dd0-88f3-f24f71716916" (UID: "030ed361-be9d-4dd0-88f3-f24f71716916"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.411691 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/030ed361-be9d-4dd0-88f3-f24f71716916-kube-api-access-zxl2l" (OuterVolumeSpecName: "kube-api-access-zxl2l") pod "030ed361-be9d-4dd0-88f3-f24f71716916" (UID: "030ed361-be9d-4dd0-88f3-f24f71716916"). InnerVolumeSpecName "kube-api-access-zxl2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.431057 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/030ed361-be9d-4dd0-88f3-f24f71716916-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "030ed361-be9d-4dd0-88f3-f24f71716916" (UID: "030ed361-be9d-4dd0-88f3-f24f71716916"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.503696 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxl2l\" (UniqueName: \"kubernetes.io/projected/030ed361-be9d-4dd0-88f3-f24f71716916-kube-api-access-zxl2l\") on node \"crc\" DevicePath \"\"" Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.503723 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/030ed361-be9d-4dd0-88f3-f24f71716916-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.503732 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/030ed361-be9d-4dd0-88f3-f24f71716916-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.836656 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wdrqx" event={"ID":"436804c7-70e5-42ef-b7e5-14c02f4fd1ac","Type":"ContainerStarted","Data":"6d929c2faca8245e0052b2edc4207ffcde7d32a8eef4f6fce828b81838fe960e"} Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.846008 4950 generic.go:334] "Generic (PLEG): container finished" podID="030ed361-be9d-4dd0-88f3-f24f71716916" containerID="795a062a912929b083e90772f199efd2e3c04cdf098eada2226907293a4cf87e" exitCode=0 Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.846059 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jvg9t" Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.846063 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvg9t" event={"ID":"030ed361-be9d-4dd0-88f3-f24f71716916","Type":"ContainerDied","Data":"795a062a912929b083e90772f199efd2e3c04cdf098eada2226907293a4cf87e"} Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.846128 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvg9t" event={"ID":"030ed361-be9d-4dd0-88f3-f24f71716916","Type":"ContainerDied","Data":"2b031d442fd333620264c76424af08fb1dda57b75dfbe5fffe5aad5dbefc7c27"} Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.846153 4950 scope.go:117] "RemoveContainer" containerID="795a062a912929b083e90772f199efd2e3c04cdf098eada2226907293a4cf87e" Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.862361 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wdrqx" podStartSLOduration=2.171242095 podStartE2EDuration="5.862336714s" podCreationTimestamp="2026-03-18 20:44:16 +0000 UTC" firstStartedPulling="2026-03-18 20:44:17.763753181 +0000 UTC m=+2271.004595049" lastFinishedPulling="2026-03-18 20:44:21.45484779 +0000 UTC m=+2274.695689668" observedRunningTime="2026-03-18 20:44:21.855527635 +0000 UTC m=+2275.096369513" watchObservedRunningTime="2026-03-18 20:44:21.862336714 +0000 UTC m=+2275.103178582" Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.881168 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jvg9t"] Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.885145 4950 scope.go:117] "RemoveContainer" containerID="13c621d8f148e6775f41927c66d035e32acfa835e16b51700f8e6d93c227b22f" Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.893131 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jvg9t"] Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.901602 4950 scope.go:117] "RemoveContainer" containerID="5efef99e3a644e1da9f175cb7afa70a6f18bdb74dd6854421b696c3da253736f" Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.917294 4950 scope.go:117] "RemoveContainer" containerID="795a062a912929b083e90772f199efd2e3c04cdf098eada2226907293a4cf87e" Mar 18 20:44:21 crc kubenswrapper[4950]: E0318 20:44:21.917760 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"795a062a912929b083e90772f199efd2e3c04cdf098eada2226907293a4cf87e\": container with ID starting with 795a062a912929b083e90772f199efd2e3c04cdf098eada2226907293a4cf87e not found: ID does not exist" containerID="795a062a912929b083e90772f199efd2e3c04cdf098eada2226907293a4cf87e" Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.917891 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"795a062a912929b083e90772f199efd2e3c04cdf098eada2226907293a4cf87e"} err="failed to get container status \"795a062a912929b083e90772f199efd2e3c04cdf098eada2226907293a4cf87e\": rpc error: code = NotFound desc = could not find container \"795a062a912929b083e90772f199efd2e3c04cdf098eada2226907293a4cf87e\": container with ID starting with 795a062a912929b083e90772f199efd2e3c04cdf098eada2226907293a4cf87e not found: ID does not exist" Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.918028 4950 scope.go:117] "RemoveContainer" containerID="13c621d8f148e6775f41927c66d035e32acfa835e16b51700f8e6d93c227b22f" Mar 18 20:44:21 crc kubenswrapper[4950]: E0318 20:44:21.918488 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13c621d8f148e6775f41927c66d035e32acfa835e16b51700f8e6d93c227b22f\": container with ID starting with 13c621d8f148e6775f41927c66d035e32acfa835e16b51700f8e6d93c227b22f not found: ID does not exist" containerID="13c621d8f148e6775f41927c66d035e32acfa835e16b51700f8e6d93c227b22f" Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.918512 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13c621d8f148e6775f41927c66d035e32acfa835e16b51700f8e6d93c227b22f"} err="failed to get container status \"13c621d8f148e6775f41927c66d035e32acfa835e16b51700f8e6d93c227b22f\": rpc error: code = NotFound desc = could not find container \"13c621d8f148e6775f41927c66d035e32acfa835e16b51700f8e6d93c227b22f\": container with ID starting with 13c621d8f148e6775f41927c66d035e32acfa835e16b51700f8e6d93c227b22f not found: ID does not exist" Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.918531 4950 scope.go:117] "RemoveContainer" containerID="5efef99e3a644e1da9f175cb7afa70a6f18bdb74dd6854421b696c3da253736f" Mar 18 20:44:21 crc kubenswrapper[4950]: E0318 20:44:21.920750 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5efef99e3a644e1da9f175cb7afa70a6f18bdb74dd6854421b696c3da253736f\": container with ID starting with 5efef99e3a644e1da9f175cb7afa70a6f18bdb74dd6854421b696c3da253736f not found: ID does not exist" containerID="5efef99e3a644e1da9f175cb7afa70a6f18bdb74dd6854421b696c3da253736f" Mar 18 20:44:21 crc kubenswrapper[4950]: I0318 20:44:21.920821 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5efef99e3a644e1da9f175cb7afa70a6f18bdb74dd6854421b696c3da253736f"} err="failed to get container status \"5efef99e3a644e1da9f175cb7afa70a6f18bdb74dd6854421b696c3da253736f\": rpc error: code = NotFound desc = could not find container \"5efef99e3a644e1da9f175cb7afa70a6f18bdb74dd6854421b696c3da253736f\": container with ID starting with 5efef99e3a644e1da9f175cb7afa70a6f18bdb74dd6854421b696c3da253736f not found: ID does not exist" Mar 18 20:44:23 crc kubenswrapper[4950]: I0318 20:44:23.495932 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="030ed361-be9d-4dd0-88f3-f24f71716916" path="/var/lib/kubelet/pods/030ed361-be9d-4dd0-88f3-f24f71716916/volumes" Mar 18 20:44:26 crc kubenswrapper[4950]: I0318 20:44:26.544702 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wdrqx" Mar 18 20:44:26 crc kubenswrapper[4950]: I0318 20:44:26.546592 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wdrqx" Mar 18 20:44:26 crc kubenswrapper[4950]: I0318 20:44:26.634580 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wdrqx" Mar 18 20:44:26 crc kubenswrapper[4950]: I0318 20:44:26.987718 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wdrqx" Mar 18 20:44:28 crc kubenswrapper[4950]: I0318 20:44:28.367539 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wdrqx"] Mar 18 20:44:29 crc kubenswrapper[4950]: I0318 20:44:29.931057 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wdrqx" podUID="436804c7-70e5-42ef-b7e5-14c02f4fd1ac" containerName="registry-server" containerID="cri-o://6d929c2faca8245e0052b2edc4207ffcde7d32a8eef4f6fce828b81838fe960e" gracePeriod=2 Mar 18 20:44:30 crc kubenswrapper[4950]: I0318 20:44:30.384363 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wdrqx" Mar 18 20:44:30 crc kubenswrapper[4950]: I0318 20:44:30.496444 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rks7c\" (UniqueName: \"kubernetes.io/projected/436804c7-70e5-42ef-b7e5-14c02f4fd1ac-kube-api-access-rks7c\") pod \"436804c7-70e5-42ef-b7e5-14c02f4fd1ac\" (UID: \"436804c7-70e5-42ef-b7e5-14c02f4fd1ac\") " Mar 18 20:44:30 crc kubenswrapper[4950]: I0318 20:44:30.496634 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/436804c7-70e5-42ef-b7e5-14c02f4fd1ac-catalog-content\") pod \"436804c7-70e5-42ef-b7e5-14c02f4fd1ac\" (UID: \"436804c7-70e5-42ef-b7e5-14c02f4fd1ac\") " Mar 18 20:44:30 crc kubenswrapper[4950]: I0318 20:44:30.496831 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/436804c7-70e5-42ef-b7e5-14c02f4fd1ac-utilities\") pod \"436804c7-70e5-42ef-b7e5-14c02f4fd1ac\" (UID: \"436804c7-70e5-42ef-b7e5-14c02f4fd1ac\") " Mar 18 20:44:30 crc kubenswrapper[4950]: I0318 20:44:30.497710 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/436804c7-70e5-42ef-b7e5-14c02f4fd1ac-utilities" (OuterVolumeSpecName: "utilities") pod "436804c7-70e5-42ef-b7e5-14c02f4fd1ac" (UID: "436804c7-70e5-42ef-b7e5-14c02f4fd1ac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:44:30 crc kubenswrapper[4950]: I0318 20:44:30.510691 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/436804c7-70e5-42ef-b7e5-14c02f4fd1ac-kube-api-access-rks7c" (OuterVolumeSpecName: "kube-api-access-rks7c") pod "436804c7-70e5-42ef-b7e5-14c02f4fd1ac" (UID: "436804c7-70e5-42ef-b7e5-14c02f4fd1ac"). InnerVolumeSpecName "kube-api-access-rks7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:44:30 crc kubenswrapper[4950]: I0318 20:44:30.559563 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/436804c7-70e5-42ef-b7e5-14c02f4fd1ac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "436804c7-70e5-42ef-b7e5-14c02f4fd1ac" (UID: "436804c7-70e5-42ef-b7e5-14c02f4fd1ac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:44:30 crc kubenswrapper[4950]: I0318 20:44:30.599460 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rks7c\" (UniqueName: \"kubernetes.io/projected/436804c7-70e5-42ef-b7e5-14c02f4fd1ac-kube-api-access-rks7c\") on node \"crc\" DevicePath \"\"" Mar 18 20:44:30 crc kubenswrapper[4950]: I0318 20:44:30.599497 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/436804c7-70e5-42ef-b7e5-14c02f4fd1ac-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:44:30 crc kubenswrapper[4950]: I0318 20:44:30.599509 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/436804c7-70e5-42ef-b7e5-14c02f4fd1ac-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:44:30 crc kubenswrapper[4950]: I0318 20:44:30.940513 4950 generic.go:334] "Generic (PLEG): container finished" podID="436804c7-70e5-42ef-b7e5-14c02f4fd1ac" containerID="6d929c2faca8245e0052b2edc4207ffcde7d32a8eef4f6fce828b81838fe960e" exitCode=0 Mar 18 20:44:30 crc kubenswrapper[4950]: I0318 20:44:30.940562 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wdrqx" event={"ID":"436804c7-70e5-42ef-b7e5-14c02f4fd1ac","Type":"ContainerDied","Data":"6d929c2faca8245e0052b2edc4207ffcde7d32a8eef4f6fce828b81838fe960e"} Mar 18 20:44:30 crc kubenswrapper[4950]: I0318 20:44:30.940596 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wdrqx" event={"ID":"436804c7-70e5-42ef-b7e5-14c02f4fd1ac","Type":"ContainerDied","Data":"83d1dfd0fa736078bc70ed22a11c5013450a33bb00625090e23bc64be942b4e7"} Mar 18 20:44:30 crc kubenswrapper[4950]: I0318 20:44:30.940617 4950 scope.go:117] "RemoveContainer" containerID="6d929c2faca8245e0052b2edc4207ffcde7d32a8eef4f6fce828b81838fe960e" Mar 18 20:44:30 crc kubenswrapper[4950]: I0318 20:44:30.940615 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wdrqx" Mar 18 20:44:30 crc kubenswrapper[4950]: I0318 20:44:30.956637 4950 scope.go:117] "RemoveContainer" containerID="bab42cbb6e92a3432afc62ff577fe7eadd846e6776626efa78fc4f57742d877a" Mar 18 20:44:30 crc kubenswrapper[4950]: I0318 20:44:30.981167 4950 scope.go:117] "RemoveContainer" containerID="1bd43c64411fa94ec2dc328de5e59998cdaa3ffbb5e4f0103dbb6c78fe62370a" Mar 18 20:44:31 crc kubenswrapper[4950]: I0318 20:44:31.038787 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wdrqx"] Mar 18 20:44:31 crc kubenswrapper[4950]: I0318 20:44:31.039046 4950 scope.go:117] "RemoveContainer" containerID="6d929c2faca8245e0052b2edc4207ffcde7d32a8eef4f6fce828b81838fe960e" Mar 18 20:44:31 crc kubenswrapper[4950]: E0318 20:44:31.039817 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d929c2faca8245e0052b2edc4207ffcde7d32a8eef4f6fce828b81838fe960e\": container with ID starting with 6d929c2faca8245e0052b2edc4207ffcde7d32a8eef4f6fce828b81838fe960e not found: ID does not exist" containerID="6d929c2faca8245e0052b2edc4207ffcde7d32a8eef4f6fce828b81838fe960e" Mar 18 20:44:31 crc kubenswrapper[4950]: I0318 20:44:31.039854 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d929c2faca8245e0052b2edc4207ffcde7d32a8eef4f6fce828b81838fe960e"} err="failed to get container status \"6d929c2faca8245e0052b2edc4207ffcde7d32a8eef4f6fce828b81838fe960e\": rpc error: code = NotFound desc = could not find container \"6d929c2faca8245e0052b2edc4207ffcde7d32a8eef4f6fce828b81838fe960e\": container with ID starting with 6d929c2faca8245e0052b2edc4207ffcde7d32a8eef4f6fce828b81838fe960e not found: ID does not exist" Mar 18 20:44:31 crc kubenswrapper[4950]: I0318 20:44:31.039878 4950 scope.go:117] "RemoveContainer" containerID="bab42cbb6e92a3432afc62ff577fe7eadd846e6776626efa78fc4f57742d877a" Mar 18 20:44:31 crc kubenswrapper[4950]: E0318 20:44:31.040130 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bab42cbb6e92a3432afc62ff577fe7eadd846e6776626efa78fc4f57742d877a\": container with ID starting with bab42cbb6e92a3432afc62ff577fe7eadd846e6776626efa78fc4f57742d877a not found: ID does not exist" containerID="bab42cbb6e92a3432afc62ff577fe7eadd846e6776626efa78fc4f57742d877a" Mar 18 20:44:31 crc kubenswrapper[4950]: I0318 20:44:31.040174 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bab42cbb6e92a3432afc62ff577fe7eadd846e6776626efa78fc4f57742d877a"} err="failed to get container status \"bab42cbb6e92a3432afc62ff577fe7eadd846e6776626efa78fc4f57742d877a\": rpc error: code = NotFound desc = could not find container \"bab42cbb6e92a3432afc62ff577fe7eadd846e6776626efa78fc4f57742d877a\": container with ID starting with bab42cbb6e92a3432afc62ff577fe7eadd846e6776626efa78fc4f57742d877a not found: ID does not exist" Mar 18 20:44:31 crc kubenswrapper[4950]: I0318 20:44:31.040204 4950 scope.go:117] "RemoveContainer" containerID="1bd43c64411fa94ec2dc328de5e59998cdaa3ffbb5e4f0103dbb6c78fe62370a" Mar 18 20:44:31 crc kubenswrapper[4950]: E0318 20:44:31.040472 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bd43c64411fa94ec2dc328de5e59998cdaa3ffbb5e4f0103dbb6c78fe62370a\": container with ID starting with 1bd43c64411fa94ec2dc328de5e59998cdaa3ffbb5e4f0103dbb6c78fe62370a not found: ID does not exist" containerID="1bd43c64411fa94ec2dc328de5e59998cdaa3ffbb5e4f0103dbb6c78fe62370a" Mar 18 20:44:31 crc kubenswrapper[4950]: I0318 20:44:31.040495 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bd43c64411fa94ec2dc328de5e59998cdaa3ffbb5e4f0103dbb6c78fe62370a"} err="failed to get container status \"1bd43c64411fa94ec2dc328de5e59998cdaa3ffbb5e4f0103dbb6c78fe62370a\": rpc error: code = NotFound desc = could not find container \"1bd43c64411fa94ec2dc328de5e59998cdaa3ffbb5e4f0103dbb6c78fe62370a\": container with ID starting with 1bd43c64411fa94ec2dc328de5e59998cdaa3ffbb5e4f0103dbb6c78fe62370a not found: ID does not exist" Mar 18 20:44:31 crc kubenswrapper[4950]: I0318 20:44:31.048269 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wdrqx"] Mar 18 20:44:31 crc kubenswrapper[4950]: I0318 20:44:31.491045 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="436804c7-70e5-42ef-b7e5-14c02f4fd1ac" path="/var/lib/kubelet/pods/436804c7-70e5-42ef-b7e5-14c02f4fd1ac/volumes" Mar 18 20:44:33 crc kubenswrapper[4950]: I0318 20:44:33.836714 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:44:33 crc kubenswrapper[4950]: I0318 20:44:33.837016 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:44:33 crc kubenswrapper[4950]: I0318 20:44:33.837074 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:44:33 crc kubenswrapper[4950]: I0318 20:44:33.837916 4950 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0eb938fd8ae6f1d971cd0e6b1a12f847fe46b0c1dd5c1077e4c34d45028879ca"} pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 20:44:33 crc kubenswrapper[4950]: I0318 20:44:33.837974 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" containerID="cri-o://0eb938fd8ae6f1d971cd0e6b1a12f847fe46b0c1dd5c1077e4c34d45028879ca" gracePeriod=600 Mar 18 20:44:33 crc kubenswrapper[4950]: I0318 20:44:33.974673 4950 generic.go:334] "Generic (PLEG): container finished" podID="4048b439-3266-46e7-9de0-22377efacc46" containerID="0eb938fd8ae6f1d971cd0e6b1a12f847fe46b0c1dd5c1077e4c34d45028879ca" exitCode=0 Mar 18 20:44:33 crc kubenswrapper[4950]: I0318 20:44:33.974722 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerDied","Data":"0eb938fd8ae6f1d971cd0e6b1a12f847fe46b0c1dd5c1077e4c34d45028879ca"} Mar 18 20:44:33 crc kubenswrapper[4950]: I0318 20:44:33.974758 4950 scope.go:117] "RemoveContainer" containerID="f57b4a97cc7648822ae6bc2fcd058e2471d722e28d65db3f78edef86676ef53e" Mar 18 20:44:34 crc kubenswrapper[4950]: I0318 20:44:34.992918 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerStarted","Data":"75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18"} Mar 18 20:44:49 crc kubenswrapper[4950]: I0318 20:44:49.820470 4950 scope.go:117] "RemoveContainer" containerID="bfb5f2081a3871a172aa70452dcf9a517756cd507b27dae22e2b7c5c719cfa38" Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.186568 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564445-wtmkn"] Mar 18 20:45:00 crc kubenswrapper[4950]: E0318 20:45:00.187369 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="030ed361-be9d-4dd0-88f3-f24f71716916" containerName="extract-content" Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.187382 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="030ed361-be9d-4dd0-88f3-f24f71716916" containerName="extract-content" Mar 18 20:45:00 crc kubenswrapper[4950]: E0318 20:45:00.187399 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="030ed361-be9d-4dd0-88f3-f24f71716916" containerName="registry-server" Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.187428 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="030ed361-be9d-4dd0-88f3-f24f71716916" containerName="registry-server" Mar 18 20:45:00 crc kubenswrapper[4950]: E0318 20:45:00.187441 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="436804c7-70e5-42ef-b7e5-14c02f4fd1ac" containerName="registry-server" Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.187447 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="436804c7-70e5-42ef-b7e5-14c02f4fd1ac" containerName="registry-server" Mar 18 20:45:00 crc kubenswrapper[4950]: E0318 20:45:00.187457 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="436804c7-70e5-42ef-b7e5-14c02f4fd1ac" containerName="extract-content" Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.187464 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="436804c7-70e5-42ef-b7e5-14c02f4fd1ac" containerName="extract-content" Mar 18 20:45:00 crc kubenswrapper[4950]: E0318 20:45:00.187480 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="436804c7-70e5-42ef-b7e5-14c02f4fd1ac" containerName="extract-utilities" Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.187486 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="436804c7-70e5-42ef-b7e5-14c02f4fd1ac" containerName="extract-utilities" Mar 18 20:45:00 crc kubenswrapper[4950]: E0318 20:45:00.187496 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="030ed361-be9d-4dd0-88f3-f24f71716916" containerName="extract-utilities" Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.187502 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="030ed361-be9d-4dd0-88f3-f24f71716916" containerName="extract-utilities" Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.187664 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="436804c7-70e5-42ef-b7e5-14c02f4fd1ac" containerName="registry-server" Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.187674 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="030ed361-be9d-4dd0-88f3-f24f71716916" containerName="registry-server" Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.188215 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564445-wtmkn" Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.190305 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.190560 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.207730 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564445-wtmkn"] Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.280354 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a66e031-b5dd-49e5-91e9-35ea7071c207-secret-volume\") pod \"collect-profiles-29564445-wtmkn\" (UID: \"1a66e031-b5dd-49e5-91e9-35ea7071c207\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564445-wtmkn" Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.280544 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a66e031-b5dd-49e5-91e9-35ea7071c207-config-volume\") pod \"collect-profiles-29564445-wtmkn\" (UID: \"1a66e031-b5dd-49e5-91e9-35ea7071c207\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564445-wtmkn" Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.280574 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtgfl\" (UniqueName: \"kubernetes.io/projected/1a66e031-b5dd-49e5-91e9-35ea7071c207-kube-api-access-jtgfl\") pod \"collect-profiles-29564445-wtmkn\" (UID: \"1a66e031-b5dd-49e5-91e9-35ea7071c207\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564445-wtmkn" Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.382231 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a66e031-b5dd-49e5-91e9-35ea7071c207-secret-volume\") pod \"collect-profiles-29564445-wtmkn\" (UID: \"1a66e031-b5dd-49e5-91e9-35ea7071c207\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564445-wtmkn" Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.382319 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a66e031-b5dd-49e5-91e9-35ea7071c207-config-volume\") pod \"collect-profiles-29564445-wtmkn\" (UID: \"1a66e031-b5dd-49e5-91e9-35ea7071c207\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564445-wtmkn" Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.382352 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtgfl\" (UniqueName: \"kubernetes.io/projected/1a66e031-b5dd-49e5-91e9-35ea7071c207-kube-api-access-jtgfl\") pod \"collect-profiles-29564445-wtmkn\" (UID: \"1a66e031-b5dd-49e5-91e9-35ea7071c207\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564445-wtmkn" Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.384189 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a66e031-b5dd-49e5-91e9-35ea7071c207-config-volume\") pod \"collect-profiles-29564445-wtmkn\" (UID: \"1a66e031-b5dd-49e5-91e9-35ea7071c207\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564445-wtmkn" Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.416470 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtgfl\" (UniqueName: \"kubernetes.io/projected/1a66e031-b5dd-49e5-91e9-35ea7071c207-kube-api-access-jtgfl\") pod \"collect-profiles-29564445-wtmkn\" (UID: \"1a66e031-b5dd-49e5-91e9-35ea7071c207\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564445-wtmkn" Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.420098 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a66e031-b5dd-49e5-91e9-35ea7071c207-secret-volume\") pod \"collect-profiles-29564445-wtmkn\" (UID: \"1a66e031-b5dd-49e5-91e9-35ea7071c207\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564445-wtmkn" Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.520459 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564445-wtmkn" Mar 18 20:45:00 crc kubenswrapper[4950]: I0318 20:45:00.981946 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564445-wtmkn"] Mar 18 20:45:01 crc kubenswrapper[4950]: I0318 20:45:01.218272 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564445-wtmkn" event={"ID":"1a66e031-b5dd-49e5-91e9-35ea7071c207","Type":"ContainerStarted","Data":"db2c55f91c2c35b9c837ceca97531e7241f2e3130c5bc62b34d7f2b9a6b1b649"} Mar 18 20:45:01 crc kubenswrapper[4950]: I0318 20:45:01.218567 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564445-wtmkn" event={"ID":"1a66e031-b5dd-49e5-91e9-35ea7071c207","Type":"ContainerStarted","Data":"582f76d7f32419c06f8ad21d80aed60b641b411c5a73b48bb7437cdff00bce95"} Mar 18 20:45:01 crc kubenswrapper[4950]: I0318 20:45:01.245272 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29564445-wtmkn" podStartSLOduration=1.2452492 podStartE2EDuration="1.2452492s" podCreationTimestamp="2026-03-18 20:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:45:01.235850273 +0000 UTC m=+2314.476692141" watchObservedRunningTime="2026-03-18 20:45:01.2452492 +0000 UTC m=+2314.486091078" Mar 18 20:45:02 crc kubenswrapper[4950]: I0318 20:45:02.227790 4950 generic.go:334] "Generic (PLEG): container finished" podID="1a66e031-b5dd-49e5-91e9-35ea7071c207" containerID="db2c55f91c2c35b9c837ceca97531e7241f2e3130c5bc62b34d7f2b9a6b1b649" exitCode=0 Mar 18 20:45:02 crc kubenswrapper[4950]: I0318 20:45:02.228107 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564445-wtmkn" event={"ID":"1a66e031-b5dd-49e5-91e9-35ea7071c207","Type":"ContainerDied","Data":"db2c55f91c2c35b9c837ceca97531e7241f2e3130c5bc62b34d7f2b9a6b1b649"} Mar 18 20:45:03 crc kubenswrapper[4950]: I0318 20:45:03.661758 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564445-wtmkn" Mar 18 20:45:03 crc kubenswrapper[4950]: I0318 20:45:03.777032 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a66e031-b5dd-49e5-91e9-35ea7071c207-secret-volume\") pod \"1a66e031-b5dd-49e5-91e9-35ea7071c207\" (UID: \"1a66e031-b5dd-49e5-91e9-35ea7071c207\") " Mar 18 20:45:03 crc kubenswrapper[4950]: I0318 20:45:03.777085 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a66e031-b5dd-49e5-91e9-35ea7071c207-config-volume\") pod \"1a66e031-b5dd-49e5-91e9-35ea7071c207\" (UID: \"1a66e031-b5dd-49e5-91e9-35ea7071c207\") " Mar 18 20:45:03 crc kubenswrapper[4950]: I0318 20:45:03.777217 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jtgfl\" (UniqueName: \"kubernetes.io/projected/1a66e031-b5dd-49e5-91e9-35ea7071c207-kube-api-access-jtgfl\") pod \"1a66e031-b5dd-49e5-91e9-35ea7071c207\" (UID: \"1a66e031-b5dd-49e5-91e9-35ea7071c207\") " Mar 18 20:45:03 crc kubenswrapper[4950]: I0318 20:45:03.777810 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a66e031-b5dd-49e5-91e9-35ea7071c207-config-volume" (OuterVolumeSpecName: "config-volume") pod "1a66e031-b5dd-49e5-91e9-35ea7071c207" (UID: "1a66e031-b5dd-49e5-91e9-35ea7071c207"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:45:03 crc kubenswrapper[4950]: I0318 20:45:03.783665 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a66e031-b5dd-49e5-91e9-35ea7071c207-kube-api-access-jtgfl" (OuterVolumeSpecName: "kube-api-access-jtgfl") pod "1a66e031-b5dd-49e5-91e9-35ea7071c207" (UID: "1a66e031-b5dd-49e5-91e9-35ea7071c207"). InnerVolumeSpecName "kube-api-access-jtgfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:45:03 crc kubenswrapper[4950]: I0318 20:45:03.784917 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a66e031-b5dd-49e5-91e9-35ea7071c207-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1a66e031-b5dd-49e5-91e9-35ea7071c207" (UID: "1a66e031-b5dd-49e5-91e9-35ea7071c207"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:45:03 crc kubenswrapper[4950]: I0318 20:45:03.879148 4950 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a66e031-b5dd-49e5-91e9-35ea7071c207-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 18 20:45:03 crc kubenswrapper[4950]: I0318 20:45:03.879796 4950 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a66e031-b5dd-49e5-91e9-35ea7071c207-config-volume\") on node \"crc\" DevicePath \"\"" Mar 18 20:45:03 crc kubenswrapper[4950]: I0318 20:45:03.879822 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jtgfl\" (UniqueName: \"kubernetes.io/projected/1a66e031-b5dd-49e5-91e9-35ea7071c207-kube-api-access-jtgfl\") on node \"crc\" DevicePath \"\"" Mar 18 20:45:04 crc kubenswrapper[4950]: I0318 20:45:04.245520 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564445-wtmkn" event={"ID":"1a66e031-b5dd-49e5-91e9-35ea7071c207","Type":"ContainerDied","Data":"582f76d7f32419c06f8ad21d80aed60b641b411c5a73b48bb7437cdff00bce95"} Mar 18 20:45:04 crc kubenswrapper[4950]: I0318 20:45:04.245557 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="582f76d7f32419c06f8ad21d80aed60b641b411c5a73b48bb7437cdff00bce95" Mar 18 20:45:04 crc kubenswrapper[4950]: I0318 20:45:04.245588 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564445-wtmkn" Mar 18 20:45:04 crc kubenswrapper[4950]: I0318 20:45:04.319101 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564400-d9vqv"] Mar 18 20:45:04 crc kubenswrapper[4950]: I0318 20:45:04.332672 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564400-d9vqv"] Mar 18 20:45:05 crc kubenswrapper[4950]: I0318 20:45:05.494270 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="164efb7c-0b83-42ed-8d30-acc8a1ff1794" path="/var/lib/kubelet/pods/164efb7c-0b83-42ed-8d30-acc8a1ff1794/volumes" Mar 18 20:45:28 crc kubenswrapper[4950]: I0318 20:45:28.434396 4950 generic.go:334] "Generic (PLEG): container finished" podID="8230645b-e426-4c0e-a28b-565ba98d18a1" containerID="55881ed90e1103f82e4b8dd127583137efe45648c960f1fba288c8cec343e9d4" exitCode=0 Mar 18 20:45:28 crc kubenswrapper[4950]: I0318 20:45:28.434995 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" event={"ID":"8230645b-e426-4c0e-a28b-565ba98d18a1","Type":"ContainerDied","Data":"55881ed90e1103f82e4b8dd127583137efe45648c960f1fba288c8cec343e9d4"} Mar 18 20:45:29 crc kubenswrapper[4950]: I0318 20:45:29.825697 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" Mar 18 20:45:29 crc kubenswrapper[4950]: I0318 20:45:29.961016 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-bootstrap-combined-ca-bundle\") pod \"8230645b-e426-4c0e-a28b-565ba98d18a1\" (UID: \"8230645b-e426-4c0e-a28b-565ba98d18a1\") " Mar 18 20:45:29 crc kubenswrapper[4950]: I0318 20:45:29.961106 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-ssh-key-openstack-edpm-ipam\") pod \"8230645b-e426-4c0e-a28b-565ba98d18a1\" (UID: \"8230645b-e426-4c0e-a28b-565ba98d18a1\") " Mar 18 20:45:29 crc kubenswrapper[4950]: I0318 20:45:29.961223 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-inventory\") pod \"8230645b-e426-4c0e-a28b-565ba98d18a1\" (UID: \"8230645b-e426-4c0e-a28b-565ba98d18a1\") " Mar 18 20:45:29 crc kubenswrapper[4950]: I0318 20:45:29.961328 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdjkr\" (UniqueName: \"kubernetes.io/projected/8230645b-e426-4c0e-a28b-565ba98d18a1-kube-api-access-sdjkr\") pod \"8230645b-e426-4c0e-a28b-565ba98d18a1\" (UID: \"8230645b-e426-4c0e-a28b-565ba98d18a1\") " Mar 18 20:45:29 crc kubenswrapper[4950]: I0318 20:45:29.961403 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-ceph\") pod \"8230645b-e426-4c0e-a28b-565ba98d18a1\" (UID: \"8230645b-e426-4c0e-a28b-565ba98d18a1\") " Mar 18 20:45:29 crc kubenswrapper[4950]: I0318 20:45:29.967595 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8230645b-e426-4c0e-a28b-565ba98d18a1-kube-api-access-sdjkr" (OuterVolumeSpecName: "kube-api-access-sdjkr") pod "8230645b-e426-4c0e-a28b-565ba98d18a1" (UID: "8230645b-e426-4c0e-a28b-565ba98d18a1"). InnerVolumeSpecName "kube-api-access-sdjkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:45:29 crc kubenswrapper[4950]: I0318 20:45:29.967644 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-ceph" (OuterVolumeSpecName: "ceph") pod "8230645b-e426-4c0e-a28b-565ba98d18a1" (UID: "8230645b-e426-4c0e-a28b-565ba98d18a1"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:45:29 crc kubenswrapper[4950]: I0318 20:45:29.972790 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "8230645b-e426-4c0e-a28b-565ba98d18a1" (UID: "8230645b-e426-4c0e-a28b-565ba98d18a1"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:45:29 crc kubenswrapper[4950]: I0318 20:45:29.993086 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-inventory" (OuterVolumeSpecName: "inventory") pod "8230645b-e426-4c0e-a28b-565ba98d18a1" (UID: "8230645b-e426-4c0e-a28b-565ba98d18a1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:45:29 crc kubenswrapper[4950]: I0318 20:45:29.995541 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8230645b-e426-4c0e-a28b-565ba98d18a1" (UID: "8230645b-e426-4c0e-a28b-565ba98d18a1"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.063613 4950 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.063647 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.063660 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.063671 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdjkr\" (UniqueName: \"kubernetes.io/projected/8230645b-e426-4c0e-a28b-565ba98d18a1-kube-api-access-sdjkr\") on node \"crc\" DevicePath \"\"" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.063684 4950 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8230645b-e426-4c0e-a28b-565ba98d18a1-ceph\") on node \"crc\" DevicePath \"\"" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.452327 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" event={"ID":"8230645b-e426-4c0e-a28b-565ba98d18a1","Type":"ContainerDied","Data":"5d6f999c3707f06a8af2371ead063e1219e2953d72f2dbfda2236a1ff8637bc1"} Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.452846 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d6f999c3707f06a8af2371ead063e1219e2953d72f2dbfda2236a1ff8637bc1" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.452464 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.592302 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q"] Mar 18 20:45:30 crc kubenswrapper[4950]: E0318 20:45:30.592721 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a66e031-b5dd-49e5-91e9-35ea7071c207" containerName="collect-profiles" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.592740 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a66e031-b5dd-49e5-91e9-35ea7071c207" containerName="collect-profiles" Mar 18 20:45:30 crc kubenswrapper[4950]: E0318 20:45:30.592781 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8230645b-e426-4c0e-a28b-565ba98d18a1" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.592788 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="8230645b-e426-4c0e-a28b-565ba98d18a1" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.592943 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="8230645b-e426-4c0e-a28b-565ba98d18a1" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.592958 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a66e031-b5dd-49e5-91e9-35ea7071c207" containerName="collect-profiles" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.593571 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.596663 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.596968 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.597121 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.597272 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.597143 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.614625 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q"] Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.773722 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4dc08cb-4f1e-4430-b522-62f890be4336-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q\" (UID: \"f4dc08cb-4f1e-4430-b522-62f890be4336\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.774120 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f4dc08cb-4f1e-4430-b522-62f890be4336-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q\" (UID: \"f4dc08cb-4f1e-4430-b522-62f890be4336\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.774337 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f4dc08cb-4f1e-4430-b522-62f890be4336-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q\" (UID: \"f4dc08cb-4f1e-4430-b522-62f890be4336\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.774508 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzcvb\" (UniqueName: \"kubernetes.io/projected/f4dc08cb-4f1e-4430-b522-62f890be4336-kube-api-access-tzcvb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q\" (UID: \"f4dc08cb-4f1e-4430-b522-62f890be4336\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.876388 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4dc08cb-4f1e-4430-b522-62f890be4336-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q\" (UID: \"f4dc08cb-4f1e-4430-b522-62f890be4336\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.876643 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f4dc08cb-4f1e-4430-b522-62f890be4336-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q\" (UID: \"f4dc08cb-4f1e-4430-b522-62f890be4336\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.877044 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f4dc08cb-4f1e-4430-b522-62f890be4336-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q\" (UID: \"f4dc08cb-4f1e-4430-b522-62f890be4336\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.877109 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzcvb\" (UniqueName: \"kubernetes.io/projected/f4dc08cb-4f1e-4430-b522-62f890be4336-kube-api-access-tzcvb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q\" (UID: \"f4dc08cb-4f1e-4430-b522-62f890be4336\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.880499 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4dc08cb-4f1e-4430-b522-62f890be4336-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q\" (UID: \"f4dc08cb-4f1e-4430-b522-62f890be4336\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.880993 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f4dc08cb-4f1e-4430-b522-62f890be4336-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q\" (UID: \"f4dc08cb-4f1e-4430-b522-62f890be4336\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.896733 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f4dc08cb-4f1e-4430-b522-62f890be4336-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q\" (UID: \"f4dc08cb-4f1e-4430-b522-62f890be4336\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.899291 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzcvb\" (UniqueName: \"kubernetes.io/projected/f4dc08cb-4f1e-4430-b522-62f890be4336-kube-api-access-tzcvb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q\" (UID: \"f4dc08cb-4f1e-4430-b522-62f890be4336\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q" Mar 18 20:45:30 crc kubenswrapper[4950]: I0318 20:45:30.928694 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q" Mar 18 20:45:31 crc kubenswrapper[4950]: I0318 20:45:31.442318 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q"] Mar 18 20:45:31 crc kubenswrapper[4950]: I0318 20:45:31.459184 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q" event={"ID":"f4dc08cb-4f1e-4430-b522-62f890be4336","Type":"ContainerStarted","Data":"c28e215b8a2f56066f4dffb46f4bc63fb4c495d51681131842233db473e1d7f4"} Mar 18 20:45:32 crc kubenswrapper[4950]: I0318 20:45:32.468352 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q" event={"ID":"f4dc08cb-4f1e-4430-b522-62f890be4336","Type":"ContainerStarted","Data":"620fd1652e6409628d7e8a3e1792ce5f7c6deb90433a60b856ccc29b775a5faf"} Mar 18 20:45:32 crc kubenswrapper[4950]: I0318 20:45:32.486981 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q" podStartSLOduration=2.03439308 podStartE2EDuration="2.486958948s" podCreationTimestamp="2026-03-18 20:45:30 +0000 UTC" firstStartedPulling="2026-03-18 20:45:31.453565982 +0000 UTC m=+2344.694407850" lastFinishedPulling="2026-03-18 20:45:31.90613185 +0000 UTC m=+2345.146973718" observedRunningTime="2026-03-18 20:45:32.485470799 +0000 UTC m=+2345.726312667" watchObservedRunningTime="2026-03-18 20:45:32.486958948 +0000 UTC m=+2345.727800816" Mar 18 20:45:49 crc kubenswrapper[4950]: I0318 20:45:49.983831 4950 scope.go:117] "RemoveContainer" containerID="25e62c249dbad8da15635226eb6e8947d89177ad45c9ad3022cf2b7827bfcaff" Mar 18 20:45:57 crc kubenswrapper[4950]: I0318 20:45:57.672932 4950 generic.go:334] "Generic (PLEG): container finished" podID="f4dc08cb-4f1e-4430-b522-62f890be4336" containerID="620fd1652e6409628d7e8a3e1792ce5f7c6deb90433a60b856ccc29b775a5faf" exitCode=0 Mar 18 20:45:57 crc kubenswrapper[4950]: I0318 20:45:57.673070 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q" event={"ID":"f4dc08cb-4f1e-4430-b522-62f890be4336","Type":"ContainerDied","Data":"620fd1652e6409628d7e8a3e1792ce5f7c6deb90433a60b856ccc29b775a5faf"} Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.251854 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.360078 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4dc08cb-4f1e-4430-b522-62f890be4336-inventory\") pod \"f4dc08cb-4f1e-4430-b522-62f890be4336\" (UID: \"f4dc08cb-4f1e-4430-b522-62f890be4336\") " Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.360224 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f4dc08cb-4f1e-4430-b522-62f890be4336-ceph\") pod \"f4dc08cb-4f1e-4430-b522-62f890be4336\" (UID: \"f4dc08cb-4f1e-4430-b522-62f890be4336\") " Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.360304 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f4dc08cb-4f1e-4430-b522-62f890be4336-ssh-key-openstack-edpm-ipam\") pod \"f4dc08cb-4f1e-4430-b522-62f890be4336\" (UID: \"f4dc08cb-4f1e-4430-b522-62f890be4336\") " Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.360338 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzcvb\" (UniqueName: \"kubernetes.io/projected/f4dc08cb-4f1e-4430-b522-62f890be4336-kube-api-access-tzcvb\") pod \"f4dc08cb-4f1e-4430-b522-62f890be4336\" (UID: \"f4dc08cb-4f1e-4430-b522-62f890be4336\") " Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.370022 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4dc08cb-4f1e-4430-b522-62f890be4336-ceph" (OuterVolumeSpecName: "ceph") pod "f4dc08cb-4f1e-4430-b522-62f890be4336" (UID: "f4dc08cb-4f1e-4430-b522-62f890be4336"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.377625 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4dc08cb-4f1e-4430-b522-62f890be4336-kube-api-access-tzcvb" (OuterVolumeSpecName: "kube-api-access-tzcvb") pod "f4dc08cb-4f1e-4430-b522-62f890be4336" (UID: "f4dc08cb-4f1e-4430-b522-62f890be4336"). InnerVolumeSpecName "kube-api-access-tzcvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.394517 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4dc08cb-4f1e-4430-b522-62f890be4336-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f4dc08cb-4f1e-4430-b522-62f890be4336" (UID: "f4dc08cb-4f1e-4430-b522-62f890be4336"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.421597 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4dc08cb-4f1e-4430-b522-62f890be4336-inventory" (OuterVolumeSpecName: "inventory") pod "f4dc08cb-4f1e-4430-b522-62f890be4336" (UID: "f4dc08cb-4f1e-4430-b522-62f890be4336"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.463451 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f4dc08cb-4f1e-4430-b522-62f890be4336-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.463513 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzcvb\" (UniqueName: \"kubernetes.io/projected/f4dc08cb-4f1e-4430-b522-62f890be4336-kube-api-access-tzcvb\") on node \"crc\" DevicePath \"\"" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.463531 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4dc08cb-4f1e-4430-b522-62f890be4336-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.463545 4950 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f4dc08cb-4f1e-4430-b522-62f890be4336-ceph\") on node \"crc\" DevicePath \"\"" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.697846 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q" event={"ID":"f4dc08cb-4f1e-4430-b522-62f890be4336","Type":"ContainerDied","Data":"c28e215b8a2f56066f4dffb46f4bc63fb4c495d51681131842233db473e1d7f4"} Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.698127 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c28e215b8a2f56066f4dffb46f4bc63fb4c495d51681131842233db473e1d7f4" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.697915 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.798217 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw"] Mar 18 20:45:59 crc kubenswrapper[4950]: E0318 20:45:59.798581 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4dc08cb-4f1e-4430-b522-62f890be4336" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.798597 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4dc08cb-4f1e-4430-b522-62f890be4336" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.798751 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4dc08cb-4f1e-4430-b522-62f890be4336" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.799323 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.801893 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.802016 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.802498 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.802979 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.803171 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.812506 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw"] Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.871598 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd54300e-2b25-4bb8-b649-ac546072be73-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw\" (UID: \"dd54300e-2b25-4bb8-b649-ac546072be73\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.871664 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k5ff\" (UniqueName: \"kubernetes.io/projected/dd54300e-2b25-4bb8-b649-ac546072be73-kube-api-access-7k5ff\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw\" (UID: \"dd54300e-2b25-4bb8-b649-ac546072be73\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.871863 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dd54300e-2b25-4bb8-b649-ac546072be73-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw\" (UID: \"dd54300e-2b25-4bb8-b649-ac546072be73\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.872022 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dd54300e-2b25-4bb8-b649-ac546072be73-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw\" (UID: \"dd54300e-2b25-4bb8-b649-ac546072be73\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.973145 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k5ff\" (UniqueName: \"kubernetes.io/projected/dd54300e-2b25-4bb8-b649-ac546072be73-kube-api-access-7k5ff\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw\" (UID: \"dd54300e-2b25-4bb8-b649-ac546072be73\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.973220 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dd54300e-2b25-4bb8-b649-ac546072be73-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw\" (UID: \"dd54300e-2b25-4bb8-b649-ac546072be73\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.973322 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dd54300e-2b25-4bb8-b649-ac546072be73-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw\" (UID: \"dd54300e-2b25-4bb8-b649-ac546072be73\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.973358 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd54300e-2b25-4bb8-b649-ac546072be73-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw\" (UID: \"dd54300e-2b25-4bb8-b649-ac546072be73\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.976676 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd54300e-2b25-4bb8-b649-ac546072be73-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw\" (UID: \"dd54300e-2b25-4bb8-b649-ac546072be73\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.976896 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dd54300e-2b25-4bb8-b649-ac546072be73-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw\" (UID: \"dd54300e-2b25-4bb8-b649-ac546072be73\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.977236 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dd54300e-2b25-4bb8-b649-ac546072be73-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw\" (UID: \"dd54300e-2b25-4bb8-b649-ac546072be73\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw" Mar 18 20:45:59 crc kubenswrapper[4950]: I0318 20:45:59.989052 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k5ff\" (UniqueName: \"kubernetes.io/projected/dd54300e-2b25-4bb8-b649-ac546072be73-kube-api-access-7k5ff\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw\" (UID: \"dd54300e-2b25-4bb8-b649-ac546072be73\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw" Mar 18 20:46:00 crc kubenswrapper[4950]: I0318 20:46:00.117491 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw" Mar 18 20:46:00 crc kubenswrapper[4950]: I0318 20:46:00.145814 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564446-zpc8l"] Mar 18 20:46:00 crc kubenswrapper[4950]: I0318 20:46:00.147245 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564446-zpc8l" Mar 18 20:46:00 crc kubenswrapper[4950]: I0318 20:46:00.152241 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:46:00 crc kubenswrapper[4950]: I0318 20:46:00.152700 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:46:00 crc kubenswrapper[4950]: I0318 20:46:00.152809 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:46:00 crc kubenswrapper[4950]: I0318 20:46:00.182727 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564446-zpc8l"] Mar 18 20:46:00 crc kubenswrapper[4950]: I0318 20:46:00.278110 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4rfs\" (UniqueName: \"kubernetes.io/projected/0ab7b23e-8e83-4f5a-a7dd-e7b2a6ffcb94-kube-api-access-r4rfs\") pod \"auto-csr-approver-29564446-zpc8l\" (UID: \"0ab7b23e-8e83-4f5a-a7dd-e7b2a6ffcb94\") " pod="openshift-infra/auto-csr-approver-29564446-zpc8l" Mar 18 20:46:00 crc kubenswrapper[4950]: I0318 20:46:00.379966 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4rfs\" (UniqueName: \"kubernetes.io/projected/0ab7b23e-8e83-4f5a-a7dd-e7b2a6ffcb94-kube-api-access-r4rfs\") pod \"auto-csr-approver-29564446-zpc8l\" (UID: \"0ab7b23e-8e83-4f5a-a7dd-e7b2a6ffcb94\") " pod="openshift-infra/auto-csr-approver-29564446-zpc8l" Mar 18 20:46:00 crc kubenswrapper[4950]: I0318 20:46:00.401805 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4rfs\" (UniqueName: \"kubernetes.io/projected/0ab7b23e-8e83-4f5a-a7dd-e7b2a6ffcb94-kube-api-access-r4rfs\") pod \"auto-csr-approver-29564446-zpc8l\" (UID: \"0ab7b23e-8e83-4f5a-a7dd-e7b2a6ffcb94\") " pod="openshift-infra/auto-csr-approver-29564446-zpc8l" Mar 18 20:46:00 crc kubenswrapper[4950]: I0318 20:46:00.568981 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564446-zpc8l" Mar 18 20:46:00 crc kubenswrapper[4950]: I0318 20:46:00.702685 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw"] Mar 18 20:46:01 crc kubenswrapper[4950]: I0318 20:46:01.120017 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564446-zpc8l"] Mar 18 20:46:01 crc kubenswrapper[4950]: I0318 20:46:01.713751 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564446-zpc8l" event={"ID":"0ab7b23e-8e83-4f5a-a7dd-e7b2a6ffcb94","Type":"ContainerStarted","Data":"07e8c4b64cb6bb42cb7a734cfb27b18e4b4580e933546746d0009a3974055e83"} Mar 18 20:46:01 crc kubenswrapper[4950]: I0318 20:46:01.720723 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw" event={"ID":"dd54300e-2b25-4bb8-b649-ac546072be73","Type":"ContainerStarted","Data":"64bf24f105aea1113cb340f0eb525873e324a943f28dd5929e2ebac0cc298889"} Mar 18 20:46:01 crc kubenswrapper[4950]: I0318 20:46:01.720751 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw" event={"ID":"dd54300e-2b25-4bb8-b649-ac546072be73","Type":"ContainerStarted","Data":"49b553f17fedb7e668bff30f7bc66c00a97f680e1f1e8b8a1f3e4475415da2e6"} Mar 18 20:46:01 crc kubenswrapper[4950]: I0318 20:46:01.760630 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw" podStartSLOduration=2.3308268 podStartE2EDuration="2.760613569s" podCreationTimestamp="2026-03-18 20:45:59 +0000 UTC" firstStartedPulling="2026-03-18 20:46:00.722499231 +0000 UTC m=+2373.963341099" lastFinishedPulling="2026-03-18 20:46:01.15228599 +0000 UTC m=+2374.393127868" observedRunningTime="2026-03-18 20:46:01.75302647 +0000 UTC m=+2374.993868338" watchObservedRunningTime="2026-03-18 20:46:01.760613569 +0000 UTC m=+2375.001455437" Mar 18 20:46:03 crc kubenswrapper[4950]: I0318 20:46:03.742105 4950 generic.go:334] "Generic (PLEG): container finished" podID="0ab7b23e-8e83-4f5a-a7dd-e7b2a6ffcb94" containerID="cd20209c9a713c1281bce5860c56014ea0a7da9d820a225d3a2b432bfa9edb30" exitCode=0 Mar 18 20:46:03 crc kubenswrapper[4950]: I0318 20:46:03.742253 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564446-zpc8l" event={"ID":"0ab7b23e-8e83-4f5a-a7dd-e7b2a6ffcb94","Type":"ContainerDied","Data":"cd20209c9a713c1281bce5860c56014ea0a7da9d820a225d3a2b432bfa9edb30"} Mar 18 20:46:05 crc kubenswrapper[4950]: I0318 20:46:05.119765 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564446-zpc8l" Mar 18 20:46:05 crc kubenswrapper[4950]: I0318 20:46:05.287096 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4rfs\" (UniqueName: \"kubernetes.io/projected/0ab7b23e-8e83-4f5a-a7dd-e7b2a6ffcb94-kube-api-access-r4rfs\") pod \"0ab7b23e-8e83-4f5a-a7dd-e7b2a6ffcb94\" (UID: \"0ab7b23e-8e83-4f5a-a7dd-e7b2a6ffcb94\") " Mar 18 20:46:05 crc kubenswrapper[4950]: I0318 20:46:05.292873 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ab7b23e-8e83-4f5a-a7dd-e7b2a6ffcb94-kube-api-access-r4rfs" (OuterVolumeSpecName: "kube-api-access-r4rfs") pod "0ab7b23e-8e83-4f5a-a7dd-e7b2a6ffcb94" (UID: "0ab7b23e-8e83-4f5a-a7dd-e7b2a6ffcb94"). InnerVolumeSpecName "kube-api-access-r4rfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:46:05 crc kubenswrapper[4950]: I0318 20:46:05.388968 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4rfs\" (UniqueName: \"kubernetes.io/projected/0ab7b23e-8e83-4f5a-a7dd-e7b2a6ffcb94-kube-api-access-r4rfs\") on node \"crc\" DevicePath \"\"" Mar 18 20:46:05 crc kubenswrapper[4950]: I0318 20:46:05.766273 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564446-zpc8l" event={"ID":"0ab7b23e-8e83-4f5a-a7dd-e7b2a6ffcb94","Type":"ContainerDied","Data":"07e8c4b64cb6bb42cb7a734cfb27b18e4b4580e933546746d0009a3974055e83"} Mar 18 20:46:05 crc kubenswrapper[4950]: I0318 20:46:05.766334 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07e8c4b64cb6bb42cb7a734cfb27b18e4b4580e933546746d0009a3974055e83" Mar 18 20:46:05 crc kubenswrapper[4950]: I0318 20:46:05.766351 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564446-zpc8l" Mar 18 20:46:06 crc kubenswrapper[4950]: I0318 20:46:06.194559 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564440-jqvbk"] Mar 18 20:46:06 crc kubenswrapper[4950]: I0318 20:46:06.200694 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564440-jqvbk"] Mar 18 20:46:06 crc kubenswrapper[4950]: I0318 20:46:06.782641 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw" event={"ID":"dd54300e-2b25-4bb8-b649-ac546072be73","Type":"ContainerDied","Data":"64bf24f105aea1113cb340f0eb525873e324a943f28dd5929e2ebac0cc298889"} Mar 18 20:46:06 crc kubenswrapper[4950]: I0318 20:46:06.782675 4950 generic.go:334] "Generic (PLEG): container finished" podID="dd54300e-2b25-4bb8-b649-ac546072be73" containerID="64bf24f105aea1113cb340f0eb525873e324a943f28dd5929e2ebac0cc298889" exitCode=0 Mar 18 20:46:07 crc kubenswrapper[4950]: I0318 20:46:07.513150 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8149017b-bc82-441c-8ede-ec04b1ab0614" path="/var/lib/kubelet/pods/8149017b-bc82-441c-8ede-ec04b1ab0614/volumes" Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.221257 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw" Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.267576 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7k5ff\" (UniqueName: \"kubernetes.io/projected/dd54300e-2b25-4bb8-b649-ac546072be73-kube-api-access-7k5ff\") pod \"dd54300e-2b25-4bb8-b649-ac546072be73\" (UID: \"dd54300e-2b25-4bb8-b649-ac546072be73\") " Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.267697 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd54300e-2b25-4bb8-b649-ac546072be73-inventory\") pod \"dd54300e-2b25-4bb8-b649-ac546072be73\" (UID: \"dd54300e-2b25-4bb8-b649-ac546072be73\") " Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.267730 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dd54300e-2b25-4bb8-b649-ac546072be73-ceph\") pod \"dd54300e-2b25-4bb8-b649-ac546072be73\" (UID: \"dd54300e-2b25-4bb8-b649-ac546072be73\") " Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.267761 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dd54300e-2b25-4bb8-b649-ac546072be73-ssh-key-openstack-edpm-ipam\") pod \"dd54300e-2b25-4bb8-b649-ac546072be73\" (UID: \"dd54300e-2b25-4bb8-b649-ac546072be73\") " Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.274920 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd54300e-2b25-4bb8-b649-ac546072be73-ceph" (OuterVolumeSpecName: "ceph") pod "dd54300e-2b25-4bb8-b649-ac546072be73" (UID: "dd54300e-2b25-4bb8-b649-ac546072be73"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.278154 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd54300e-2b25-4bb8-b649-ac546072be73-kube-api-access-7k5ff" (OuterVolumeSpecName: "kube-api-access-7k5ff") pod "dd54300e-2b25-4bb8-b649-ac546072be73" (UID: "dd54300e-2b25-4bb8-b649-ac546072be73"). InnerVolumeSpecName "kube-api-access-7k5ff". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.298567 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd54300e-2b25-4bb8-b649-ac546072be73-inventory" (OuterVolumeSpecName: "inventory") pod "dd54300e-2b25-4bb8-b649-ac546072be73" (UID: "dd54300e-2b25-4bb8-b649-ac546072be73"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.302918 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd54300e-2b25-4bb8-b649-ac546072be73-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "dd54300e-2b25-4bb8-b649-ac546072be73" (UID: "dd54300e-2b25-4bb8-b649-ac546072be73"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.369994 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd54300e-2b25-4bb8-b649-ac546072be73-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.370400 4950 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dd54300e-2b25-4bb8-b649-ac546072be73-ceph\") on node \"crc\" DevicePath \"\"" Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.370460 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dd54300e-2b25-4bb8-b649-ac546072be73-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.370494 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7k5ff\" (UniqueName: \"kubernetes.io/projected/dd54300e-2b25-4bb8-b649-ac546072be73-kube-api-access-7k5ff\") on node \"crc\" DevicePath \"\"" Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.805976 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw" event={"ID":"dd54300e-2b25-4bb8-b649-ac546072be73","Type":"ContainerDied","Data":"49b553f17fedb7e668bff30f7bc66c00a97f680e1f1e8b8a1f3e4475415da2e6"} Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.806607 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49b553f17fedb7e668bff30f7bc66c00a97f680e1f1e8b8a1f3e4475415da2e6" Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.806074 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw" Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.923636 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff"] Mar 18 20:46:08 crc kubenswrapper[4950]: E0318 20:46:08.924027 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ab7b23e-8e83-4f5a-a7dd-e7b2a6ffcb94" containerName="oc" Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.924046 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ab7b23e-8e83-4f5a-a7dd-e7b2a6ffcb94" containerName="oc" Mar 18 20:46:08 crc kubenswrapper[4950]: E0318 20:46:08.924070 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd54300e-2b25-4bb8-b649-ac546072be73" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.924081 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd54300e-2b25-4bb8-b649-ac546072be73" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.924338 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ab7b23e-8e83-4f5a-a7dd-e7b2a6ffcb94" containerName="oc" Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.924366 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd54300e-2b25-4bb8-b649-ac546072be73" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.925072 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff" Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.927704 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.938709 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff"] Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.941861 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.942212 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.942381 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:46:08 crc kubenswrapper[4950]: I0318 20:46:08.942534 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:46:09 crc kubenswrapper[4950]: I0318 20:46:08.983836 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djw9s\" (UniqueName: \"kubernetes.io/projected/29a51c30-1655-4c6c-9e74-9b8797196361-kube-api-access-djw9s\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-w5fff\" (UID: \"29a51c30-1655-4c6c-9e74-9b8797196361\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff" Mar 18 20:46:09 crc kubenswrapper[4950]: I0318 20:46:08.983900 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29a51c30-1655-4c6c-9e74-9b8797196361-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-w5fff\" (UID: \"29a51c30-1655-4c6c-9e74-9b8797196361\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff" Mar 18 20:46:09 crc kubenswrapper[4950]: I0318 20:46:08.983955 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/29a51c30-1655-4c6c-9e74-9b8797196361-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-w5fff\" (UID: \"29a51c30-1655-4c6c-9e74-9b8797196361\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff" Mar 18 20:46:09 crc kubenswrapper[4950]: I0318 20:46:08.984035 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/29a51c30-1655-4c6c-9e74-9b8797196361-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-w5fff\" (UID: \"29a51c30-1655-4c6c-9e74-9b8797196361\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff" Mar 18 20:46:09 crc kubenswrapper[4950]: I0318 20:46:09.085772 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/29a51c30-1655-4c6c-9e74-9b8797196361-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-w5fff\" (UID: \"29a51c30-1655-4c6c-9e74-9b8797196361\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff" Mar 18 20:46:09 crc kubenswrapper[4950]: I0318 20:46:09.085892 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/29a51c30-1655-4c6c-9e74-9b8797196361-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-w5fff\" (UID: \"29a51c30-1655-4c6c-9e74-9b8797196361\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff" Mar 18 20:46:09 crc kubenswrapper[4950]: I0318 20:46:09.085936 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djw9s\" (UniqueName: \"kubernetes.io/projected/29a51c30-1655-4c6c-9e74-9b8797196361-kube-api-access-djw9s\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-w5fff\" (UID: \"29a51c30-1655-4c6c-9e74-9b8797196361\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff" Mar 18 20:46:09 crc kubenswrapper[4950]: I0318 20:46:09.085972 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29a51c30-1655-4c6c-9e74-9b8797196361-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-w5fff\" (UID: \"29a51c30-1655-4c6c-9e74-9b8797196361\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff" Mar 18 20:46:09 crc kubenswrapper[4950]: I0318 20:46:09.090019 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29a51c30-1655-4c6c-9e74-9b8797196361-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-w5fff\" (UID: \"29a51c30-1655-4c6c-9e74-9b8797196361\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff" Mar 18 20:46:09 crc kubenswrapper[4950]: I0318 20:46:09.090084 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/29a51c30-1655-4c6c-9e74-9b8797196361-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-w5fff\" (UID: \"29a51c30-1655-4c6c-9e74-9b8797196361\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff" Mar 18 20:46:09 crc kubenswrapper[4950]: I0318 20:46:09.091796 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/29a51c30-1655-4c6c-9e74-9b8797196361-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-w5fff\" (UID: \"29a51c30-1655-4c6c-9e74-9b8797196361\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff" Mar 18 20:46:09 crc kubenswrapper[4950]: I0318 20:46:09.106454 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djw9s\" (UniqueName: \"kubernetes.io/projected/29a51c30-1655-4c6c-9e74-9b8797196361-kube-api-access-djw9s\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-w5fff\" (UID: \"29a51c30-1655-4c6c-9e74-9b8797196361\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff" Mar 18 20:46:09 crc kubenswrapper[4950]: I0318 20:46:09.249230 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff" Mar 18 20:46:09 crc kubenswrapper[4950]: I0318 20:46:09.829725 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff"] Mar 18 20:46:10 crc kubenswrapper[4950]: I0318 20:46:10.825555 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff" event={"ID":"29a51c30-1655-4c6c-9e74-9b8797196361","Type":"ContainerStarted","Data":"c25e999b0101ba4b46ab821d36794e7d8aa71128f26b97ddecfa2fb149d9cbda"} Mar 18 20:46:10 crc kubenswrapper[4950]: I0318 20:46:10.825902 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff" event={"ID":"29a51c30-1655-4c6c-9e74-9b8797196361","Type":"ContainerStarted","Data":"ac9ad19adb1081bf283f4095d188bcb3087d3f05261d1eb955fc2108bebe54a8"} Mar 18 20:46:10 crc kubenswrapper[4950]: I0318 20:46:10.847834 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff" podStartSLOduration=2.216661226 podStartE2EDuration="2.847819255s" podCreationTimestamp="2026-03-18 20:46:08 +0000 UTC" firstStartedPulling="2026-03-18 20:46:09.837305381 +0000 UTC m=+2383.078147249" lastFinishedPulling="2026-03-18 20:46:10.46846342 +0000 UTC m=+2383.709305278" observedRunningTime="2026-03-18 20:46:10.843778979 +0000 UTC m=+2384.084620847" watchObservedRunningTime="2026-03-18 20:46:10.847819255 +0000 UTC m=+2384.088661113" Mar 18 20:46:50 crc kubenswrapper[4950]: I0318 20:46:50.047631 4950 scope.go:117] "RemoveContainer" containerID="4c36c400239f08c0485c4f59a33cd5f31b2e05e90ada501a1492b3d74e3af080" Mar 18 20:46:50 crc kubenswrapper[4950]: I0318 20:46:50.149845 4950 generic.go:334] "Generic (PLEG): container finished" podID="29a51c30-1655-4c6c-9e74-9b8797196361" containerID="c25e999b0101ba4b46ab821d36794e7d8aa71128f26b97ddecfa2fb149d9cbda" exitCode=0 Mar 18 20:46:50 crc kubenswrapper[4950]: I0318 20:46:50.150159 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff" event={"ID":"29a51c30-1655-4c6c-9e74-9b8797196361","Type":"ContainerDied","Data":"c25e999b0101ba4b46ab821d36794e7d8aa71128f26b97ddecfa2fb149d9cbda"} Mar 18 20:46:51 crc kubenswrapper[4950]: I0318 20:46:51.623463 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff" Mar 18 20:46:51 crc kubenswrapper[4950]: I0318 20:46:51.704231 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djw9s\" (UniqueName: \"kubernetes.io/projected/29a51c30-1655-4c6c-9e74-9b8797196361-kube-api-access-djw9s\") pod \"29a51c30-1655-4c6c-9e74-9b8797196361\" (UID: \"29a51c30-1655-4c6c-9e74-9b8797196361\") " Mar 18 20:46:51 crc kubenswrapper[4950]: I0318 20:46:51.705235 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/29a51c30-1655-4c6c-9e74-9b8797196361-ssh-key-openstack-edpm-ipam\") pod \"29a51c30-1655-4c6c-9e74-9b8797196361\" (UID: \"29a51c30-1655-4c6c-9e74-9b8797196361\") " Mar 18 20:46:51 crc kubenswrapper[4950]: I0318 20:46:51.705453 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/29a51c30-1655-4c6c-9e74-9b8797196361-ceph\") pod \"29a51c30-1655-4c6c-9e74-9b8797196361\" (UID: \"29a51c30-1655-4c6c-9e74-9b8797196361\") " Mar 18 20:46:51 crc kubenswrapper[4950]: I0318 20:46:51.705764 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29a51c30-1655-4c6c-9e74-9b8797196361-inventory\") pod \"29a51c30-1655-4c6c-9e74-9b8797196361\" (UID: \"29a51c30-1655-4c6c-9e74-9b8797196361\") " Mar 18 20:46:51 crc kubenswrapper[4950]: I0318 20:46:51.710304 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29a51c30-1655-4c6c-9e74-9b8797196361-ceph" (OuterVolumeSpecName: "ceph") pod "29a51c30-1655-4c6c-9e74-9b8797196361" (UID: "29a51c30-1655-4c6c-9e74-9b8797196361"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:46:51 crc kubenswrapper[4950]: I0318 20:46:51.710638 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29a51c30-1655-4c6c-9e74-9b8797196361-kube-api-access-djw9s" (OuterVolumeSpecName: "kube-api-access-djw9s") pod "29a51c30-1655-4c6c-9e74-9b8797196361" (UID: "29a51c30-1655-4c6c-9e74-9b8797196361"). InnerVolumeSpecName "kube-api-access-djw9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:46:51 crc kubenswrapper[4950]: I0318 20:46:51.732990 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29a51c30-1655-4c6c-9e74-9b8797196361-inventory" (OuterVolumeSpecName: "inventory") pod "29a51c30-1655-4c6c-9e74-9b8797196361" (UID: "29a51c30-1655-4c6c-9e74-9b8797196361"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:46:51 crc kubenswrapper[4950]: I0318 20:46:51.744059 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29a51c30-1655-4c6c-9e74-9b8797196361-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "29a51c30-1655-4c6c-9e74-9b8797196361" (UID: "29a51c30-1655-4c6c-9e74-9b8797196361"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:46:51 crc kubenswrapper[4950]: I0318 20:46:51.808097 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/29a51c30-1655-4c6c-9e74-9b8797196361-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:46:51 crc kubenswrapper[4950]: I0318 20:46:51.808137 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djw9s\" (UniqueName: \"kubernetes.io/projected/29a51c30-1655-4c6c-9e74-9b8797196361-kube-api-access-djw9s\") on node \"crc\" DevicePath \"\"" Mar 18 20:46:51 crc kubenswrapper[4950]: I0318 20:46:51.808149 4950 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/29a51c30-1655-4c6c-9e74-9b8797196361-ceph\") on node \"crc\" DevicePath \"\"" Mar 18 20:46:51 crc kubenswrapper[4950]: I0318 20:46:51.808161 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29a51c30-1655-4c6c-9e74-9b8797196361-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.170552 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff" event={"ID":"29a51c30-1655-4c6c-9e74-9b8797196361","Type":"ContainerDied","Data":"ac9ad19adb1081bf283f4095d188bcb3087d3f05261d1eb955fc2108bebe54a8"} Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.170946 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac9ad19adb1081bf283f4095d188bcb3087d3f05261d1eb955fc2108bebe54a8" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.170770 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-w5fff" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.272543 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f"] Mar 18 20:46:52 crc kubenswrapper[4950]: E0318 20:46:52.272991 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a51c30-1655-4c6c-9e74-9b8797196361" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.273012 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a51c30-1655-4c6c-9e74-9b8797196361" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.273241 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="29a51c30-1655-4c6c-9e74-9b8797196361" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.274011 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.278012 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.278303 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.278602 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.279007 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.280303 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.293376 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f"] Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.419087 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e53816d0-c4c0-412f-9568-af0a96c87145-ssh-key-openstack-edpm-ipam\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f\" (UID: \"e53816d0-c4c0-412f-9568-af0a96c87145\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.419144 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4hl4\" (UniqueName: \"kubernetes.io/projected/e53816d0-c4c0-412f-9568-af0a96c87145-kube-api-access-l4hl4\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f\" (UID: \"e53816d0-c4c0-412f-9568-af0a96c87145\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.419219 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e53816d0-c4c0-412f-9568-af0a96c87145-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f\" (UID: \"e53816d0-c4c0-412f-9568-af0a96c87145\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.419385 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e53816d0-c4c0-412f-9568-af0a96c87145-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f\" (UID: \"e53816d0-c4c0-412f-9568-af0a96c87145\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.521021 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e53816d0-c4c0-412f-9568-af0a96c87145-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f\" (UID: \"e53816d0-c4c0-412f-9568-af0a96c87145\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.521130 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e53816d0-c4c0-412f-9568-af0a96c87145-ssh-key-openstack-edpm-ipam\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f\" (UID: \"e53816d0-c4c0-412f-9568-af0a96c87145\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.521152 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4hl4\" (UniqueName: \"kubernetes.io/projected/e53816d0-c4c0-412f-9568-af0a96c87145-kube-api-access-l4hl4\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f\" (UID: \"e53816d0-c4c0-412f-9568-af0a96c87145\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.521175 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e53816d0-c4c0-412f-9568-af0a96c87145-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f\" (UID: \"e53816d0-c4c0-412f-9568-af0a96c87145\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.525848 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e53816d0-c4c0-412f-9568-af0a96c87145-ssh-key-openstack-edpm-ipam\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f\" (UID: \"e53816d0-c4c0-412f-9568-af0a96c87145\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.527125 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e53816d0-c4c0-412f-9568-af0a96c87145-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f\" (UID: \"e53816d0-c4c0-412f-9568-af0a96c87145\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.529106 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e53816d0-c4c0-412f-9568-af0a96c87145-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f\" (UID: \"e53816d0-c4c0-412f-9568-af0a96c87145\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.541781 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4hl4\" (UniqueName: \"kubernetes.io/projected/e53816d0-c4c0-412f-9568-af0a96c87145-kube-api-access-l4hl4\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f\" (UID: \"e53816d0-c4c0-412f-9568-af0a96c87145\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f" Mar 18 20:46:52 crc kubenswrapper[4950]: I0318 20:46:52.592291 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f" Mar 18 20:46:53 crc kubenswrapper[4950]: I0318 20:46:53.094353 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f"] Mar 18 20:46:53 crc kubenswrapper[4950]: I0318 20:46:53.178447 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f" event={"ID":"e53816d0-c4c0-412f-9568-af0a96c87145","Type":"ContainerStarted","Data":"1627851d03ccc8912993e3863de7ada91baee6f44f5ad391cf744b946ab5f849"} Mar 18 20:46:54 crc kubenswrapper[4950]: I0318 20:46:54.187627 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f" event={"ID":"e53816d0-c4c0-412f-9568-af0a96c87145","Type":"ContainerStarted","Data":"d0dfa6f4b9c7f125220535fa3a87368d8da8e108903fd31bb7216641d1c3e063"} Mar 18 20:46:58 crc kubenswrapper[4950]: I0318 20:46:58.219221 4950 generic.go:334] "Generic (PLEG): container finished" podID="e53816d0-c4c0-412f-9568-af0a96c87145" containerID="d0dfa6f4b9c7f125220535fa3a87368d8da8e108903fd31bb7216641d1c3e063" exitCode=0 Mar 18 20:46:58 crc kubenswrapper[4950]: I0318 20:46:58.219288 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f" event={"ID":"e53816d0-c4c0-412f-9568-af0a96c87145","Type":"ContainerDied","Data":"d0dfa6f4b9c7f125220535fa3a87368d8da8e108903fd31bb7216641d1c3e063"} Mar 18 20:46:59 crc kubenswrapper[4950]: I0318 20:46:59.722866 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f" Mar 18 20:46:59 crc kubenswrapper[4950]: I0318 20:46:59.866169 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e53816d0-c4c0-412f-9568-af0a96c87145-ssh-key-openstack-edpm-ipam\") pod \"e53816d0-c4c0-412f-9568-af0a96c87145\" (UID: \"e53816d0-c4c0-412f-9568-af0a96c87145\") " Mar 18 20:46:59 crc kubenswrapper[4950]: I0318 20:46:59.866226 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e53816d0-c4c0-412f-9568-af0a96c87145-ceph\") pod \"e53816d0-c4c0-412f-9568-af0a96c87145\" (UID: \"e53816d0-c4c0-412f-9568-af0a96c87145\") " Mar 18 20:46:59 crc kubenswrapper[4950]: I0318 20:46:59.866300 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e53816d0-c4c0-412f-9568-af0a96c87145-inventory\") pod \"e53816d0-c4c0-412f-9568-af0a96c87145\" (UID: \"e53816d0-c4c0-412f-9568-af0a96c87145\") " Mar 18 20:46:59 crc kubenswrapper[4950]: I0318 20:46:59.866494 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4hl4\" (UniqueName: \"kubernetes.io/projected/e53816d0-c4c0-412f-9568-af0a96c87145-kube-api-access-l4hl4\") pod \"e53816d0-c4c0-412f-9568-af0a96c87145\" (UID: \"e53816d0-c4c0-412f-9568-af0a96c87145\") " Mar 18 20:46:59 crc kubenswrapper[4950]: I0318 20:46:59.872494 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e53816d0-c4c0-412f-9568-af0a96c87145-ceph" (OuterVolumeSpecName: "ceph") pod "e53816d0-c4c0-412f-9568-af0a96c87145" (UID: "e53816d0-c4c0-412f-9568-af0a96c87145"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:46:59 crc kubenswrapper[4950]: I0318 20:46:59.873513 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e53816d0-c4c0-412f-9568-af0a96c87145-kube-api-access-l4hl4" (OuterVolumeSpecName: "kube-api-access-l4hl4") pod "e53816d0-c4c0-412f-9568-af0a96c87145" (UID: "e53816d0-c4c0-412f-9568-af0a96c87145"). InnerVolumeSpecName "kube-api-access-l4hl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:46:59 crc kubenswrapper[4950]: I0318 20:46:59.897082 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e53816d0-c4c0-412f-9568-af0a96c87145-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e53816d0-c4c0-412f-9568-af0a96c87145" (UID: "e53816d0-c4c0-412f-9568-af0a96c87145"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:46:59 crc kubenswrapper[4950]: I0318 20:46:59.897105 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e53816d0-c4c0-412f-9568-af0a96c87145-inventory" (OuterVolumeSpecName: "inventory") pod "e53816d0-c4c0-412f-9568-af0a96c87145" (UID: "e53816d0-c4c0-412f-9568-af0a96c87145"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:46:59 crc kubenswrapper[4950]: I0318 20:46:59.968253 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4hl4\" (UniqueName: \"kubernetes.io/projected/e53816d0-c4c0-412f-9568-af0a96c87145-kube-api-access-l4hl4\") on node \"crc\" DevicePath \"\"" Mar 18 20:46:59 crc kubenswrapper[4950]: I0318 20:46:59.968298 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e53816d0-c4c0-412f-9568-af0a96c87145-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:46:59 crc kubenswrapper[4950]: I0318 20:46:59.968314 4950 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e53816d0-c4c0-412f-9568-af0a96c87145-ceph\") on node \"crc\" DevicePath \"\"" Mar 18 20:46:59 crc kubenswrapper[4950]: I0318 20:46:59.968328 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e53816d0-c4c0-412f-9568-af0a96c87145-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.240204 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f" event={"ID":"e53816d0-c4c0-412f-9568-af0a96c87145","Type":"ContainerDied","Data":"1627851d03ccc8912993e3863de7ada91baee6f44f5ad391cf744b946ab5f849"} Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.240547 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1627851d03ccc8912993e3863de7ada91baee6f44f5ad391cf744b946ab5f849" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.240609 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.350435 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb"] Mar 18 20:47:00 crc kubenswrapper[4950]: E0318 20:47:00.350782 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e53816d0-c4c0-412f-9568-af0a96c87145" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.350797 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="e53816d0-c4c0-412f-9568-af0a96c87145" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.351000 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="e53816d0-c4c0-412f-9568-af0a96c87145" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.351565 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.357888 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.358024 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.358058 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.358026 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.358476 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.371255 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb"] Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.480787 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb\" (UID: \"3f18ebe5-e07d-40a7-911e-8d3ad683c78c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.481042 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsf6k\" (UniqueName: \"kubernetes.io/projected/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-kube-api-access-zsf6k\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb\" (UID: \"3f18ebe5-e07d-40a7-911e-8d3ad683c78c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.481091 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb\" (UID: \"3f18ebe5-e07d-40a7-911e-8d3ad683c78c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.481340 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb\" (UID: \"3f18ebe5-e07d-40a7-911e-8d3ad683c78c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.583688 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsf6k\" (UniqueName: \"kubernetes.io/projected/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-kube-api-access-zsf6k\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb\" (UID: \"3f18ebe5-e07d-40a7-911e-8d3ad683c78c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.584117 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb\" (UID: \"3f18ebe5-e07d-40a7-911e-8d3ad683c78c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.584300 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb\" (UID: \"3f18ebe5-e07d-40a7-911e-8d3ad683c78c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.584352 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb\" (UID: \"3f18ebe5-e07d-40a7-911e-8d3ad683c78c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.588966 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb\" (UID: \"3f18ebe5-e07d-40a7-911e-8d3ad683c78c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.589869 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb\" (UID: \"3f18ebe5-e07d-40a7-911e-8d3ad683c78c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.598377 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb\" (UID: \"3f18ebe5-e07d-40a7-911e-8d3ad683c78c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.623335 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsf6k\" (UniqueName: \"kubernetes.io/projected/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-kube-api-access-zsf6k\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb\" (UID: \"3f18ebe5-e07d-40a7-911e-8d3ad683c78c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb" Mar 18 20:47:00 crc kubenswrapper[4950]: I0318 20:47:00.691429 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb" Mar 18 20:47:01 crc kubenswrapper[4950]: I0318 20:47:01.147765 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb"] Mar 18 20:47:01 crc kubenswrapper[4950]: I0318 20:47:01.249023 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb" event={"ID":"3f18ebe5-e07d-40a7-911e-8d3ad683c78c","Type":"ContainerStarted","Data":"49c79891a82b9876674b5aed4ec78d0e4bd04c9bb69aabe0338e8885802eb20c"} Mar 18 20:47:02 crc kubenswrapper[4950]: I0318 20:47:02.262128 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb" event={"ID":"3f18ebe5-e07d-40a7-911e-8d3ad683c78c","Type":"ContainerStarted","Data":"7649ff8cd2c5e3e62f6bfd701c251142bc8d2084cd41f6e168bd60d9fcb23100"} Mar 18 20:47:02 crc kubenswrapper[4950]: I0318 20:47:02.293857 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb" podStartSLOduration=1.867473998 podStartE2EDuration="2.293838218s" podCreationTimestamp="2026-03-18 20:47:00 +0000 UTC" firstStartedPulling="2026-03-18 20:47:01.151596624 +0000 UTC m=+2434.392438492" lastFinishedPulling="2026-03-18 20:47:01.577960834 +0000 UTC m=+2434.818802712" observedRunningTime="2026-03-18 20:47:02.288042725 +0000 UTC m=+2435.528884603" watchObservedRunningTime="2026-03-18 20:47:02.293838218 +0000 UTC m=+2435.534680086" Mar 18 20:47:03 crc kubenswrapper[4950]: I0318 20:47:03.836025 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:47:03 crc kubenswrapper[4950]: I0318 20:47:03.836700 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:47:33 crc kubenswrapper[4950]: I0318 20:47:33.835754 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:47:33 crc kubenswrapper[4950]: I0318 20:47:33.836449 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:47:54 crc kubenswrapper[4950]: I0318 20:47:54.706096 4950 generic.go:334] "Generic (PLEG): container finished" podID="3f18ebe5-e07d-40a7-911e-8d3ad683c78c" containerID="7649ff8cd2c5e3e62f6bfd701c251142bc8d2084cd41f6e168bd60d9fcb23100" exitCode=0 Mar 18 20:47:54 crc kubenswrapper[4950]: I0318 20:47:54.706190 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb" event={"ID":"3f18ebe5-e07d-40a7-911e-8d3ad683c78c","Type":"ContainerDied","Data":"7649ff8cd2c5e3e62f6bfd701c251142bc8d2084cd41f6e168bd60d9fcb23100"} Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.136206 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb" Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.265325 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-ceph\") pod \"3f18ebe5-e07d-40a7-911e-8d3ad683c78c\" (UID: \"3f18ebe5-e07d-40a7-911e-8d3ad683c78c\") " Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.266405 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-inventory\") pod \"3f18ebe5-e07d-40a7-911e-8d3ad683c78c\" (UID: \"3f18ebe5-e07d-40a7-911e-8d3ad683c78c\") " Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.266456 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-ssh-key-openstack-edpm-ipam\") pod \"3f18ebe5-e07d-40a7-911e-8d3ad683c78c\" (UID: \"3f18ebe5-e07d-40a7-911e-8d3ad683c78c\") " Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.266508 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsf6k\" (UniqueName: \"kubernetes.io/projected/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-kube-api-access-zsf6k\") pod \"3f18ebe5-e07d-40a7-911e-8d3ad683c78c\" (UID: \"3f18ebe5-e07d-40a7-911e-8d3ad683c78c\") " Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.271761 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-kube-api-access-zsf6k" (OuterVolumeSpecName: "kube-api-access-zsf6k") pod "3f18ebe5-e07d-40a7-911e-8d3ad683c78c" (UID: "3f18ebe5-e07d-40a7-911e-8d3ad683c78c"). InnerVolumeSpecName "kube-api-access-zsf6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.271939 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-ceph" (OuterVolumeSpecName: "ceph") pod "3f18ebe5-e07d-40a7-911e-8d3ad683c78c" (UID: "3f18ebe5-e07d-40a7-911e-8d3ad683c78c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.300855 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-inventory" (OuterVolumeSpecName: "inventory") pod "3f18ebe5-e07d-40a7-911e-8d3ad683c78c" (UID: "3f18ebe5-e07d-40a7-911e-8d3ad683c78c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.301389 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3f18ebe5-e07d-40a7-911e-8d3ad683c78c" (UID: "3f18ebe5-e07d-40a7-911e-8d3ad683c78c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.369049 4950 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-ceph\") on node \"crc\" DevicePath \"\"" Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.369096 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.369116 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.369134 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsf6k\" (UniqueName: \"kubernetes.io/projected/3f18ebe5-e07d-40a7-911e-8d3ad683c78c-kube-api-access-zsf6k\") on node \"crc\" DevicePath \"\"" Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.724818 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb" event={"ID":"3f18ebe5-e07d-40a7-911e-8d3ad683c78c","Type":"ContainerDied","Data":"49c79891a82b9876674b5aed4ec78d0e4bd04c9bb69aabe0338e8885802eb20c"} Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.725088 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49c79891a82b9876674b5aed4ec78d0e4bd04c9bb69aabe0338e8885802eb20c" Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.724873 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb" Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.882865 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-zt89n"] Mar 18 20:47:56 crc kubenswrapper[4950]: E0318 20:47:56.883241 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f18ebe5-e07d-40a7-911e-8d3ad683c78c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.883262 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f18ebe5-e07d-40a7-911e-8d3ad683c78c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.883466 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f18ebe5-e07d-40a7-911e-8d3ad683c78c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.884018 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-zt89n" Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.886502 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.886807 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.886953 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.887098 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.889118 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.892564 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-zt89n"] Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.985373 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d304aae2-4025-4dea-980b-2f3b8c656a8e-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-zt89n\" (UID: \"d304aae2-4025-4dea-980b-2f3b8c656a8e\") " pod="openstack/ssh-known-hosts-edpm-deployment-zt89n" Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.985506 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d304aae2-4025-4dea-980b-2f3b8c656a8e-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-zt89n\" (UID: \"d304aae2-4025-4dea-980b-2f3b8c656a8e\") " pod="openstack/ssh-known-hosts-edpm-deployment-zt89n" Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.985551 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d304aae2-4025-4dea-980b-2f3b8c656a8e-ceph\") pod \"ssh-known-hosts-edpm-deployment-zt89n\" (UID: \"d304aae2-4025-4dea-980b-2f3b8c656a8e\") " pod="openstack/ssh-known-hosts-edpm-deployment-zt89n" Mar 18 20:47:56 crc kubenswrapper[4950]: I0318 20:47:56.985603 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2jh2\" (UniqueName: \"kubernetes.io/projected/d304aae2-4025-4dea-980b-2f3b8c656a8e-kube-api-access-f2jh2\") pod \"ssh-known-hosts-edpm-deployment-zt89n\" (UID: \"d304aae2-4025-4dea-980b-2f3b8c656a8e\") " pod="openstack/ssh-known-hosts-edpm-deployment-zt89n" Mar 18 20:47:57 crc kubenswrapper[4950]: I0318 20:47:57.087327 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d304aae2-4025-4dea-980b-2f3b8c656a8e-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-zt89n\" (UID: \"d304aae2-4025-4dea-980b-2f3b8c656a8e\") " pod="openstack/ssh-known-hosts-edpm-deployment-zt89n" Mar 18 20:47:57 crc kubenswrapper[4950]: I0318 20:47:57.087377 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d304aae2-4025-4dea-980b-2f3b8c656a8e-ceph\") pod \"ssh-known-hosts-edpm-deployment-zt89n\" (UID: \"d304aae2-4025-4dea-980b-2f3b8c656a8e\") " pod="openstack/ssh-known-hosts-edpm-deployment-zt89n" Mar 18 20:47:57 crc kubenswrapper[4950]: I0318 20:47:57.087428 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2jh2\" (UniqueName: \"kubernetes.io/projected/d304aae2-4025-4dea-980b-2f3b8c656a8e-kube-api-access-f2jh2\") pod \"ssh-known-hosts-edpm-deployment-zt89n\" (UID: \"d304aae2-4025-4dea-980b-2f3b8c656a8e\") " pod="openstack/ssh-known-hosts-edpm-deployment-zt89n" Mar 18 20:47:57 crc kubenswrapper[4950]: I0318 20:47:57.087492 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d304aae2-4025-4dea-980b-2f3b8c656a8e-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-zt89n\" (UID: \"d304aae2-4025-4dea-980b-2f3b8c656a8e\") " pod="openstack/ssh-known-hosts-edpm-deployment-zt89n" Mar 18 20:47:57 crc kubenswrapper[4950]: I0318 20:47:57.099604 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d304aae2-4025-4dea-980b-2f3b8c656a8e-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-zt89n\" (UID: \"d304aae2-4025-4dea-980b-2f3b8c656a8e\") " pod="openstack/ssh-known-hosts-edpm-deployment-zt89n" Mar 18 20:47:57 crc kubenswrapper[4950]: I0318 20:47:57.104281 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d304aae2-4025-4dea-980b-2f3b8c656a8e-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-zt89n\" (UID: \"d304aae2-4025-4dea-980b-2f3b8c656a8e\") " pod="openstack/ssh-known-hosts-edpm-deployment-zt89n" Mar 18 20:47:57 crc kubenswrapper[4950]: I0318 20:47:57.104595 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d304aae2-4025-4dea-980b-2f3b8c656a8e-ceph\") pod \"ssh-known-hosts-edpm-deployment-zt89n\" (UID: \"d304aae2-4025-4dea-980b-2f3b8c656a8e\") " pod="openstack/ssh-known-hosts-edpm-deployment-zt89n" Mar 18 20:47:57 crc kubenswrapper[4950]: I0318 20:47:57.109887 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2jh2\" (UniqueName: \"kubernetes.io/projected/d304aae2-4025-4dea-980b-2f3b8c656a8e-kube-api-access-f2jh2\") pod \"ssh-known-hosts-edpm-deployment-zt89n\" (UID: \"d304aae2-4025-4dea-980b-2f3b8c656a8e\") " pod="openstack/ssh-known-hosts-edpm-deployment-zt89n" Mar 18 20:47:57 crc kubenswrapper[4950]: I0318 20:47:57.277781 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-zt89n" Mar 18 20:47:57 crc kubenswrapper[4950]: I0318 20:47:57.923124 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-zt89n"] Mar 18 20:47:58 crc kubenswrapper[4950]: I0318 20:47:58.743400 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-zt89n" event={"ID":"d304aae2-4025-4dea-980b-2f3b8c656a8e","Type":"ContainerStarted","Data":"da3ad0377c1728eba792d2d4fb07ed2414343bafd1f32b9a966c5d38300a3cfc"} Mar 18 20:47:59 crc kubenswrapper[4950]: I0318 20:47:59.754522 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-zt89n" event={"ID":"d304aae2-4025-4dea-980b-2f3b8c656a8e","Type":"ContainerStarted","Data":"1c02b285a9ba3195b6f4f093be9e511a7ed8588e26558c4ed63d894f0d911e88"} Mar 18 20:47:59 crc kubenswrapper[4950]: I0318 20:47:59.779016 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-zt89n" podStartSLOduration=3.131729695 podStartE2EDuration="3.778996408s" podCreationTimestamp="2026-03-18 20:47:56 +0000 UTC" firstStartedPulling="2026-03-18 20:47:57.930903322 +0000 UTC m=+2491.171745190" lastFinishedPulling="2026-03-18 20:47:58.578170025 +0000 UTC m=+2491.819011903" observedRunningTime="2026-03-18 20:47:59.769682143 +0000 UTC m=+2493.010524021" watchObservedRunningTime="2026-03-18 20:47:59.778996408 +0000 UTC m=+2493.019838276" Mar 18 20:48:00 crc kubenswrapper[4950]: I0318 20:48:00.136156 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564448-cbwbn"] Mar 18 20:48:00 crc kubenswrapper[4950]: I0318 20:48:00.138127 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564448-cbwbn" Mar 18 20:48:00 crc kubenswrapper[4950]: I0318 20:48:00.141028 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:48:00 crc kubenswrapper[4950]: I0318 20:48:00.141080 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:48:00 crc kubenswrapper[4950]: I0318 20:48:00.143373 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:48:00 crc kubenswrapper[4950]: I0318 20:48:00.151156 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564448-cbwbn"] Mar 18 20:48:00 crc kubenswrapper[4950]: I0318 20:48:00.246759 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qjjt\" (UniqueName: \"kubernetes.io/projected/26ce5d7b-072e-407e-b8c5-4b5610711ee7-kube-api-access-9qjjt\") pod \"auto-csr-approver-29564448-cbwbn\" (UID: \"26ce5d7b-072e-407e-b8c5-4b5610711ee7\") " pod="openshift-infra/auto-csr-approver-29564448-cbwbn" Mar 18 20:48:00 crc kubenswrapper[4950]: I0318 20:48:00.348703 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qjjt\" (UniqueName: \"kubernetes.io/projected/26ce5d7b-072e-407e-b8c5-4b5610711ee7-kube-api-access-9qjjt\") pod \"auto-csr-approver-29564448-cbwbn\" (UID: \"26ce5d7b-072e-407e-b8c5-4b5610711ee7\") " pod="openshift-infra/auto-csr-approver-29564448-cbwbn" Mar 18 20:48:00 crc kubenswrapper[4950]: I0318 20:48:00.371312 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qjjt\" (UniqueName: \"kubernetes.io/projected/26ce5d7b-072e-407e-b8c5-4b5610711ee7-kube-api-access-9qjjt\") pod \"auto-csr-approver-29564448-cbwbn\" (UID: \"26ce5d7b-072e-407e-b8c5-4b5610711ee7\") " pod="openshift-infra/auto-csr-approver-29564448-cbwbn" Mar 18 20:48:00 crc kubenswrapper[4950]: I0318 20:48:00.465765 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564448-cbwbn" Mar 18 20:48:00 crc kubenswrapper[4950]: W0318 20:48:00.951283 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26ce5d7b_072e_407e_b8c5_4b5610711ee7.slice/crio-baee6741643f41b1806b66c54744ff67c7a1d7883a0ab32f270be6c82d05e8df WatchSource:0}: Error finding container baee6741643f41b1806b66c54744ff67c7a1d7883a0ab32f270be6c82d05e8df: Status 404 returned error can't find the container with id baee6741643f41b1806b66c54744ff67c7a1d7883a0ab32f270be6c82d05e8df Mar 18 20:48:00 crc kubenswrapper[4950]: I0318 20:48:00.962053 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564448-cbwbn"] Mar 18 20:48:01 crc kubenswrapper[4950]: I0318 20:48:01.787515 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564448-cbwbn" event={"ID":"26ce5d7b-072e-407e-b8c5-4b5610711ee7","Type":"ContainerStarted","Data":"baee6741643f41b1806b66c54744ff67c7a1d7883a0ab32f270be6c82d05e8df"} Mar 18 20:48:02 crc kubenswrapper[4950]: I0318 20:48:02.797013 4950 generic.go:334] "Generic (PLEG): container finished" podID="26ce5d7b-072e-407e-b8c5-4b5610711ee7" containerID="2aae3f0a0a4ae8919c59beacb63d556d7f9973ea9021d03f673cc338f8e03877" exitCode=0 Mar 18 20:48:02 crc kubenswrapper[4950]: I0318 20:48:02.797150 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564448-cbwbn" event={"ID":"26ce5d7b-072e-407e-b8c5-4b5610711ee7","Type":"ContainerDied","Data":"2aae3f0a0a4ae8919c59beacb63d556d7f9973ea9021d03f673cc338f8e03877"} Mar 18 20:48:03 crc kubenswrapper[4950]: I0318 20:48:03.837810 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:48:03 crc kubenswrapper[4950]: I0318 20:48:03.837853 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:48:03 crc kubenswrapper[4950]: I0318 20:48:03.837891 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:48:03 crc kubenswrapper[4950]: I0318 20:48:03.838587 4950 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18"} pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 20:48:03 crc kubenswrapper[4950]: I0318 20:48:03.838635 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" containerID="cri-o://75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" gracePeriod=600 Mar 18 20:48:03 crc kubenswrapper[4950]: E0318 20:48:03.999377 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:48:04 crc kubenswrapper[4950]: I0318 20:48:04.114156 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564448-cbwbn" Mar 18 20:48:04 crc kubenswrapper[4950]: I0318 20:48:04.155212 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qjjt\" (UniqueName: \"kubernetes.io/projected/26ce5d7b-072e-407e-b8c5-4b5610711ee7-kube-api-access-9qjjt\") pod \"26ce5d7b-072e-407e-b8c5-4b5610711ee7\" (UID: \"26ce5d7b-072e-407e-b8c5-4b5610711ee7\") " Mar 18 20:48:04 crc kubenswrapper[4950]: I0318 20:48:04.161829 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26ce5d7b-072e-407e-b8c5-4b5610711ee7-kube-api-access-9qjjt" (OuterVolumeSpecName: "kube-api-access-9qjjt") pod "26ce5d7b-072e-407e-b8c5-4b5610711ee7" (UID: "26ce5d7b-072e-407e-b8c5-4b5610711ee7"). InnerVolumeSpecName "kube-api-access-9qjjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:48:04 crc kubenswrapper[4950]: I0318 20:48:04.257539 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qjjt\" (UniqueName: \"kubernetes.io/projected/26ce5d7b-072e-407e-b8c5-4b5610711ee7-kube-api-access-9qjjt\") on node \"crc\" DevicePath \"\"" Mar 18 20:48:04 crc kubenswrapper[4950]: I0318 20:48:04.818214 4950 generic.go:334] "Generic (PLEG): container finished" podID="4048b439-3266-46e7-9de0-22377efacc46" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" exitCode=0 Mar 18 20:48:04 crc kubenswrapper[4950]: I0318 20:48:04.818305 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerDied","Data":"75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18"} Mar 18 20:48:04 crc kubenswrapper[4950]: I0318 20:48:04.818694 4950 scope.go:117] "RemoveContainer" containerID="0eb938fd8ae6f1d971cd0e6b1a12f847fe46b0c1dd5c1077e4c34d45028879ca" Mar 18 20:48:04 crc kubenswrapper[4950]: I0318 20:48:04.819857 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:48:04 crc kubenswrapper[4950]: E0318 20:48:04.820287 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:48:04 crc kubenswrapper[4950]: I0318 20:48:04.822052 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564448-cbwbn" event={"ID":"26ce5d7b-072e-407e-b8c5-4b5610711ee7","Type":"ContainerDied","Data":"baee6741643f41b1806b66c54744ff67c7a1d7883a0ab32f270be6c82d05e8df"} Mar 18 20:48:04 crc kubenswrapper[4950]: I0318 20:48:04.822091 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="baee6741643f41b1806b66c54744ff67c7a1d7883a0ab32f270be6c82d05e8df" Mar 18 20:48:04 crc kubenswrapper[4950]: I0318 20:48:04.822169 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564448-cbwbn" Mar 18 20:48:05 crc kubenswrapper[4950]: I0318 20:48:05.184397 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564442-k9z45"] Mar 18 20:48:05 crc kubenswrapper[4950]: I0318 20:48:05.190688 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564442-k9z45"] Mar 18 20:48:05 crc kubenswrapper[4950]: I0318 20:48:05.489118 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe68a5f4-7e8b-42a7-bcab-fbd81aead72e" path="/var/lib/kubelet/pods/fe68a5f4-7e8b-42a7-bcab-fbd81aead72e/volumes" Mar 18 20:48:08 crc kubenswrapper[4950]: I0318 20:48:08.857931 4950 generic.go:334] "Generic (PLEG): container finished" podID="d304aae2-4025-4dea-980b-2f3b8c656a8e" containerID="1c02b285a9ba3195b6f4f093be9e511a7ed8588e26558c4ed63d894f0d911e88" exitCode=0 Mar 18 20:48:08 crc kubenswrapper[4950]: I0318 20:48:08.858036 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-zt89n" event={"ID":"d304aae2-4025-4dea-980b-2f3b8c656a8e","Type":"ContainerDied","Data":"1c02b285a9ba3195b6f4f093be9e511a7ed8588e26558c4ed63d894f0d911e88"} Mar 18 20:48:10 crc kubenswrapper[4950]: I0318 20:48:10.275034 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-zt89n" Mar 18 20:48:10 crc kubenswrapper[4950]: I0318 20:48:10.385165 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2jh2\" (UniqueName: \"kubernetes.io/projected/d304aae2-4025-4dea-980b-2f3b8c656a8e-kube-api-access-f2jh2\") pod \"d304aae2-4025-4dea-980b-2f3b8c656a8e\" (UID: \"d304aae2-4025-4dea-980b-2f3b8c656a8e\") " Mar 18 20:48:10 crc kubenswrapper[4950]: I0318 20:48:10.385247 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d304aae2-4025-4dea-980b-2f3b8c656a8e-ssh-key-openstack-edpm-ipam\") pod \"d304aae2-4025-4dea-980b-2f3b8c656a8e\" (UID: \"d304aae2-4025-4dea-980b-2f3b8c656a8e\") " Mar 18 20:48:10 crc kubenswrapper[4950]: I0318 20:48:10.385267 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d304aae2-4025-4dea-980b-2f3b8c656a8e-ceph\") pod \"d304aae2-4025-4dea-980b-2f3b8c656a8e\" (UID: \"d304aae2-4025-4dea-980b-2f3b8c656a8e\") " Mar 18 20:48:10 crc kubenswrapper[4950]: I0318 20:48:10.385447 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d304aae2-4025-4dea-980b-2f3b8c656a8e-inventory-0\") pod \"d304aae2-4025-4dea-980b-2f3b8c656a8e\" (UID: \"d304aae2-4025-4dea-980b-2f3b8c656a8e\") " Mar 18 20:48:10 crc kubenswrapper[4950]: I0318 20:48:10.395690 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d304aae2-4025-4dea-980b-2f3b8c656a8e-ceph" (OuterVolumeSpecName: "ceph") pod "d304aae2-4025-4dea-980b-2f3b8c656a8e" (UID: "d304aae2-4025-4dea-980b-2f3b8c656a8e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:48:10 crc kubenswrapper[4950]: I0318 20:48:10.398342 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d304aae2-4025-4dea-980b-2f3b8c656a8e-kube-api-access-f2jh2" (OuterVolumeSpecName: "kube-api-access-f2jh2") pod "d304aae2-4025-4dea-980b-2f3b8c656a8e" (UID: "d304aae2-4025-4dea-980b-2f3b8c656a8e"). InnerVolumeSpecName "kube-api-access-f2jh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:48:10 crc kubenswrapper[4950]: I0318 20:48:10.409716 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d304aae2-4025-4dea-980b-2f3b8c656a8e-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "d304aae2-4025-4dea-980b-2f3b8c656a8e" (UID: "d304aae2-4025-4dea-980b-2f3b8c656a8e"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:48:10 crc kubenswrapper[4950]: I0318 20:48:10.410027 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d304aae2-4025-4dea-980b-2f3b8c656a8e-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d304aae2-4025-4dea-980b-2f3b8c656a8e" (UID: "d304aae2-4025-4dea-980b-2f3b8c656a8e"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:48:10 crc kubenswrapper[4950]: I0318 20:48:10.489177 4950 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d304aae2-4025-4dea-980b-2f3b8c656a8e-inventory-0\") on node \"crc\" DevicePath \"\"" Mar 18 20:48:10 crc kubenswrapper[4950]: I0318 20:48:10.489215 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2jh2\" (UniqueName: \"kubernetes.io/projected/d304aae2-4025-4dea-980b-2f3b8c656a8e-kube-api-access-f2jh2\") on node \"crc\" DevicePath \"\"" Mar 18 20:48:10 crc kubenswrapper[4950]: I0318 20:48:10.489229 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d304aae2-4025-4dea-980b-2f3b8c656a8e-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:48:10 crc kubenswrapper[4950]: I0318 20:48:10.489243 4950 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d304aae2-4025-4dea-980b-2f3b8c656a8e-ceph\") on node \"crc\" DevicePath \"\"" Mar 18 20:48:10 crc kubenswrapper[4950]: I0318 20:48:10.883472 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-zt89n" event={"ID":"d304aae2-4025-4dea-980b-2f3b8c656a8e","Type":"ContainerDied","Data":"da3ad0377c1728eba792d2d4fb07ed2414343bafd1f32b9a966c5d38300a3cfc"} Mar 18 20:48:10 crc kubenswrapper[4950]: I0318 20:48:10.883561 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-zt89n" Mar 18 20:48:10 crc kubenswrapper[4950]: I0318 20:48:10.883589 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da3ad0377c1728eba792d2d4fb07ed2414343bafd1f32b9a966c5d38300a3cfc" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.003915 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb"] Mar 18 20:48:11 crc kubenswrapper[4950]: E0318 20:48:11.004333 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26ce5d7b-072e-407e-b8c5-4b5610711ee7" containerName="oc" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.004355 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="26ce5d7b-072e-407e-b8c5-4b5610711ee7" containerName="oc" Mar 18 20:48:11 crc kubenswrapper[4950]: E0318 20:48:11.004375 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d304aae2-4025-4dea-980b-2f3b8c656a8e" containerName="ssh-known-hosts-edpm-deployment" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.004386 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="d304aae2-4025-4dea-980b-2f3b8c656a8e" containerName="ssh-known-hosts-edpm-deployment" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.004619 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="d304aae2-4025-4dea-980b-2f3b8c656a8e" containerName="ssh-known-hosts-edpm-deployment" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.005021 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="26ce5d7b-072e-407e-b8c5-4b5610711ee7" containerName="oc" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.005765 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.011052 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.012278 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.012323 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.012546 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.015369 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.019785 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb"] Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.102980 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5e58539e-bccd-459c-b1f7-2696ffe9a402-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gsrdb\" (UID: \"5e58539e-bccd-459c-b1f7-2696ffe9a402\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.103312 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e58539e-bccd-459c-b1f7-2696ffe9a402-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gsrdb\" (UID: \"5e58539e-bccd-459c-b1f7-2696ffe9a402\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.103356 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsb5b\" (UniqueName: \"kubernetes.io/projected/5e58539e-bccd-459c-b1f7-2696ffe9a402-kube-api-access-bsb5b\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gsrdb\" (UID: \"5e58539e-bccd-459c-b1f7-2696ffe9a402\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.103379 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e58539e-bccd-459c-b1f7-2696ffe9a402-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gsrdb\" (UID: \"5e58539e-bccd-459c-b1f7-2696ffe9a402\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.204936 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5e58539e-bccd-459c-b1f7-2696ffe9a402-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gsrdb\" (UID: \"5e58539e-bccd-459c-b1f7-2696ffe9a402\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.205064 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e58539e-bccd-459c-b1f7-2696ffe9a402-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gsrdb\" (UID: \"5e58539e-bccd-459c-b1f7-2696ffe9a402\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.205121 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsb5b\" (UniqueName: \"kubernetes.io/projected/5e58539e-bccd-459c-b1f7-2696ffe9a402-kube-api-access-bsb5b\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gsrdb\" (UID: \"5e58539e-bccd-459c-b1f7-2696ffe9a402\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.205156 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e58539e-bccd-459c-b1f7-2696ffe9a402-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gsrdb\" (UID: \"5e58539e-bccd-459c-b1f7-2696ffe9a402\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.209885 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e58539e-bccd-459c-b1f7-2696ffe9a402-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gsrdb\" (UID: \"5e58539e-bccd-459c-b1f7-2696ffe9a402\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.209926 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5e58539e-bccd-459c-b1f7-2696ffe9a402-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gsrdb\" (UID: \"5e58539e-bccd-459c-b1f7-2696ffe9a402\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.210075 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e58539e-bccd-459c-b1f7-2696ffe9a402-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gsrdb\" (UID: \"5e58539e-bccd-459c-b1f7-2696ffe9a402\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.222191 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsb5b\" (UniqueName: \"kubernetes.io/projected/5e58539e-bccd-459c-b1f7-2696ffe9a402-kube-api-access-bsb5b\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-gsrdb\" (UID: \"5e58539e-bccd-459c-b1f7-2696ffe9a402\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.323279 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb" Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.826104 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb"] Mar 18 20:48:11 crc kubenswrapper[4950]: W0318 20:48:11.833216 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e58539e_bccd_459c_b1f7_2696ffe9a402.slice/crio-e581e2f9f776387bc490f4e4112001c7f9e546d8a38f748255cf213d3746d894 WatchSource:0}: Error finding container e581e2f9f776387bc490f4e4112001c7f9e546d8a38f748255cf213d3746d894: Status 404 returned error can't find the container with id e581e2f9f776387bc490f4e4112001c7f9e546d8a38f748255cf213d3746d894 Mar 18 20:48:11 crc kubenswrapper[4950]: I0318 20:48:11.891483 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb" event={"ID":"5e58539e-bccd-459c-b1f7-2696ffe9a402","Type":"ContainerStarted","Data":"e581e2f9f776387bc490f4e4112001c7f9e546d8a38f748255cf213d3746d894"} Mar 18 20:48:12 crc kubenswrapper[4950]: I0318 20:48:12.261143 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6ct7j"] Mar 18 20:48:12 crc kubenswrapper[4950]: I0318 20:48:12.266722 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6ct7j" Mar 18 20:48:12 crc kubenswrapper[4950]: I0318 20:48:12.289321 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6ct7j"] Mar 18 20:48:12 crc kubenswrapper[4950]: I0318 20:48:12.330026 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11ce1f58-227b-4e48-89fd-282411a9536d-utilities\") pod \"redhat-operators-6ct7j\" (UID: \"11ce1f58-227b-4e48-89fd-282411a9536d\") " pod="openshift-marketplace/redhat-operators-6ct7j" Mar 18 20:48:12 crc kubenswrapper[4950]: I0318 20:48:12.330080 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11ce1f58-227b-4e48-89fd-282411a9536d-catalog-content\") pod \"redhat-operators-6ct7j\" (UID: \"11ce1f58-227b-4e48-89fd-282411a9536d\") " pod="openshift-marketplace/redhat-operators-6ct7j" Mar 18 20:48:12 crc kubenswrapper[4950]: I0318 20:48:12.330133 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58c7m\" (UniqueName: \"kubernetes.io/projected/11ce1f58-227b-4e48-89fd-282411a9536d-kube-api-access-58c7m\") pod \"redhat-operators-6ct7j\" (UID: \"11ce1f58-227b-4e48-89fd-282411a9536d\") " pod="openshift-marketplace/redhat-operators-6ct7j" Mar 18 20:48:12 crc kubenswrapper[4950]: I0318 20:48:12.432277 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11ce1f58-227b-4e48-89fd-282411a9536d-utilities\") pod \"redhat-operators-6ct7j\" (UID: \"11ce1f58-227b-4e48-89fd-282411a9536d\") " pod="openshift-marketplace/redhat-operators-6ct7j" Mar 18 20:48:12 crc kubenswrapper[4950]: I0318 20:48:12.432826 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11ce1f58-227b-4e48-89fd-282411a9536d-utilities\") pod \"redhat-operators-6ct7j\" (UID: \"11ce1f58-227b-4e48-89fd-282411a9536d\") " pod="openshift-marketplace/redhat-operators-6ct7j" Mar 18 20:48:12 crc kubenswrapper[4950]: I0318 20:48:12.432917 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11ce1f58-227b-4e48-89fd-282411a9536d-catalog-content\") pod \"redhat-operators-6ct7j\" (UID: \"11ce1f58-227b-4e48-89fd-282411a9536d\") " pod="openshift-marketplace/redhat-operators-6ct7j" Mar 18 20:48:12 crc kubenswrapper[4950]: I0318 20:48:12.432401 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11ce1f58-227b-4e48-89fd-282411a9536d-catalog-content\") pod \"redhat-operators-6ct7j\" (UID: \"11ce1f58-227b-4e48-89fd-282411a9536d\") " pod="openshift-marketplace/redhat-operators-6ct7j" Mar 18 20:48:12 crc kubenswrapper[4950]: I0318 20:48:12.433079 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58c7m\" (UniqueName: \"kubernetes.io/projected/11ce1f58-227b-4e48-89fd-282411a9536d-kube-api-access-58c7m\") pod \"redhat-operators-6ct7j\" (UID: \"11ce1f58-227b-4e48-89fd-282411a9536d\") " pod="openshift-marketplace/redhat-operators-6ct7j" Mar 18 20:48:12 crc kubenswrapper[4950]: I0318 20:48:12.454009 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58c7m\" (UniqueName: \"kubernetes.io/projected/11ce1f58-227b-4e48-89fd-282411a9536d-kube-api-access-58c7m\") pod \"redhat-operators-6ct7j\" (UID: \"11ce1f58-227b-4e48-89fd-282411a9536d\") " pod="openshift-marketplace/redhat-operators-6ct7j" Mar 18 20:48:12 crc kubenswrapper[4950]: I0318 20:48:12.516252 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6ct7j" Mar 18 20:48:12 crc kubenswrapper[4950]: I0318 20:48:12.899339 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb" event={"ID":"5e58539e-bccd-459c-b1f7-2696ffe9a402","Type":"ContainerStarted","Data":"e26d175db399a40a4896c760734965c2d2883f140f2fe3430c1b9cab4a94a322"} Mar 18 20:48:12 crc kubenswrapper[4950]: I0318 20:48:12.916147 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb" podStartSLOduration=2.495096119 podStartE2EDuration="2.916129359s" podCreationTimestamp="2026-03-18 20:48:10 +0000 UTC" firstStartedPulling="2026-03-18 20:48:11.83580208 +0000 UTC m=+2505.076643958" lastFinishedPulling="2026-03-18 20:48:12.25683533 +0000 UTC m=+2505.497677198" observedRunningTime="2026-03-18 20:48:12.912002921 +0000 UTC m=+2506.152844789" watchObservedRunningTime="2026-03-18 20:48:12.916129359 +0000 UTC m=+2506.156971227" Mar 18 20:48:12 crc kubenswrapper[4950]: I0318 20:48:12.963605 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6ct7j"] Mar 18 20:48:12 crc kubenswrapper[4950]: W0318 20:48:12.971997 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11ce1f58_227b_4e48_89fd_282411a9536d.slice/crio-2fbe6b4fd4a0b4c26c64b9e2d3ef43b4753a93c90a7956de6db85dbc73d3892d WatchSource:0}: Error finding container 2fbe6b4fd4a0b4c26c64b9e2d3ef43b4753a93c90a7956de6db85dbc73d3892d: Status 404 returned error can't find the container with id 2fbe6b4fd4a0b4c26c64b9e2d3ef43b4753a93c90a7956de6db85dbc73d3892d Mar 18 20:48:13 crc kubenswrapper[4950]: I0318 20:48:13.907242 4950 generic.go:334] "Generic (PLEG): container finished" podID="11ce1f58-227b-4e48-89fd-282411a9536d" containerID="f665917117257d5c7f2a0e2286a2ddc6f0d6b445c2465829055038eee2d94b98" exitCode=0 Mar 18 20:48:13 crc kubenswrapper[4950]: I0318 20:48:13.907347 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6ct7j" event={"ID":"11ce1f58-227b-4e48-89fd-282411a9536d","Type":"ContainerDied","Data":"f665917117257d5c7f2a0e2286a2ddc6f0d6b445c2465829055038eee2d94b98"} Mar 18 20:48:13 crc kubenswrapper[4950]: I0318 20:48:13.907649 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6ct7j" event={"ID":"11ce1f58-227b-4e48-89fd-282411a9536d","Type":"ContainerStarted","Data":"2fbe6b4fd4a0b4c26c64b9e2d3ef43b4753a93c90a7956de6db85dbc73d3892d"} Mar 18 20:48:15 crc kubenswrapper[4950]: I0318 20:48:15.930308 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6ct7j" event={"ID":"11ce1f58-227b-4e48-89fd-282411a9536d","Type":"ContainerStarted","Data":"7b85bd1aaaa9fe52c1abd3598f323231e9fce3a87fca71010125ab5687d9c26d"} Mar 18 20:48:19 crc kubenswrapper[4950]: I0318 20:48:19.969570 4950 generic.go:334] "Generic (PLEG): container finished" podID="11ce1f58-227b-4e48-89fd-282411a9536d" containerID="7b85bd1aaaa9fe52c1abd3598f323231e9fce3a87fca71010125ab5687d9c26d" exitCode=0 Mar 18 20:48:19 crc kubenswrapper[4950]: I0318 20:48:19.969654 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6ct7j" event={"ID":"11ce1f58-227b-4e48-89fd-282411a9536d","Type":"ContainerDied","Data":"7b85bd1aaaa9fe52c1abd3598f323231e9fce3a87fca71010125ab5687d9c26d"} Mar 18 20:48:20 crc kubenswrapper[4950]: I0318 20:48:20.480435 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:48:20 crc kubenswrapper[4950]: E0318 20:48:20.481197 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:48:20 crc kubenswrapper[4950]: I0318 20:48:20.984850 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6ct7j" event={"ID":"11ce1f58-227b-4e48-89fd-282411a9536d","Type":"ContainerStarted","Data":"bb9ca240759e754dab229b5f922bf267c217979af7deaef18e1b86d2b8900b87"} Mar 18 20:48:21 crc kubenswrapper[4950]: I0318 20:48:21.019921 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6ct7j" podStartSLOduration=2.49298828 podStartE2EDuration="9.019892291s" podCreationTimestamp="2026-03-18 20:48:12 +0000 UTC" firstStartedPulling="2026-03-18 20:48:13.909080402 +0000 UTC m=+2507.149922270" lastFinishedPulling="2026-03-18 20:48:20.435984413 +0000 UTC m=+2513.676826281" observedRunningTime="2026-03-18 20:48:21.008339627 +0000 UTC m=+2514.249181505" watchObservedRunningTime="2026-03-18 20:48:21.019892291 +0000 UTC m=+2514.260734169" Mar 18 20:48:21 crc kubenswrapper[4950]: I0318 20:48:21.993315 4950 generic.go:334] "Generic (PLEG): container finished" podID="5e58539e-bccd-459c-b1f7-2696ffe9a402" containerID="e26d175db399a40a4896c760734965c2d2883f140f2fe3430c1b9cab4a94a322" exitCode=0 Mar 18 20:48:21 crc kubenswrapper[4950]: I0318 20:48:21.993375 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb" event={"ID":"5e58539e-bccd-459c-b1f7-2696ffe9a402","Type":"ContainerDied","Data":"e26d175db399a40a4896c760734965c2d2883f140f2fe3430c1b9cab4a94a322"} Mar 18 20:48:22 crc kubenswrapper[4950]: I0318 20:48:22.516343 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6ct7j" Mar 18 20:48:22 crc kubenswrapper[4950]: I0318 20:48:22.516742 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6ct7j" Mar 18 20:48:23 crc kubenswrapper[4950]: I0318 20:48:23.463977 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb" Mar 18 20:48:23 crc kubenswrapper[4950]: I0318 20:48:23.539850 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e58539e-bccd-459c-b1f7-2696ffe9a402-inventory\") pod \"5e58539e-bccd-459c-b1f7-2696ffe9a402\" (UID: \"5e58539e-bccd-459c-b1f7-2696ffe9a402\") " Mar 18 20:48:23 crc kubenswrapper[4950]: I0318 20:48:23.539939 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsb5b\" (UniqueName: \"kubernetes.io/projected/5e58539e-bccd-459c-b1f7-2696ffe9a402-kube-api-access-bsb5b\") pod \"5e58539e-bccd-459c-b1f7-2696ffe9a402\" (UID: \"5e58539e-bccd-459c-b1f7-2696ffe9a402\") " Mar 18 20:48:23 crc kubenswrapper[4950]: I0318 20:48:23.540028 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5e58539e-bccd-459c-b1f7-2696ffe9a402-ceph\") pod \"5e58539e-bccd-459c-b1f7-2696ffe9a402\" (UID: \"5e58539e-bccd-459c-b1f7-2696ffe9a402\") " Mar 18 20:48:23 crc kubenswrapper[4950]: I0318 20:48:23.540081 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e58539e-bccd-459c-b1f7-2696ffe9a402-ssh-key-openstack-edpm-ipam\") pod \"5e58539e-bccd-459c-b1f7-2696ffe9a402\" (UID: \"5e58539e-bccd-459c-b1f7-2696ffe9a402\") " Mar 18 20:48:23 crc kubenswrapper[4950]: I0318 20:48:23.546731 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e58539e-bccd-459c-b1f7-2696ffe9a402-kube-api-access-bsb5b" (OuterVolumeSpecName: "kube-api-access-bsb5b") pod "5e58539e-bccd-459c-b1f7-2696ffe9a402" (UID: "5e58539e-bccd-459c-b1f7-2696ffe9a402"). InnerVolumeSpecName "kube-api-access-bsb5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:48:23 crc kubenswrapper[4950]: I0318 20:48:23.548717 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e58539e-bccd-459c-b1f7-2696ffe9a402-ceph" (OuterVolumeSpecName: "ceph") pod "5e58539e-bccd-459c-b1f7-2696ffe9a402" (UID: "5e58539e-bccd-459c-b1f7-2696ffe9a402"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:48:23 crc kubenswrapper[4950]: I0318 20:48:23.565903 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6ct7j" podUID="11ce1f58-227b-4e48-89fd-282411a9536d" containerName="registry-server" probeResult="failure" output=< Mar 18 20:48:23 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 20:48:23 crc kubenswrapper[4950]: > Mar 18 20:48:23 crc kubenswrapper[4950]: I0318 20:48:23.574539 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e58539e-bccd-459c-b1f7-2696ffe9a402-inventory" (OuterVolumeSpecName: "inventory") pod "5e58539e-bccd-459c-b1f7-2696ffe9a402" (UID: "5e58539e-bccd-459c-b1f7-2696ffe9a402"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:48:23 crc kubenswrapper[4950]: I0318 20:48:23.599586 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e58539e-bccd-459c-b1f7-2696ffe9a402-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5e58539e-bccd-459c-b1f7-2696ffe9a402" (UID: "5e58539e-bccd-459c-b1f7-2696ffe9a402"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:48:23 crc kubenswrapper[4950]: I0318 20:48:23.642602 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e58539e-bccd-459c-b1f7-2696ffe9a402-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:48:23 crc kubenswrapper[4950]: I0318 20:48:23.642638 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsb5b\" (UniqueName: \"kubernetes.io/projected/5e58539e-bccd-459c-b1f7-2696ffe9a402-kube-api-access-bsb5b\") on node \"crc\" DevicePath \"\"" Mar 18 20:48:23 crc kubenswrapper[4950]: I0318 20:48:23.642651 4950 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5e58539e-bccd-459c-b1f7-2696ffe9a402-ceph\") on node \"crc\" DevicePath \"\"" Mar 18 20:48:23 crc kubenswrapper[4950]: I0318 20:48:23.642660 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e58539e-bccd-459c-b1f7-2696ffe9a402-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.012633 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb" event={"ID":"5e58539e-bccd-459c-b1f7-2696ffe9a402","Type":"ContainerDied","Data":"e581e2f9f776387bc490f4e4112001c7f9e546d8a38f748255cf213d3746d894"} Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.012670 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e581e2f9f776387bc490f4e4112001c7f9e546d8a38f748255cf213d3746d894" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.013010 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-gsrdb" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.091320 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4"] Mar 18 20:48:24 crc kubenswrapper[4950]: E0318 20:48:24.091711 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e58539e-bccd-459c-b1f7-2696ffe9a402" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.091731 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e58539e-bccd-459c-b1f7-2696ffe9a402" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.091925 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e58539e-bccd-459c-b1f7-2696ffe9a402" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.092498 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.097523 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.097757 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.097777 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.098750 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.100283 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4"] Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.100354 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.150672 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9z54x\" (UniqueName: \"kubernetes.io/projected/afdcd712-fe52-40ce-a610-82b7dd39a59c-kube-api-access-9z54x\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4\" (UID: \"afdcd712-fe52-40ce-a610-82b7dd39a59c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.150750 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afdcd712-fe52-40ce-a610-82b7dd39a59c-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4\" (UID: \"afdcd712-fe52-40ce-a610-82b7dd39a59c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.150808 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/afdcd712-fe52-40ce-a610-82b7dd39a59c-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4\" (UID: \"afdcd712-fe52-40ce-a610-82b7dd39a59c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.150835 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/afdcd712-fe52-40ce-a610-82b7dd39a59c-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4\" (UID: \"afdcd712-fe52-40ce-a610-82b7dd39a59c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.252673 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/afdcd712-fe52-40ce-a610-82b7dd39a59c-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4\" (UID: \"afdcd712-fe52-40ce-a610-82b7dd39a59c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.252737 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/afdcd712-fe52-40ce-a610-82b7dd39a59c-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4\" (UID: \"afdcd712-fe52-40ce-a610-82b7dd39a59c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.252810 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9z54x\" (UniqueName: \"kubernetes.io/projected/afdcd712-fe52-40ce-a610-82b7dd39a59c-kube-api-access-9z54x\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4\" (UID: \"afdcd712-fe52-40ce-a610-82b7dd39a59c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.252863 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afdcd712-fe52-40ce-a610-82b7dd39a59c-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4\" (UID: \"afdcd712-fe52-40ce-a610-82b7dd39a59c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.256286 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/afdcd712-fe52-40ce-a610-82b7dd39a59c-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4\" (UID: \"afdcd712-fe52-40ce-a610-82b7dd39a59c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.258512 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/afdcd712-fe52-40ce-a610-82b7dd39a59c-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4\" (UID: \"afdcd712-fe52-40ce-a610-82b7dd39a59c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.261255 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afdcd712-fe52-40ce-a610-82b7dd39a59c-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4\" (UID: \"afdcd712-fe52-40ce-a610-82b7dd39a59c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.269957 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9z54x\" (UniqueName: \"kubernetes.io/projected/afdcd712-fe52-40ce-a610-82b7dd39a59c-kube-api-access-9z54x\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4\" (UID: \"afdcd712-fe52-40ce-a610-82b7dd39a59c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.410102 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4" Mar 18 20:48:24 crc kubenswrapper[4950]: I0318 20:48:24.980205 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4"] Mar 18 20:48:25 crc kubenswrapper[4950]: I0318 20:48:25.028924 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4" event={"ID":"afdcd712-fe52-40ce-a610-82b7dd39a59c","Type":"ContainerStarted","Data":"e592b638e427bfb52272b5d71b8ce4aaca6fcbe2664dcafeb055664f0e4c04e6"} Mar 18 20:48:26 crc kubenswrapper[4950]: I0318 20:48:26.038367 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4" event={"ID":"afdcd712-fe52-40ce-a610-82b7dd39a59c","Type":"ContainerStarted","Data":"bfdceb2a83c53497f33146cbf192d3031ab35042d15e28681720a50634d4f4cd"} Mar 18 20:48:26 crc kubenswrapper[4950]: I0318 20:48:26.080132 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4" podStartSLOduration=1.66006463 podStartE2EDuration="2.080104654s" podCreationTimestamp="2026-03-18 20:48:24 +0000 UTC" firstStartedPulling="2026-03-18 20:48:24.987848772 +0000 UTC m=+2518.228690640" lastFinishedPulling="2026-03-18 20:48:25.407888796 +0000 UTC m=+2518.648730664" observedRunningTime="2026-03-18 20:48:26.072510845 +0000 UTC m=+2519.313352723" watchObservedRunningTime="2026-03-18 20:48:26.080104654 +0000 UTC m=+2519.320946532" Mar 18 20:48:33 crc kubenswrapper[4950]: I0318 20:48:33.580818 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6ct7j" podUID="11ce1f58-227b-4e48-89fd-282411a9536d" containerName="registry-server" probeResult="failure" output=< Mar 18 20:48:33 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 20:48:33 crc kubenswrapper[4950]: > Mar 18 20:48:34 crc kubenswrapper[4950]: I0318 20:48:34.480453 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:48:34 crc kubenswrapper[4950]: E0318 20:48:34.480744 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:48:38 crc kubenswrapper[4950]: I0318 20:48:38.155212 4950 generic.go:334] "Generic (PLEG): container finished" podID="afdcd712-fe52-40ce-a610-82b7dd39a59c" containerID="bfdceb2a83c53497f33146cbf192d3031ab35042d15e28681720a50634d4f4cd" exitCode=0 Mar 18 20:48:38 crc kubenswrapper[4950]: I0318 20:48:38.155274 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4" event={"ID":"afdcd712-fe52-40ce-a610-82b7dd39a59c","Type":"ContainerDied","Data":"bfdceb2a83c53497f33146cbf192d3031ab35042d15e28681720a50634d4f4cd"} Mar 18 20:48:39 crc kubenswrapper[4950]: I0318 20:48:39.616038 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4" Mar 18 20:48:39 crc kubenswrapper[4950]: I0318 20:48:39.795877 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/afdcd712-fe52-40ce-a610-82b7dd39a59c-ssh-key-openstack-edpm-ipam\") pod \"afdcd712-fe52-40ce-a610-82b7dd39a59c\" (UID: \"afdcd712-fe52-40ce-a610-82b7dd39a59c\") " Mar 18 20:48:39 crc kubenswrapper[4950]: I0318 20:48:39.796207 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9z54x\" (UniqueName: \"kubernetes.io/projected/afdcd712-fe52-40ce-a610-82b7dd39a59c-kube-api-access-9z54x\") pod \"afdcd712-fe52-40ce-a610-82b7dd39a59c\" (UID: \"afdcd712-fe52-40ce-a610-82b7dd39a59c\") " Mar 18 20:48:39 crc kubenswrapper[4950]: I0318 20:48:39.796271 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/afdcd712-fe52-40ce-a610-82b7dd39a59c-ceph\") pod \"afdcd712-fe52-40ce-a610-82b7dd39a59c\" (UID: \"afdcd712-fe52-40ce-a610-82b7dd39a59c\") " Mar 18 20:48:39 crc kubenswrapper[4950]: I0318 20:48:39.796377 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afdcd712-fe52-40ce-a610-82b7dd39a59c-inventory\") pod \"afdcd712-fe52-40ce-a610-82b7dd39a59c\" (UID: \"afdcd712-fe52-40ce-a610-82b7dd39a59c\") " Mar 18 20:48:39 crc kubenswrapper[4950]: I0318 20:48:39.820585 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afdcd712-fe52-40ce-a610-82b7dd39a59c-ceph" (OuterVolumeSpecName: "ceph") pod "afdcd712-fe52-40ce-a610-82b7dd39a59c" (UID: "afdcd712-fe52-40ce-a610-82b7dd39a59c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:48:39 crc kubenswrapper[4950]: I0318 20:48:39.821602 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afdcd712-fe52-40ce-a610-82b7dd39a59c-kube-api-access-9z54x" (OuterVolumeSpecName: "kube-api-access-9z54x") pod "afdcd712-fe52-40ce-a610-82b7dd39a59c" (UID: "afdcd712-fe52-40ce-a610-82b7dd39a59c"). InnerVolumeSpecName "kube-api-access-9z54x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:48:39 crc kubenswrapper[4950]: I0318 20:48:39.844087 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afdcd712-fe52-40ce-a610-82b7dd39a59c-inventory" (OuterVolumeSpecName: "inventory") pod "afdcd712-fe52-40ce-a610-82b7dd39a59c" (UID: "afdcd712-fe52-40ce-a610-82b7dd39a59c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:48:39 crc kubenswrapper[4950]: I0318 20:48:39.850666 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afdcd712-fe52-40ce-a610-82b7dd39a59c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "afdcd712-fe52-40ce-a610-82b7dd39a59c" (UID: "afdcd712-fe52-40ce-a610-82b7dd39a59c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:48:39 crc kubenswrapper[4950]: I0318 20:48:39.899053 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afdcd712-fe52-40ce-a610-82b7dd39a59c-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:48:39 crc kubenswrapper[4950]: I0318 20:48:39.899084 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/afdcd712-fe52-40ce-a610-82b7dd39a59c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:48:39 crc kubenswrapper[4950]: I0318 20:48:39.899096 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9z54x\" (UniqueName: \"kubernetes.io/projected/afdcd712-fe52-40ce-a610-82b7dd39a59c-kube-api-access-9z54x\") on node \"crc\" DevicePath \"\"" Mar 18 20:48:39 crc kubenswrapper[4950]: I0318 20:48:39.899107 4950 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/afdcd712-fe52-40ce-a610-82b7dd39a59c-ceph\") on node \"crc\" DevicePath \"\"" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.175860 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4" event={"ID":"afdcd712-fe52-40ce-a610-82b7dd39a59c","Type":"ContainerDied","Data":"e592b638e427bfb52272b5d71b8ce4aaca6fcbe2664dcafeb055664f0e4c04e6"} Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.175909 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e592b638e427bfb52272b5d71b8ce4aaca6fcbe2664dcafeb055664f0e4c04e6" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.175963 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.305924 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf"] Mar 18 20:48:40 crc kubenswrapper[4950]: E0318 20:48:40.306714 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afdcd712-fe52-40ce-a610-82b7dd39a59c" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.307660 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="afdcd712-fe52-40ce-a610-82b7dd39a59c" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.308239 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="afdcd712-fe52-40ce-a610-82b7dd39a59c" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.309260 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.312241 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.312668 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.314696 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.314719 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.315836 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.315821 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.316116 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.316223 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.320239 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf"] Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.407931 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.407990 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.408027 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.408108 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.408190 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.408217 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.408246 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dt8f\" (UniqueName: \"kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-kube-api-access-9dt8f\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.408282 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.408335 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.408361 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.408408 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.408827 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.408971 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.510343 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.510388 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.510424 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dt8f\" (UniqueName: \"kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-kube-api-access-9dt8f\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.510464 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.510526 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.510548 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.510584 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.510622 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.510654 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.510678 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.510694 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.510721 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.510747 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.516875 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.517745 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.518569 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.520127 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.522097 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.522484 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.522505 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.523186 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.523324 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.526309 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.527721 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.528560 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.529577 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dt8f\" (UniqueName: \"kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-kube-api-access-9dt8f\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-472gf\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:40 crc kubenswrapper[4950]: I0318 20:48:40.627782 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:48:41 crc kubenswrapper[4950]: I0318 20:48:41.132016 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf"] Mar 18 20:48:41 crc kubenswrapper[4950]: I0318 20:48:41.135647 4950 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 20:48:41 crc kubenswrapper[4950]: I0318 20:48:41.185003 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" event={"ID":"096e1bc2-b416-48d1-bc00-082fb7023df9","Type":"ContainerStarted","Data":"3c8691b9a223c3de7dc9afe6d4bbe1796e948c4cbcecebb13b75387372ece449"} Mar 18 20:48:42 crc kubenswrapper[4950]: I0318 20:48:42.197053 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" event={"ID":"096e1bc2-b416-48d1-bc00-082fb7023df9","Type":"ContainerStarted","Data":"28dc09f6ce28f9111bf2b8cd0a68a88687f9c0031a96be6cd8c96218df46d505"} Mar 18 20:48:42 crc kubenswrapper[4950]: I0318 20:48:42.226203 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" podStartSLOduration=1.748696982 podStartE2EDuration="2.226184025s" podCreationTimestamp="2026-03-18 20:48:40 +0000 UTC" firstStartedPulling="2026-03-18 20:48:41.135232737 +0000 UTC m=+2534.376074615" lastFinishedPulling="2026-03-18 20:48:41.61271978 +0000 UTC m=+2534.853561658" observedRunningTime="2026-03-18 20:48:42.223220517 +0000 UTC m=+2535.464062385" watchObservedRunningTime="2026-03-18 20:48:42.226184025 +0000 UTC m=+2535.467025903" Mar 18 20:48:43 crc kubenswrapper[4950]: I0318 20:48:43.561385 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6ct7j" podUID="11ce1f58-227b-4e48-89fd-282411a9536d" containerName="registry-server" probeResult="failure" output=< Mar 18 20:48:43 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 20:48:43 crc kubenswrapper[4950]: > Mar 18 20:48:46 crc kubenswrapper[4950]: I0318 20:48:46.479614 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:48:46 crc kubenswrapper[4950]: E0318 20:48:46.480291 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:48:50 crc kubenswrapper[4950]: I0318 20:48:50.150482 4950 scope.go:117] "RemoveContainer" containerID="4476e98b688feef0f2d052155367dfc73a22f97ece6566f6363dc47cd0ebb8ec" Mar 18 20:48:52 crc kubenswrapper[4950]: I0318 20:48:52.599220 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6ct7j" Mar 18 20:48:52 crc kubenswrapper[4950]: I0318 20:48:52.651461 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6ct7j" Mar 18 20:48:52 crc kubenswrapper[4950]: I0318 20:48:52.840575 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6ct7j"] Mar 18 20:48:54 crc kubenswrapper[4950]: I0318 20:48:54.293614 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6ct7j" podUID="11ce1f58-227b-4e48-89fd-282411a9536d" containerName="registry-server" containerID="cri-o://bb9ca240759e754dab229b5f922bf267c217979af7deaef18e1b86d2b8900b87" gracePeriod=2 Mar 18 20:48:54 crc kubenswrapper[4950]: I0318 20:48:54.737300 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6ct7j" Mar 18 20:48:54 crc kubenswrapper[4950]: I0318 20:48:54.811054 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11ce1f58-227b-4e48-89fd-282411a9536d-utilities\") pod \"11ce1f58-227b-4e48-89fd-282411a9536d\" (UID: \"11ce1f58-227b-4e48-89fd-282411a9536d\") " Mar 18 20:48:54 crc kubenswrapper[4950]: I0318 20:48:54.811092 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58c7m\" (UniqueName: \"kubernetes.io/projected/11ce1f58-227b-4e48-89fd-282411a9536d-kube-api-access-58c7m\") pod \"11ce1f58-227b-4e48-89fd-282411a9536d\" (UID: \"11ce1f58-227b-4e48-89fd-282411a9536d\") " Mar 18 20:48:54 crc kubenswrapper[4950]: I0318 20:48:54.811119 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11ce1f58-227b-4e48-89fd-282411a9536d-catalog-content\") pod \"11ce1f58-227b-4e48-89fd-282411a9536d\" (UID: \"11ce1f58-227b-4e48-89fd-282411a9536d\") " Mar 18 20:48:54 crc kubenswrapper[4950]: I0318 20:48:54.812081 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11ce1f58-227b-4e48-89fd-282411a9536d-utilities" (OuterVolumeSpecName: "utilities") pod "11ce1f58-227b-4e48-89fd-282411a9536d" (UID: "11ce1f58-227b-4e48-89fd-282411a9536d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:48:54 crc kubenswrapper[4950]: I0318 20:48:54.820030 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11ce1f58-227b-4e48-89fd-282411a9536d-kube-api-access-58c7m" (OuterVolumeSpecName: "kube-api-access-58c7m") pod "11ce1f58-227b-4e48-89fd-282411a9536d" (UID: "11ce1f58-227b-4e48-89fd-282411a9536d"). InnerVolumeSpecName "kube-api-access-58c7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:48:54 crc kubenswrapper[4950]: I0318 20:48:54.912517 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11ce1f58-227b-4e48-89fd-282411a9536d-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:48:54 crc kubenswrapper[4950]: I0318 20:48:54.912543 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58c7m\" (UniqueName: \"kubernetes.io/projected/11ce1f58-227b-4e48-89fd-282411a9536d-kube-api-access-58c7m\") on node \"crc\" DevicePath \"\"" Mar 18 20:48:54 crc kubenswrapper[4950]: I0318 20:48:54.960328 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11ce1f58-227b-4e48-89fd-282411a9536d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "11ce1f58-227b-4e48-89fd-282411a9536d" (UID: "11ce1f58-227b-4e48-89fd-282411a9536d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:48:55 crc kubenswrapper[4950]: I0318 20:48:55.014248 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11ce1f58-227b-4e48-89fd-282411a9536d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:48:55 crc kubenswrapper[4950]: I0318 20:48:55.306046 4950 generic.go:334] "Generic (PLEG): container finished" podID="11ce1f58-227b-4e48-89fd-282411a9536d" containerID="bb9ca240759e754dab229b5f922bf267c217979af7deaef18e1b86d2b8900b87" exitCode=0 Mar 18 20:48:55 crc kubenswrapper[4950]: I0318 20:48:55.306091 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6ct7j" event={"ID":"11ce1f58-227b-4e48-89fd-282411a9536d","Type":"ContainerDied","Data":"bb9ca240759e754dab229b5f922bf267c217979af7deaef18e1b86d2b8900b87"} Mar 18 20:48:55 crc kubenswrapper[4950]: I0318 20:48:55.306121 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6ct7j" event={"ID":"11ce1f58-227b-4e48-89fd-282411a9536d","Type":"ContainerDied","Data":"2fbe6b4fd4a0b4c26c64b9e2d3ef43b4753a93c90a7956de6db85dbc73d3892d"} Mar 18 20:48:55 crc kubenswrapper[4950]: I0318 20:48:55.306147 4950 scope.go:117] "RemoveContainer" containerID="bb9ca240759e754dab229b5f922bf267c217979af7deaef18e1b86d2b8900b87" Mar 18 20:48:55 crc kubenswrapper[4950]: I0318 20:48:55.306202 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6ct7j" Mar 18 20:48:55 crc kubenswrapper[4950]: I0318 20:48:55.339679 4950 scope.go:117] "RemoveContainer" containerID="7b85bd1aaaa9fe52c1abd3598f323231e9fce3a87fca71010125ab5687d9c26d" Mar 18 20:48:55 crc kubenswrapper[4950]: I0318 20:48:55.375900 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6ct7j"] Mar 18 20:48:55 crc kubenswrapper[4950]: I0318 20:48:55.389658 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6ct7j"] Mar 18 20:48:55 crc kubenswrapper[4950]: I0318 20:48:55.402848 4950 scope.go:117] "RemoveContainer" containerID="f665917117257d5c7f2a0e2286a2ddc6f0d6b445c2465829055038eee2d94b98" Mar 18 20:48:55 crc kubenswrapper[4950]: I0318 20:48:55.445750 4950 scope.go:117] "RemoveContainer" containerID="bb9ca240759e754dab229b5f922bf267c217979af7deaef18e1b86d2b8900b87" Mar 18 20:48:55 crc kubenswrapper[4950]: E0318 20:48:55.447054 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb9ca240759e754dab229b5f922bf267c217979af7deaef18e1b86d2b8900b87\": container with ID starting with bb9ca240759e754dab229b5f922bf267c217979af7deaef18e1b86d2b8900b87 not found: ID does not exist" containerID="bb9ca240759e754dab229b5f922bf267c217979af7deaef18e1b86d2b8900b87" Mar 18 20:48:55 crc kubenswrapper[4950]: I0318 20:48:55.447094 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb9ca240759e754dab229b5f922bf267c217979af7deaef18e1b86d2b8900b87"} err="failed to get container status \"bb9ca240759e754dab229b5f922bf267c217979af7deaef18e1b86d2b8900b87\": rpc error: code = NotFound desc = could not find container \"bb9ca240759e754dab229b5f922bf267c217979af7deaef18e1b86d2b8900b87\": container with ID starting with bb9ca240759e754dab229b5f922bf267c217979af7deaef18e1b86d2b8900b87 not found: ID does not exist" Mar 18 20:48:55 crc kubenswrapper[4950]: I0318 20:48:55.447121 4950 scope.go:117] "RemoveContainer" containerID="7b85bd1aaaa9fe52c1abd3598f323231e9fce3a87fca71010125ab5687d9c26d" Mar 18 20:48:55 crc kubenswrapper[4950]: E0318 20:48:55.447453 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b85bd1aaaa9fe52c1abd3598f323231e9fce3a87fca71010125ab5687d9c26d\": container with ID starting with 7b85bd1aaaa9fe52c1abd3598f323231e9fce3a87fca71010125ab5687d9c26d not found: ID does not exist" containerID="7b85bd1aaaa9fe52c1abd3598f323231e9fce3a87fca71010125ab5687d9c26d" Mar 18 20:48:55 crc kubenswrapper[4950]: I0318 20:48:55.447471 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b85bd1aaaa9fe52c1abd3598f323231e9fce3a87fca71010125ab5687d9c26d"} err="failed to get container status \"7b85bd1aaaa9fe52c1abd3598f323231e9fce3a87fca71010125ab5687d9c26d\": rpc error: code = NotFound desc = could not find container \"7b85bd1aaaa9fe52c1abd3598f323231e9fce3a87fca71010125ab5687d9c26d\": container with ID starting with 7b85bd1aaaa9fe52c1abd3598f323231e9fce3a87fca71010125ab5687d9c26d not found: ID does not exist" Mar 18 20:48:55 crc kubenswrapper[4950]: I0318 20:48:55.447483 4950 scope.go:117] "RemoveContainer" containerID="f665917117257d5c7f2a0e2286a2ddc6f0d6b445c2465829055038eee2d94b98" Mar 18 20:48:55 crc kubenswrapper[4950]: E0318 20:48:55.447695 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f665917117257d5c7f2a0e2286a2ddc6f0d6b445c2465829055038eee2d94b98\": container with ID starting with f665917117257d5c7f2a0e2286a2ddc6f0d6b445c2465829055038eee2d94b98 not found: ID does not exist" containerID="f665917117257d5c7f2a0e2286a2ddc6f0d6b445c2465829055038eee2d94b98" Mar 18 20:48:55 crc kubenswrapper[4950]: I0318 20:48:55.447717 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f665917117257d5c7f2a0e2286a2ddc6f0d6b445c2465829055038eee2d94b98"} err="failed to get container status \"f665917117257d5c7f2a0e2286a2ddc6f0d6b445c2465829055038eee2d94b98\": rpc error: code = NotFound desc = could not find container \"f665917117257d5c7f2a0e2286a2ddc6f0d6b445c2465829055038eee2d94b98\": container with ID starting with f665917117257d5c7f2a0e2286a2ddc6f0d6b445c2465829055038eee2d94b98 not found: ID does not exist" Mar 18 20:48:55 crc kubenswrapper[4950]: I0318 20:48:55.492583 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11ce1f58-227b-4e48-89fd-282411a9536d" path="/var/lib/kubelet/pods/11ce1f58-227b-4e48-89fd-282411a9536d/volumes" Mar 18 20:48:59 crc kubenswrapper[4950]: I0318 20:48:59.479458 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:48:59 crc kubenswrapper[4950]: E0318 20:48:59.481646 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:49:14 crc kubenswrapper[4950]: I0318 20:49:14.480764 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:49:14 crc kubenswrapper[4950]: E0318 20:49:14.482020 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:49:16 crc kubenswrapper[4950]: I0318 20:49:16.520298 4950 generic.go:334] "Generic (PLEG): container finished" podID="096e1bc2-b416-48d1-bc00-082fb7023df9" containerID="28dc09f6ce28f9111bf2b8cd0a68a88687f9c0031a96be6cd8c96218df46d505" exitCode=0 Mar 18 20:49:16 crc kubenswrapper[4950]: I0318 20:49:16.520380 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" event={"ID":"096e1bc2-b416-48d1-bc00-082fb7023df9","Type":"ContainerDied","Data":"28dc09f6ce28f9111bf2b8cd0a68a88687f9c0031a96be6cd8c96218df46d505"} Mar 18 20:49:17 crc kubenswrapper[4950]: I0318 20:49:17.993504 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.162386 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"096e1bc2-b416-48d1-bc00-082fb7023df9\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.162472 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-nova-combined-ca-bundle\") pod \"096e1bc2-b416-48d1-bc00-082fb7023df9\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.162505 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-libvirt-combined-ca-bundle\") pod \"096e1bc2-b416-48d1-bc00-082fb7023df9\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.162561 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"096e1bc2-b416-48d1-bc00-082fb7023df9\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.162584 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-inventory\") pod \"096e1bc2-b416-48d1-bc00-082fb7023df9\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.162621 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-ssh-key-openstack-edpm-ipam\") pod \"096e1bc2-b416-48d1-bc00-082fb7023df9\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.162684 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"096e1bc2-b416-48d1-bc00-082fb7023df9\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.162703 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-repo-setup-combined-ca-bundle\") pod \"096e1bc2-b416-48d1-bc00-082fb7023df9\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.162735 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dt8f\" (UniqueName: \"kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-kube-api-access-9dt8f\") pod \"096e1bc2-b416-48d1-bc00-082fb7023df9\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.162756 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-ovn-combined-ca-bundle\") pod \"096e1bc2-b416-48d1-bc00-082fb7023df9\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.162814 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-ceph\") pod \"096e1bc2-b416-48d1-bc00-082fb7023df9\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.162840 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-neutron-metadata-combined-ca-bundle\") pod \"096e1bc2-b416-48d1-bc00-082fb7023df9\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.162876 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-bootstrap-combined-ca-bundle\") pod \"096e1bc2-b416-48d1-bc00-082fb7023df9\" (UID: \"096e1bc2-b416-48d1-bc00-082fb7023df9\") " Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.170024 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "096e1bc2-b416-48d1-bc00-082fb7023df9" (UID: "096e1bc2-b416-48d1-bc00-082fb7023df9"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.175279 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "096e1bc2-b416-48d1-bc00-082fb7023df9" (UID: "096e1bc2-b416-48d1-bc00-082fb7023df9"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.175853 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "096e1bc2-b416-48d1-bc00-082fb7023df9" (UID: "096e1bc2-b416-48d1-bc00-082fb7023df9"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.175895 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "096e1bc2-b416-48d1-bc00-082fb7023df9" (UID: "096e1bc2-b416-48d1-bc00-082fb7023df9"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.175904 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-kube-api-access-9dt8f" (OuterVolumeSpecName: "kube-api-access-9dt8f") pod "096e1bc2-b416-48d1-bc00-082fb7023df9" (UID: "096e1bc2-b416-48d1-bc00-082fb7023df9"). InnerVolumeSpecName "kube-api-access-9dt8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.175944 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-ceph" (OuterVolumeSpecName: "ceph") pod "096e1bc2-b416-48d1-bc00-082fb7023df9" (UID: "096e1bc2-b416-48d1-bc00-082fb7023df9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.176112 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "096e1bc2-b416-48d1-bc00-082fb7023df9" (UID: "096e1bc2-b416-48d1-bc00-082fb7023df9"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.178939 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "096e1bc2-b416-48d1-bc00-082fb7023df9" (UID: "096e1bc2-b416-48d1-bc00-082fb7023df9"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.178993 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "096e1bc2-b416-48d1-bc00-082fb7023df9" (UID: "096e1bc2-b416-48d1-bc00-082fb7023df9"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.179304 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "096e1bc2-b416-48d1-bc00-082fb7023df9" (UID: "096e1bc2-b416-48d1-bc00-082fb7023df9"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.181400 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "096e1bc2-b416-48d1-bc00-082fb7023df9" (UID: "096e1bc2-b416-48d1-bc00-082fb7023df9"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.197831 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "096e1bc2-b416-48d1-bc00-082fb7023df9" (UID: "096e1bc2-b416-48d1-bc00-082fb7023df9"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.198507 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-inventory" (OuterVolumeSpecName: "inventory") pod "096e1bc2-b416-48d1-bc00-082fb7023df9" (UID: "096e1bc2-b416-48d1-bc00-082fb7023df9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.267531 4950 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.267565 4950 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.267611 4950 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.267621 4950 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.267631 4950 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.267643 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.267651 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.267717 4950 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.267732 4950 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.267777 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dt8f\" (UniqueName: \"kubernetes.io/projected/096e1bc2-b416-48d1-bc00-082fb7023df9-kube-api-access-9dt8f\") on node \"crc\" DevicePath \"\"" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.267790 4950 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.267802 4950 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-ceph\") on node \"crc\" DevicePath \"\"" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.267841 4950 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096e1bc2-b416-48d1-bc00-082fb7023df9-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.541851 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" event={"ID":"096e1bc2-b416-48d1-bc00-082fb7023df9","Type":"ContainerDied","Data":"3c8691b9a223c3de7dc9afe6d4bbe1796e948c4cbcecebb13b75387372ece449"} Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.542143 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c8691b9a223c3de7dc9afe6d4bbe1796e948c4cbcecebb13b75387372ece449" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.542204 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-472gf" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.691898 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj"] Mar 18 20:49:18 crc kubenswrapper[4950]: E0318 20:49:18.692524 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11ce1f58-227b-4e48-89fd-282411a9536d" containerName="registry-server" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.692556 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="11ce1f58-227b-4e48-89fd-282411a9536d" containerName="registry-server" Mar 18 20:49:18 crc kubenswrapper[4950]: E0318 20:49:18.692580 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11ce1f58-227b-4e48-89fd-282411a9536d" containerName="extract-content" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.692595 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="11ce1f58-227b-4e48-89fd-282411a9536d" containerName="extract-content" Mar 18 20:49:18 crc kubenswrapper[4950]: E0318 20:49:18.692622 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11ce1f58-227b-4e48-89fd-282411a9536d" containerName="extract-utilities" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.692635 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="11ce1f58-227b-4e48-89fd-282411a9536d" containerName="extract-utilities" Mar 18 20:49:18 crc kubenswrapper[4950]: E0318 20:49:18.692696 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="096e1bc2-b416-48d1-bc00-082fb7023df9" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.692710 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="096e1bc2-b416-48d1-bc00-082fb7023df9" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.693038 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="11ce1f58-227b-4e48-89fd-282411a9536d" containerName="registry-server" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.693076 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="096e1bc2-b416-48d1-bc00-082fb7023df9" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.694015 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.699493 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.699855 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.700119 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.700356 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.700595 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.705701 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj"] Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.882701 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2e43754d-5118-4f9d-afe0-c70c6d4e7589-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj\" (UID: \"2e43754d-5118-4f9d-afe0-c70c6d4e7589\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.882791 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e43754d-5118-4f9d-afe0-c70c6d4e7589-ssh-key-openstack-edpm-ipam\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj\" (UID: \"2e43754d-5118-4f9d-afe0-c70c6d4e7589\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.882844 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svbrt\" (UniqueName: \"kubernetes.io/projected/2e43754d-5118-4f9d-afe0-c70c6d4e7589-kube-api-access-svbrt\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj\" (UID: \"2e43754d-5118-4f9d-afe0-c70c6d4e7589\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.882937 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e43754d-5118-4f9d-afe0-c70c6d4e7589-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj\" (UID: \"2e43754d-5118-4f9d-afe0-c70c6d4e7589\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.984163 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2e43754d-5118-4f9d-afe0-c70c6d4e7589-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj\" (UID: \"2e43754d-5118-4f9d-afe0-c70c6d4e7589\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.984224 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e43754d-5118-4f9d-afe0-c70c6d4e7589-ssh-key-openstack-edpm-ipam\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj\" (UID: \"2e43754d-5118-4f9d-afe0-c70c6d4e7589\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.984291 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svbrt\" (UniqueName: \"kubernetes.io/projected/2e43754d-5118-4f9d-afe0-c70c6d4e7589-kube-api-access-svbrt\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj\" (UID: \"2e43754d-5118-4f9d-afe0-c70c6d4e7589\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.984353 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e43754d-5118-4f9d-afe0-c70c6d4e7589-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj\" (UID: \"2e43754d-5118-4f9d-afe0-c70c6d4e7589\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.990082 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e43754d-5118-4f9d-afe0-c70c6d4e7589-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj\" (UID: \"2e43754d-5118-4f9d-afe0-c70c6d4e7589\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.990237 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e43754d-5118-4f9d-afe0-c70c6d4e7589-ssh-key-openstack-edpm-ipam\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj\" (UID: \"2e43754d-5118-4f9d-afe0-c70c6d4e7589\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj" Mar 18 20:49:18 crc kubenswrapper[4950]: I0318 20:49:18.990561 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2e43754d-5118-4f9d-afe0-c70c6d4e7589-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj\" (UID: \"2e43754d-5118-4f9d-afe0-c70c6d4e7589\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj" Mar 18 20:49:19 crc kubenswrapper[4950]: I0318 20:49:19.015701 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svbrt\" (UniqueName: \"kubernetes.io/projected/2e43754d-5118-4f9d-afe0-c70c6d4e7589-kube-api-access-svbrt\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj\" (UID: \"2e43754d-5118-4f9d-afe0-c70c6d4e7589\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj" Mar 18 20:49:19 crc kubenswrapper[4950]: I0318 20:49:19.312852 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj" Mar 18 20:49:19 crc kubenswrapper[4950]: I0318 20:49:19.864865 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj"] Mar 18 20:49:20 crc kubenswrapper[4950]: I0318 20:49:20.559309 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj" event={"ID":"2e43754d-5118-4f9d-afe0-c70c6d4e7589","Type":"ContainerStarted","Data":"aee8d80bf10113060773d76a095ff7c4f494ad182cbb977916b5b5d1a5579643"} Mar 18 20:49:21 crc kubenswrapper[4950]: I0318 20:49:21.578187 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj" event={"ID":"2e43754d-5118-4f9d-afe0-c70c6d4e7589","Type":"ContainerStarted","Data":"4796dc74e8358fda1f5fe25280ce3598f34d5ad03b88422879fa0df9e18d09fb"} Mar 18 20:49:21 crc kubenswrapper[4950]: I0318 20:49:21.601108 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj" podStartSLOduration=2.872355614 podStartE2EDuration="3.601084126s" podCreationTimestamp="2026-03-18 20:49:18 +0000 UTC" firstStartedPulling="2026-03-18 20:49:19.875869538 +0000 UTC m=+2573.116711406" lastFinishedPulling="2026-03-18 20:49:20.60459803 +0000 UTC m=+2573.845439918" observedRunningTime="2026-03-18 20:49:21.59435982 +0000 UTC m=+2574.835201688" watchObservedRunningTime="2026-03-18 20:49:21.601084126 +0000 UTC m=+2574.841925994" Mar 18 20:49:26 crc kubenswrapper[4950]: I0318 20:49:26.616185 4950 generic.go:334] "Generic (PLEG): container finished" podID="2e43754d-5118-4f9d-afe0-c70c6d4e7589" containerID="4796dc74e8358fda1f5fe25280ce3598f34d5ad03b88422879fa0df9e18d09fb" exitCode=0 Mar 18 20:49:26 crc kubenswrapper[4950]: I0318 20:49:26.616535 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj" event={"ID":"2e43754d-5118-4f9d-afe0-c70c6d4e7589","Type":"ContainerDied","Data":"4796dc74e8358fda1f5fe25280ce3598f34d5ad03b88422879fa0df9e18d09fb"} Mar 18 20:49:27 crc kubenswrapper[4950]: I0318 20:49:27.488145 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:49:27 crc kubenswrapper[4950]: E0318 20:49:27.488861 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.094802 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.179582 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e43754d-5118-4f9d-afe0-c70c6d4e7589-inventory\") pod \"2e43754d-5118-4f9d-afe0-c70c6d4e7589\" (UID: \"2e43754d-5118-4f9d-afe0-c70c6d4e7589\") " Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.179691 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e43754d-5118-4f9d-afe0-c70c6d4e7589-ssh-key-openstack-edpm-ipam\") pod \"2e43754d-5118-4f9d-afe0-c70c6d4e7589\" (UID: \"2e43754d-5118-4f9d-afe0-c70c6d4e7589\") " Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.179792 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svbrt\" (UniqueName: \"kubernetes.io/projected/2e43754d-5118-4f9d-afe0-c70c6d4e7589-kube-api-access-svbrt\") pod \"2e43754d-5118-4f9d-afe0-c70c6d4e7589\" (UID: \"2e43754d-5118-4f9d-afe0-c70c6d4e7589\") " Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.180010 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2e43754d-5118-4f9d-afe0-c70c6d4e7589-ceph\") pod \"2e43754d-5118-4f9d-afe0-c70c6d4e7589\" (UID: \"2e43754d-5118-4f9d-afe0-c70c6d4e7589\") " Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.206631 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e43754d-5118-4f9d-afe0-c70c6d4e7589-kube-api-access-svbrt" (OuterVolumeSpecName: "kube-api-access-svbrt") pod "2e43754d-5118-4f9d-afe0-c70c6d4e7589" (UID: "2e43754d-5118-4f9d-afe0-c70c6d4e7589"). InnerVolumeSpecName "kube-api-access-svbrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.210985 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e43754d-5118-4f9d-afe0-c70c6d4e7589-ceph" (OuterVolumeSpecName: "ceph") pod "2e43754d-5118-4f9d-afe0-c70c6d4e7589" (UID: "2e43754d-5118-4f9d-afe0-c70c6d4e7589"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.219515 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e43754d-5118-4f9d-afe0-c70c6d4e7589-inventory" (OuterVolumeSpecName: "inventory") pod "2e43754d-5118-4f9d-afe0-c70c6d4e7589" (UID: "2e43754d-5118-4f9d-afe0-c70c6d4e7589"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.272616 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e43754d-5118-4f9d-afe0-c70c6d4e7589-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2e43754d-5118-4f9d-afe0-c70c6d4e7589" (UID: "2e43754d-5118-4f9d-afe0-c70c6d4e7589"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.282069 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e43754d-5118-4f9d-afe0-c70c6d4e7589-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.282341 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e43754d-5118-4f9d-afe0-c70c6d4e7589-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.282446 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svbrt\" (UniqueName: \"kubernetes.io/projected/2e43754d-5118-4f9d-afe0-c70c6d4e7589-kube-api-access-svbrt\") on node \"crc\" DevicePath \"\"" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.282519 4950 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2e43754d-5118-4f9d-afe0-c70c6d4e7589-ceph\") on node \"crc\" DevicePath \"\"" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.634135 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj" event={"ID":"2e43754d-5118-4f9d-afe0-c70c6d4e7589","Type":"ContainerDied","Data":"aee8d80bf10113060773d76a095ff7c4f494ad182cbb977916b5b5d1a5579643"} Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.634185 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aee8d80bf10113060773d76a095ff7c4f494ad182cbb977916b5b5d1a5579643" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.634244 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.728092 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9"] Mar 18 20:49:28 crc kubenswrapper[4950]: E0318 20:49:28.728682 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e43754d-5118-4f9d-afe0-c70c6d4e7589" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.728783 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e43754d-5118-4f9d-afe0-c70c6d4e7589" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.729060 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e43754d-5118-4f9d-afe0-c70c6d4e7589" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.729746 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.732618 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.732976 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.733065 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.733306 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.734181 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.742753 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.744068 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9"] Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.790879 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnlg9\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.790955 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjftp\" (UniqueName: \"kubernetes.io/projected/67bb928c-890d-480c-8ce7-ec7b299cb1ee-kube-api-access-fjftp\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnlg9\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.791028 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnlg9\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.791053 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnlg9\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.791112 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnlg9\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.791143 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnlg9\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.892815 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnlg9\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.893126 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnlg9\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.893185 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnlg9\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.893225 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjftp\" (UniqueName: \"kubernetes.io/projected/67bb928c-890d-480c-8ce7-ec7b299cb1ee-kube-api-access-fjftp\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnlg9\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.893272 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnlg9\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.893295 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnlg9\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.894630 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnlg9\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.896330 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnlg9\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.896507 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnlg9\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.896893 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnlg9\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.897858 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnlg9\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" Mar 18 20:49:28 crc kubenswrapper[4950]: I0318 20:49:28.910024 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjftp\" (UniqueName: \"kubernetes.io/projected/67bb928c-890d-480c-8ce7-ec7b299cb1ee-kube-api-access-fjftp\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-hnlg9\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" Mar 18 20:49:29 crc kubenswrapper[4950]: I0318 20:49:29.049814 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" Mar 18 20:49:29 crc kubenswrapper[4950]: I0318 20:49:29.654698 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9"] Mar 18 20:49:30 crc kubenswrapper[4950]: I0318 20:49:30.649123 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" event={"ID":"67bb928c-890d-480c-8ce7-ec7b299cb1ee","Type":"ContainerStarted","Data":"81db0ed1bf206d71c8d9ec808ea1b58e81c264fbcb673c24eca613623519082c"} Mar 18 20:49:30 crc kubenswrapper[4950]: I0318 20:49:30.649811 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" event={"ID":"67bb928c-890d-480c-8ce7-ec7b299cb1ee","Type":"ContainerStarted","Data":"68368e987009cac789876ef73b839d3f0c2d9b3e97c81ee1b06b51da40ec1289"} Mar 18 20:49:30 crc kubenswrapper[4950]: I0318 20:49:30.674687 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" podStartSLOduration=2.241715794 podStartE2EDuration="2.674666797s" podCreationTimestamp="2026-03-18 20:49:28 +0000 UTC" firstStartedPulling="2026-03-18 20:49:29.655391912 +0000 UTC m=+2582.896233790" lastFinishedPulling="2026-03-18 20:49:30.088342925 +0000 UTC m=+2583.329184793" observedRunningTime="2026-03-18 20:49:30.666961504 +0000 UTC m=+2583.907803382" watchObservedRunningTime="2026-03-18 20:49:30.674666797 +0000 UTC m=+2583.915508675" Mar 18 20:49:39 crc kubenswrapper[4950]: I0318 20:49:39.479585 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:49:39 crc kubenswrapper[4950]: E0318 20:49:39.480403 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:49:53 crc kubenswrapper[4950]: I0318 20:49:53.480028 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:49:53 crc kubenswrapper[4950]: E0318 20:49:53.481061 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:50:00 crc kubenswrapper[4950]: I0318 20:50:00.146857 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564450-xwsr8"] Mar 18 20:50:00 crc kubenswrapper[4950]: I0318 20:50:00.150195 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564450-xwsr8" Mar 18 20:50:00 crc kubenswrapper[4950]: I0318 20:50:00.153823 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:50:00 crc kubenswrapper[4950]: I0318 20:50:00.154855 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:50:00 crc kubenswrapper[4950]: I0318 20:50:00.155498 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:50:00 crc kubenswrapper[4950]: I0318 20:50:00.160722 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564450-xwsr8"] Mar 18 20:50:00 crc kubenswrapper[4950]: I0318 20:50:00.349736 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp7gn\" (UniqueName: \"kubernetes.io/projected/9d9d3129-446e-40a8-873f-f427bdd96d25-kube-api-access-tp7gn\") pod \"auto-csr-approver-29564450-xwsr8\" (UID: \"9d9d3129-446e-40a8-873f-f427bdd96d25\") " pod="openshift-infra/auto-csr-approver-29564450-xwsr8" Mar 18 20:50:00 crc kubenswrapper[4950]: I0318 20:50:00.451496 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp7gn\" (UniqueName: \"kubernetes.io/projected/9d9d3129-446e-40a8-873f-f427bdd96d25-kube-api-access-tp7gn\") pod \"auto-csr-approver-29564450-xwsr8\" (UID: \"9d9d3129-446e-40a8-873f-f427bdd96d25\") " pod="openshift-infra/auto-csr-approver-29564450-xwsr8" Mar 18 20:50:00 crc kubenswrapper[4950]: I0318 20:50:00.476719 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp7gn\" (UniqueName: \"kubernetes.io/projected/9d9d3129-446e-40a8-873f-f427bdd96d25-kube-api-access-tp7gn\") pod \"auto-csr-approver-29564450-xwsr8\" (UID: \"9d9d3129-446e-40a8-873f-f427bdd96d25\") " pod="openshift-infra/auto-csr-approver-29564450-xwsr8" Mar 18 20:50:00 crc kubenswrapper[4950]: I0318 20:50:00.772885 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564450-xwsr8" Mar 18 20:50:01 crc kubenswrapper[4950]: I0318 20:50:01.266632 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564450-xwsr8"] Mar 18 20:50:01 crc kubenswrapper[4950]: I0318 20:50:01.938314 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564450-xwsr8" event={"ID":"9d9d3129-446e-40a8-873f-f427bdd96d25","Type":"ContainerStarted","Data":"4393017ea33318dd6e5bce41335af3785fa6e84defbb47149530af0ae9714b83"} Mar 18 20:50:02 crc kubenswrapper[4950]: I0318 20:50:02.947239 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564450-xwsr8" event={"ID":"9d9d3129-446e-40a8-873f-f427bdd96d25","Type":"ContainerStarted","Data":"0aa0efd12d9ea9bbe4a547a35f9f99d1107d0539f9ef66bd867fa8cb4f44fc05"} Mar 18 20:50:02 crc kubenswrapper[4950]: I0318 20:50:02.970247 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29564450-xwsr8" podStartSLOduration=1.572117611 podStartE2EDuration="2.970228157s" podCreationTimestamp="2026-03-18 20:50:00 +0000 UTC" firstStartedPulling="2026-03-18 20:50:01.266667527 +0000 UTC m=+2614.507509395" lastFinishedPulling="2026-03-18 20:50:02.664778063 +0000 UTC m=+2615.905619941" observedRunningTime="2026-03-18 20:50:02.961671293 +0000 UTC m=+2616.202513151" watchObservedRunningTime="2026-03-18 20:50:02.970228157 +0000 UTC m=+2616.211070025" Mar 18 20:50:03 crc kubenswrapper[4950]: I0318 20:50:03.957553 4950 generic.go:334] "Generic (PLEG): container finished" podID="9d9d3129-446e-40a8-873f-f427bdd96d25" containerID="0aa0efd12d9ea9bbe4a547a35f9f99d1107d0539f9ef66bd867fa8cb4f44fc05" exitCode=0 Mar 18 20:50:03 crc kubenswrapper[4950]: I0318 20:50:03.958150 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564450-xwsr8" event={"ID":"9d9d3129-446e-40a8-873f-f427bdd96d25","Type":"ContainerDied","Data":"0aa0efd12d9ea9bbe4a547a35f9f99d1107d0539f9ef66bd867fa8cb4f44fc05"} Mar 18 20:50:05 crc kubenswrapper[4950]: I0318 20:50:05.308113 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564450-xwsr8" Mar 18 20:50:05 crc kubenswrapper[4950]: I0318 20:50:05.453036 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tp7gn\" (UniqueName: \"kubernetes.io/projected/9d9d3129-446e-40a8-873f-f427bdd96d25-kube-api-access-tp7gn\") pod \"9d9d3129-446e-40a8-873f-f427bdd96d25\" (UID: \"9d9d3129-446e-40a8-873f-f427bdd96d25\") " Mar 18 20:50:05 crc kubenswrapper[4950]: I0318 20:50:05.473391 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d9d3129-446e-40a8-873f-f427bdd96d25-kube-api-access-tp7gn" (OuterVolumeSpecName: "kube-api-access-tp7gn") pod "9d9d3129-446e-40a8-873f-f427bdd96d25" (UID: "9d9d3129-446e-40a8-873f-f427bdd96d25"). InnerVolumeSpecName "kube-api-access-tp7gn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:50:05 crc kubenswrapper[4950]: I0318 20:50:05.554974 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tp7gn\" (UniqueName: \"kubernetes.io/projected/9d9d3129-446e-40a8-873f-f427bdd96d25-kube-api-access-tp7gn\") on node \"crc\" DevicePath \"\"" Mar 18 20:50:05 crc kubenswrapper[4950]: I0318 20:50:05.976191 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564450-xwsr8" event={"ID":"9d9d3129-446e-40a8-873f-f427bdd96d25","Type":"ContainerDied","Data":"4393017ea33318dd6e5bce41335af3785fa6e84defbb47149530af0ae9714b83"} Mar 18 20:50:05 crc kubenswrapper[4950]: I0318 20:50:05.976486 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564450-xwsr8" Mar 18 20:50:05 crc kubenswrapper[4950]: I0318 20:50:05.976503 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4393017ea33318dd6e5bce41335af3785fa6e84defbb47149530af0ae9714b83" Mar 18 20:50:06 crc kubenswrapper[4950]: I0318 20:50:06.041567 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564444-7vvp5"] Mar 18 20:50:06 crc kubenswrapper[4950]: I0318 20:50:06.050229 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564444-7vvp5"] Mar 18 20:50:07 crc kubenswrapper[4950]: I0318 20:50:07.494974 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb7b1442-7b35-43f0-aa0c-84c5a8e6d2c9" path="/var/lib/kubelet/pods/eb7b1442-7b35-43f0-aa0c-84c5a8e6d2c9/volumes" Mar 18 20:50:07 crc kubenswrapper[4950]: I0318 20:50:07.494998 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:50:07 crc kubenswrapper[4950]: E0318 20:50:07.496018 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:50:21 crc kubenswrapper[4950]: I0318 20:50:21.481998 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:50:21 crc kubenswrapper[4950]: E0318 20:50:21.482657 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:50:36 crc kubenswrapper[4950]: I0318 20:50:36.480337 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:50:36 crc kubenswrapper[4950]: E0318 20:50:36.481115 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:50:46 crc kubenswrapper[4950]: I0318 20:50:46.357050 4950 generic.go:334] "Generic (PLEG): container finished" podID="67bb928c-890d-480c-8ce7-ec7b299cb1ee" containerID="81db0ed1bf206d71c8d9ec808ea1b58e81c264fbcb673c24eca613623519082c" exitCode=0 Mar 18 20:50:46 crc kubenswrapper[4950]: I0318 20:50:46.357686 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" event={"ID":"67bb928c-890d-480c-8ce7-ec7b299cb1ee","Type":"ContainerDied","Data":"81db0ed1bf206d71c8d9ec808ea1b58e81c264fbcb673c24eca613623519082c"} Mar 18 20:50:47 crc kubenswrapper[4950]: I0318 20:50:47.863880 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" Mar 18 20:50:47 crc kubenswrapper[4950]: I0318 20:50:47.890674 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjftp\" (UniqueName: \"kubernetes.io/projected/67bb928c-890d-480c-8ce7-ec7b299cb1ee-kube-api-access-fjftp\") pod \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " Mar 18 20:50:47 crc kubenswrapper[4950]: I0318 20:50:47.890736 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ceph\") pod \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " Mar 18 20:50:47 crc kubenswrapper[4950]: I0318 20:50:47.890821 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-inventory\") pod \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " Mar 18 20:50:47 crc kubenswrapper[4950]: I0318 20:50:47.890923 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ovncontroller-config-0\") pod \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " Mar 18 20:50:47 crc kubenswrapper[4950]: I0318 20:50:47.890976 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ovn-combined-ca-bundle\") pod \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " Mar 18 20:50:47 crc kubenswrapper[4950]: I0318 20:50:47.890994 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ssh-key-openstack-edpm-ipam\") pod \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\" (UID: \"67bb928c-890d-480c-8ce7-ec7b299cb1ee\") " Mar 18 20:50:47 crc kubenswrapper[4950]: I0318 20:50:47.899047 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ceph" (OuterVolumeSpecName: "ceph") pod "67bb928c-890d-480c-8ce7-ec7b299cb1ee" (UID: "67bb928c-890d-480c-8ce7-ec7b299cb1ee"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:50:47 crc kubenswrapper[4950]: I0318 20:50:47.905744 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67bb928c-890d-480c-8ce7-ec7b299cb1ee-kube-api-access-fjftp" (OuterVolumeSpecName: "kube-api-access-fjftp") pod "67bb928c-890d-480c-8ce7-ec7b299cb1ee" (UID: "67bb928c-890d-480c-8ce7-ec7b299cb1ee"). InnerVolumeSpecName "kube-api-access-fjftp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:50:47 crc kubenswrapper[4950]: I0318 20:50:47.909581 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "67bb928c-890d-480c-8ce7-ec7b299cb1ee" (UID: "67bb928c-890d-480c-8ce7-ec7b299cb1ee"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:50:47 crc kubenswrapper[4950]: I0318 20:50:47.926857 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-inventory" (OuterVolumeSpecName: "inventory") pod "67bb928c-890d-480c-8ce7-ec7b299cb1ee" (UID: "67bb928c-890d-480c-8ce7-ec7b299cb1ee"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:50:47 crc kubenswrapper[4950]: I0318 20:50:47.933121 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "67bb928c-890d-480c-8ce7-ec7b299cb1ee" (UID: "67bb928c-890d-480c-8ce7-ec7b299cb1ee"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:50:47 crc kubenswrapper[4950]: I0318 20:50:47.933561 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "67bb928c-890d-480c-8ce7-ec7b299cb1ee" (UID: "67bb928c-890d-480c-8ce7-ec7b299cb1ee"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:50:47 crc kubenswrapper[4950]: I0318 20:50:47.992869 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:50:47 crc kubenswrapper[4950]: I0318 20:50:47.992902 4950 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Mar 18 20:50:47 crc kubenswrapper[4950]: I0318 20:50:47.992912 4950 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:50:47 crc kubenswrapper[4950]: I0318 20:50:47.992922 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:50:47 crc kubenswrapper[4950]: I0318 20:50:47.992931 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjftp\" (UniqueName: \"kubernetes.io/projected/67bb928c-890d-480c-8ce7-ec7b299cb1ee-kube-api-access-fjftp\") on node \"crc\" DevicePath \"\"" Mar 18 20:50:47 crc kubenswrapper[4950]: I0318 20:50:47.992939 4950 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/67bb928c-890d-480c-8ce7-ec7b299cb1ee-ceph\") on node \"crc\" DevicePath \"\"" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.379991 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" event={"ID":"67bb928c-890d-480c-8ce7-ec7b299cb1ee","Type":"ContainerDied","Data":"68368e987009cac789876ef73b839d3f0c2d9b3e97c81ee1b06b51da40ec1289"} Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.380670 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68368e987009cac789876ef73b839d3f0c2d9b3e97c81ee1b06b51da40ec1289" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.380096 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-hnlg9" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.525993 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257"] Mar 18 20:50:48 crc kubenswrapper[4950]: E0318 20:50:48.527197 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d9d3129-446e-40a8-873f-f427bdd96d25" containerName="oc" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.527230 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d9d3129-446e-40a8-873f-f427bdd96d25" containerName="oc" Mar 18 20:50:48 crc kubenswrapper[4950]: E0318 20:50:48.527261 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67bb928c-890d-480c-8ce7-ec7b299cb1ee" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.527272 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="67bb928c-890d-480c-8ce7-ec7b299cb1ee" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.527682 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d9d3129-446e-40a8-873f-f427bdd96d25" containerName="oc" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.527728 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="67bb928c-890d-480c-8ce7-ec7b299cb1ee" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.528970 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.542076 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.542140 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.542181 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.542076 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.542283 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.542354 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.542661 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.547948 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257"] Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.606803 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.606872 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.606954 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px2xs\" (UniqueName: \"kubernetes.io/projected/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-kube-api-access-px2xs\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.607027 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.607124 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.607156 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.607181 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.709340 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.709567 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.709624 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.709662 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.709750 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.709870 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.710009 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px2xs\" (UniqueName: \"kubernetes.io/projected/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-kube-api-access-px2xs\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.717610 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.718032 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.721048 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.721358 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.721733 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.721979 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.727088 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px2xs\" (UniqueName: \"kubernetes.io/projected/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-kube-api-access-px2xs\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:50:48 crc kubenswrapper[4950]: I0318 20:50:48.863470 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:50:49 crc kubenswrapper[4950]: I0318 20:50:49.433217 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257"] Mar 18 20:50:49 crc kubenswrapper[4950]: W0318 20:50:49.444595 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod619538d3_bc36_4cfb_8eb7_9fad2bef5d4c.slice/crio-bc641fa445a78bc040db222e3cd44818fa08b5f19f34c85ad4b54270a2d8d081 WatchSource:0}: Error finding container bc641fa445a78bc040db222e3cd44818fa08b5f19f34c85ad4b54270a2d8d081: Status 404 returned error can't find the container with id bc641fa445a78bc040db222e3cd44818fa08b5f19f34c85ad4b54270a2d8d081 Mar 18 20:50:50 crc kubenswrapper[4950]: I0318 20:50:50.267904 4950 scope.go:117] "RemoveContainer" containerID="feb84a2195d48fdbead49f67c0f75b4e48bb65fe9e84653976a79583b17cbcc3" Mar 18 20:50:50 crc kubenswrapper[4950]: I0318 20:50:50.398669 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" event={"ID":"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c","Type":"ContainerStarted","Data":"c5981b347074ee64b2a69e6890a1c6be09384800733b69a87ff3ad1af9218e3a"} Mar 18 20:50:50 crc kubenswrapper[4950]: I0318 20:50:50.398708 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" event={"ID":"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c","Type":"ContainerStarted","Data":"bc641fa445a78bc040db222e3cd44818fa08b5f19f34c85ad4b54270a2d8d081"} Mar 18 20:50:50 crc kubenswrapper[4950]: I0318 20:50:50.415944 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" podStartSLOduration=1.797687588 podStartE2EDuration="2.415923348s" podCreationTimestamp="2026-03-18 20:50:48 +0000 UTC" firstStartedPulling="2026-03-18 20:50:49.450479297 +0000 UTC m=+2662.691321165" lastFinishedPulling="2026-03-18 20:50:50.068679756 +0000 UTC m=+2663.309556925" observedRunningTime="2026-03-18 20:50:50.413763081 +0000 UTC m=+2663.654604949" watchObservedRunningTime="2026-03-18 20:50:50.415923348 +0000 UTC m=+2663.656765226" Mar 18 20:50:51 crc kubenswrapper[4950]: I0318 20:50:51.479878 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:50:51 crc kubenswrapper[4950]: E0318 20:50:51.480193 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:51:06 crc kubenswrapper[4950]: I0318 20:51:06.480204 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:51:06 crc kubenswrapper[4950]: E0318 20:51:06.480965 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:51:21 crc kubenswrapper[4950]: I0318 20:51:21.480387 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:51:21 crc kubenswrapper[4950]: E0318 20:51:21.481600 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:51:35 crc kubenswrapper[4950]: I0318 20:51:35.485131 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:51:35 crc kubenswrapper[4950]: E0318 20:51:35.485988 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:51:47 crc kubenswrapper[4950]: I0318 20:51:47.485091 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:51:47 crc kubenswrapper[4950]: E0318 20:51:47.486081 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:51:51 crc kubenswrapper[4950]: I0318 20:51:51.012511 4950 generic.go:334] "Generic (PLEG): container finished" podID="619538d3-bc36-4cfb-8eb7-9fad2bef5d4c" containerID="c5981b347074ee64b2a69e6890a1c6be09384800733b69a87ff3ad1af9218e3a" exitCode=0 Mar 18 20:51:51 crc kubenswrapper[4950]: I0318 20:51:51.012547 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" event={"ID":"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c","Type":"ContainerDied","Data":"c5981b347074ee64b2a69e6890a1c6be09384800733b69a87ff3ad1af9218e3a"} Mar 18 20:51:52 crc kubenswrapper[4950]: I0318 20:51:52.406427 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:51:52 crc kubenswrapper[4950]: I0318 20:51:52.502636 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-ssh-key-openstack-edpm-ipam\") pod \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " Mar 18 20:51:52 crc kubenswrapper[4950]: I0318 20:51:52.503027 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-inventory\") pod \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " Mar 18 20:51:52 crc kubenswrapper[4950]: I0318 20:51:52.503130 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-nova-metadata-neutron-config-0\") pod \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " Mar 18 20:51:52 crc kubenswrapper[4950]: I0318 20:51:52.503224 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " Mar 18 20:51:52 crc kubenswrapper[4950]: I0318 20:51:52.503262 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px2xs\" (UniqueName: \"kubernetes.io/projected/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-kube-api-access-px2xs\") pod \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " Mar 18 20:51:52 crc kubenswrapper[4950]: I0318 20:51:52.503295 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-neutron-metadata-combined-ca-bundle\") pod \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " Mar 18 20:51:52 crc kubenswrapper[4950]: I0318 20:51:52.503317 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-ceph\") pod \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\" (UID: \"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c\") " Mar 18 20:51:52 crc kubenswrapper[4950]: I0318 20:51:52.508598 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-kube-api-access-px2xs" (OuterVolumeSpecName: "kube-api-access-px2xs") pod "619538d3-bc36-4cfb-8eb7-9fad2bef5d4c" (UID: "619538d3-bc36-4cfb-8eb7-9fad2bef5d4c"). InnerVolumeSpecName "kube-api-access-px2xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:51:52 crc kubenswrapper[4950]: I0318 20:51:52.508950 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "619538d3-bc36-4cfb-8eb7-9fad2bef5d4c" (UID: "619538d3-bc36-4cfb-8eb7-9fad2bef5d4c"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:51:52 crc kubenswrapper[4950]: I0318 20:51:52.509625 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-ceph" (OuterVolumeSpecName: "ceph") pod "619538d3-bc36-4cfb-8eb7-9fad2bef5d4c" (UID: "619538d3-bc36-4cfb-8eb7-9fad2bef5d4c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:51:52 crc kubenswrapper[4950]: I0318 20:51:52.531215 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "619538d3-bc36-4cfb-8eb7-9fad2bef5d4c" (UID: "619538d3-bc36-4cfb-8eb7-9fad2bef5d4c"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:51:52 crc kubenswrapper[4950]: I0318 20:51:52.540547 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "619538d3-bc36-4cfb-8eb7-9fad2bef5d4c" (UID: "619538d3-bc36-4cfb-8eb7-9fad2bef5d4c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:51:52 crc kubenswrapper[4950]: I0318 20:51:52.542871 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "619538d3-bc36-4cfb-8eb7-9fad2bef5d4c" (UID: "619538d3-bc36-4cfb-8eb7-9fad2bef5d4c"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:51:52 crc kubenswrapper[4950]: I0318 20:51:52.543036 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-inventory" (OuterVolumeSpecName: "inventory") pod "619538d3-bc36-4cfb-8eb7-9fad2bef5d4c" (UID: "619538d3-bc36-4cfb-8eb7-9fad2bef5d4c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:51:52 crc kubenswrapper[4950]: I0318 20:51:52.605888 4950 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 18 20:51:52 crc kubenswrapper[4950]: I0318 20:51:52.605920 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px2xs\" (UniqueName: \"kubernetes.io/projected/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-kube-api-access-px2xs\") on node \"crc\" DevicePath \"\"" Mar 18 20:51:52 crc kubenswrapper[4950]: I0318 20:51:52.605937 4950 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:51:52 crc kubenswrapper[4950]: I0318 20:51:52.605950 4950 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-ceph\") on node \"crc\" DevicePath \"\"" Mar 18 20:51:52 crc kubenswrapper[4950]: I0318 20:51:52.605964 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:51:52 crc kubenswrapper[4950]: I0318 20:51:52.605978 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:51:52 crc kubenswrapper[4950]: I0318 20:51:52.605991 4950 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/619538d3-bc36-4cfb-8eb7-9fad2bef5d4c-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.039088 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" event={"ID":"619538d3-bc36-4cfb-8eb7-9fad2bef5d4c","Type":"ContainerDied","Data":"bc641fa445a78bc040db222e3cd44818fa08b5f19f34c85ad4b54270a2d8d081"} Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.039182 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc641fa445a78bc040db222e3cd44818fa08b5f19f34c85ad4b54270a2d8d081" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.039216 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.291249 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp"] Mar 18 20:51:53 crc kubenswrapper[4950]: E0318 20:51:53.292177 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="619538d3-bc36-4cfb-8eb7-9fad2bef5d4c" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.292314 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="619538d3-bc36-4cfb-8eb7-9fad2bef5d4c" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.292735 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="619538d3-bc36-4cfb-8eb7-9fad2bef5d4c" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.293973 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.300849 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.302290 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.303926 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.304289 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.304671 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.306118 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.306117 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp"] Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.420324 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkcwc\" (UniqueName: \"kubernetes.io/projected/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-kube-api-access-fkcwc\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.420424 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.420458 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.420522 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.420556 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.420607 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.522332 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.522387 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.522456 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.522515 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkcwc\" (UniqueName: \"kubernetes.io/projected/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-kube-api-access-fkcwc\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.522569 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.522597 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.527688 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.531337 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.531535 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.531561 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.533755 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.543183 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkcwc\" (UniqueName: \"kubernetes.io/projected/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-kube-api-access-fkcwc\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" Mar 18 20:51:53 crc kubenswrapper[4950]: I0318 20:51:53.614954 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" Mar 18 20:51:54 crc kubenswrapper[4950]: I0318 20:51:54.120481 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp"] Mar 18 20:51:54 crc kubenswrapper[4950]: W0318 20:51:54.125625 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22d4a06f_ff3c_4fbf_9eae_8be967f1febd.slice/crio-ab99569365a39678fde7fc1ee6c6adba41ecbfec1bd62fa0d0f68c9b33a42121 WatchSource:0}: Error finding container ab99569365a39678fde7fc1ee6c6adba41ecbfec1bd62fa0d0f68c9b33a42121: Status 404 returned error can't find the container with id ab99569365a39678fde7fc1ee6c6adba41ecbfec1bd62fa0d0f68c9b33a42121 Mar 18 20:51:55 crc kubenswrapper[4950]: I0318 20:51:55.067142 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" event={"ID":"22d4a06f-ff3c-4fbf-9eae-8be967f1febd","Type":"ContainerStarted","Data":"ab99569365a39678fde7fc1ee6c6adba41ecbfec1bd62fa0d0f68c9b33a42121"} Mar 18 20:51:56 crc kubenswrapper[4950]: I0318 20:51:56.081151 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" event={"ID":"22d4a06f-ff3c-4fbf-9eae-8be967f1febd","Type":"ContainerStarted","Data":"4c80b144d282d4fdab6d4b5821131aacdba68df40fdc5be615a802fb4ab42abf"} Mar 18 20:51:56 crc kubenswrapper[4950]: I0318 20:51:56.109676 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" podStartSLOduration=2.411279609 podStartE2EDuration="3.109645715s" podCreationTimestamp="2026-03-18 20:51:53 +0000 UTC" firstStartedPulling="2026-03-18 20:51:54.127442086 +0000 UTC m=+2727.368283974" lastFinishedPulling="2026-03-18 20:51:54.825808202 +0000 UTC m=+2728.066650080" observedRunningTime="2026-03-18 20:51:56.103573666 +0000 UTC m=+2729.344415554" watchObservedRunningTime="2026-03-18 20:51:56.109645715 +0000 UTC m=+2729.350487623" Mar 18 20:51:59 crc kubenswrapper[4950]: I0318 20:51:59.480600 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:51:59 crc kubenswrapper[4950]: E0318 20:51:59.482514 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:52:00 crc kubenswrapper[4950]: I0318 20:52:00.161951 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564452-f997f"] Mar 18 20:52:00 crc kubenswrapper[4950]: I0318 20:52:00.163626 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564452-f997f" Mar 18 20:52:00 crc kubenswrapper[4950]: I0318 20:52:00.170369 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:52:00 crc kubenswrapper[4950]: I0318 20:52:00.170991 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:52:00 crc kubenswrapper[4950]: I0318 20:52:00.171476 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:52:00 crc kubenswrapper[4950]: I0318 20:52:00.182371 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564452-f997f"] Mar 18 20:52:00 crc kubenswrapper[4950]: I0318 20:52:00.252822 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xn97\" (UniqueName: \"kubernetes.io/projected/3d88011d-78a5-45a0-8cc9-939233b4cad4-kube-api-access-5xn97\") pod \"auto-csr-approver-29564452-f997f\" (UID: \"3d88011d-78a5-45a0-8cc9-939233b4cad4\") " pod="openshift-infra/auto-csr-approver-29564452-f997f" Mar 18 20:52:00 crc kubenswrapper[4950]: I0318 20:52:00.354962 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xn97\" (UniqueName: \"kubernetes.io/projected/3d88011d-78a5-45a0-8cc9-939233b4cad4-kube-api-access-5xn97\") pod \"auto-csr-approver-29564452-f997f\" (UID: \"3d88011d-78a5-45a0-8cc9-939233b4cad4\") " pod="openshift-infra/auto-csr-approver-29564452-f997f" Mar 18 20:52:00 crc kubenswrapper[4950]: I0318 20:52:00.381577 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xn97\" (UniqueName: \"kubernetes.io/projected/3d88011d-78a5-45a0-8cc9-939233b4cad4-kube-api-access-5xn97\") pod \"auto-csr-approver-29564452-f997f\" (UID: \"3d88011d-78a5-45a0-8cc9-939233b4cad4\") " pod="openshift-infra/auto-csr-approver-29564452-f997f" Mar 18 20:52:00 crc kubenswrapper[4950]: I0318 20:52:00.494561 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564452-f997f" Mar 18 20:52:00 crc kubenswrapper[4950]: I0318 20:52:00.976108 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564452-f997f"] Mar 18 20:52:01 crc kubenswrapper[4950]: I0318 20:52:01.130917 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564452-f997f" event={"ID":"3d88011d-78a5-45a0-8cc9-939233b4cad4","Type":"ContainerStarted","Data":"b62bf79588971136c522f6322771eb024a507b6df294db8e51bb45ed5c3516a9"} Mar 18 20:52:03 crc kubenswrapper[4950]: I0318 20:52:03.155821 4950 generic.go:334] "Generic (PLEG): container finished" podID="3d88011d-78a5-45a0-8cc9-939233b4cad4" containerID="33e379d85361962e47f13d952ee22ca3a03d974efd14d056aac0043f264a18f4" exitCode=0 Mar 18 20:52:03 crc kubenswrapper[4950]: I0318 20:52:03.155899 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564452-f997f" event={"ID":"3d88011d-78a5-45a0-8cc9-939233b4cad4","Type":"ContainerDied","Data":"33e379d85361962e47f13d952ee22ca3a03d974efd14d056aac0043f264a18f4"} Mar 18 20:52:04 crc kubenswrapper[4950]: I0318 20:52:04.655532 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564452-f997f" Mar 18 20:52:04 crc kubenswrapper[4950]: I0318 20:52:04.735808 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xn97\" (UniqueName: \"kubernetes.io/projected/3d88011d-78a5-45a0-8cc9-939233b4cad4-kube-api-access-5xn97\") pod \"3d88011d-78a5-45a0-8cc9-939233b4cad4\" (UID: \"3d88011d-78a5-45a0-8cc9-939233b4cad4\") " Mar 18 20:52:04 crc kubenswrapper[4950]: I0318 20:52:04.743106 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d88011d-78a5-45a0-8cc9-939233b4cad4-kube-api-access-5xn97" (OuterVolumeSpecName: "kube-api-access-5xn97") pod "3d88011d-78a5-45a0-8cc9-939233b4cad4" (UID: "3d88011d-78a5-45a0-8cc9-939233b4cad4"). InnerVolumeSpecName "kube-api-access-5xn97". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:52:04 crc kubenswrapper[4950]: I0318 20:52:04.842183 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xn97\" (UniqueName: \"kubernetes.io/projected/3d88011d-78a5-45a0-8cc9-939233b4cad4-kube-api-access-5xn97\") on node \"crc\" DevicePath \"\"" Mar 18 20:52:05 crc kubenswrapper[4950]: I0318 20:52:05.176185 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564452-f997f" event={"ID":"3d88011d-78a5-45a0-8cc9-939233b4cad4","Type":"ContainerDied","Data":"b62bf79588971136c522f6322771eb024a507b6df294db8e51bb45ed5c3516a9"} Mar 18 20:52:05 crc kubenswrapper[4950]: I0318 20:52:05.176614 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b62bf79588971136c522f6322771eb024a507b6df294db8e51bb45ed5c3516a9" Mar 18 20:52:05 crc kubenswrapper[4950]: I0318 20:52:05.176248 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564452-f997f" Mar 18 20:52:05 crc kubenswrapper[4950]: I0318 20:52:05.729594 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564446-zpc8l"] Mar 18 20:52:05 crc kubenswrapper[4950]: I0318 20:52:05.738984 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564446-zpc8l"] Mar 18 20:52:07 crc kubenswrapper[4950]: I0318 20:52:07.499629 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ab7b23e-8e83-4f5a-a7dd-e7b2a6ffcb94" path="/var/lib/kubelet/pods/0ab7b23e-8e83-4f5a-a7dd-e7b2a6ffcb94/volumes" Mar 18 20:52:14 crc kubenswrapper[4950]: I0318 20:52:14.480312 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:52:14 crc kubenswrapper[4950]: E0318 20:52:14.481084 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:52:28 crc kubenswrapper[4950]: I0318 20:52:28.479315 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:52:28 crc kubenswrapper[4950]: E0318 20:52:28.480018 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:52:39 crc kubenswrapper[4950]: I0318 20:52:39.479689 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:52:39 crc kubenswrapper[4950]: E0318 20:52:39.480451 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:52:50 crc kubenswrapper[4950]: I0318 20:52:50.364008 4950 scope.go:117] "RemoveContainer" containerID="cd20209c9a713c1281bce5860c56014ea0a7da9d820a225d3a2b432bfa9edb30" Mar 18 20:52:54 crc kubenswrapper[4950]: I0318 20:52:54.480331 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:52:54 crc kubenswrapper[4950]: E0318 20:52:54.481020 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 20:53:07 crc kubenswrapper[4950]: I0318 20:53:07.493540 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:53:08 crc kubenswrapper[4950]: I0318 20:53:08.427021 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerStarted","Data":"659b90655ee76a1c728bf80c079e29e36775b2f88753e6ed85575c3829a23583"} Mar 18 20:54:00 crc kubenswrapper[4950]: I0318 20:54:00.148281 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564454-jll6h"] Mar 18 20:54:00 crc kubenswrapper[4950]: E0318 20:54:00.150246 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d88011d-78a5-45a0-8cc9-939233b4cad4" containerName="oc" Mar 18 20:54:00 crc kubenswrapper[4950]: I0318 20:54:00.150353 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d88011d-78a5-45a0-8cc9-939233b4cad4" containerName="oc" Mar 18 20:54:00 crc kubenswrapper[4950]: I0318 20:54:00.150733 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d88011d-78a5-45a0-8cc9-939233b4cad4" containerName="oc" Mar 18 20:54:00 crc kubenswrapper[4950]: I0318 20:54:00.151619 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564454-jll6h" Mar 18 20:54:00 crc kubenswrapper[4950]: I0318 20:54:00.153372 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:54:00 crc kubenswrapper[4950]: I0318 20:54:00.154342 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:54:00 crc kubenswrapper[4950]: I0318 20:54:00.154995 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:54:00 crc kubenswrapper[4950]: I0318 20:54:00.162693 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564454-jll6h"] Mar 18 20:54:00 crc kubenswrapper[4950]: I0318 20:54:00.287481 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkhsk\" (UniqueName: \"kubernetes.io/projected/b83aea1b-8a6b-4d04-8f82-2293bc696179-kube-api-access-wkhsk\") pod \"auto-csr-approver-29564454-jll6h\" (UID: \"b83aea1b-8a6b-4d04-8f82-2293bc696179\") " pod="openshift-infra/auto-csr-approver-29564454-jll6h" Mar 18 20:54:00 crc kubenswrapper[4950]: I0318 20:54:00.389365 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkhsk\" (UniqueName: \"kubernetes.io/projected/b83aea1b-8a6b-4d04-8f82-2293bc696179-kube-api-access-wkhsk\") pod \"auto-csr-approver-29564454-jll6h\" (UID: \"b83aea1b-8a6b-4d04-8f82-2293bc696179\") " pod="openshift-infra/auto-csr-approver-29564454-jll6h" Mar 18 20:54:00 crc kubenswrapper[4950]: I0318 20:54:00.406147 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkhsk\" (UniqueName: \"kubernetes.io/projected/b83aea1b-8a6b-4d04-8f82-2293bc696179-kube-api-access-wkhsk\") pod \"auto-csr-approver-29564454-jll6h\" (UID: \"b83aea1b-8a6b-4d04-8f82-2293bc696179\") " pod="openshift-infra/auto-csr-approver-29564454-jll6h" Mar 18 20:54:00 crc kubenswrapper[4950]: I0318 20:54:00.469360 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564454-jll6h" Mar 18 20:54:00 crc kubenswrapper[4950]: I0318 20:54:00.942770 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564454-jll6h"] Mar 18 20:54:00 crc kubenswrapper[4950]: W0318 20:54:00.948743 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb83aea1b_8a6b_4d04_8f82_2293bc696179.slice/crio-fc680673a54a2820221628c5be0835cba2faed001772962a6b7632ed300a3175 WatchSource:0}: Error finding container fc680673a54a2820221628c5be0835cba2faed001772962a6b7632ed300a3175: Status 404 returned error can't find the container with id fc680673a54a2820221628c5be0835cba2faed001772962a6b7632ed300a3175 Mar 18 20:54:00 crc kubenswrapper[4950]: I0318 20:54:00.951288 4950 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 20:54:01 crc kubenswrapper[4950]: I0318 20:54:01.914404 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564454-jll6h" event={"ID":"b83aea1b-8a6b-4d04-8f82-2293bc696179","Type":"ContainerStarted","Data":"fc680673a54a2820221628c5be0835cba2faed001772962a6b7632ed300a3175"} Mar 18 20:54:02 crc kubenswrapper[4950]: I0318 20:54:02.926808 4950 generic.go:334] "Generic (PLEG): container finished" podID="b83aea1b-8a6b-4d04-8f82-2293bc696179" containerID="72e6b8ccee81fc70d3acc8fc40745fd0adb6fac9070e7724d58751929bc1766a" exitCode=0 Mar 18 20:54:02 crc kubenswrapper[4950]: I0318 20:54:02.926891 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564454-jll6h" event={"ID":"b83aea1b-8a6b-4d04-8f82-2293bc696179","Type":"ContainerDied","Data":"72e6b8ccee81fc70d3acc8fc40745fd0adb6fac9070e7724d58751929bc1766a"} Mar 18 20:54:04 crc kubenswrapper[4950]: I0318 20:54:04.704427 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564454-jll6h" Mar 18 20:54:04 crc kubenswrapper[4950]: I0318 20:54:04.886829 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkhsk\" (UniqueName: \"kubernetes.io/projected/b83aea1b-8a6b-4d04-8f82-2293bc696179-kube-api-access-wkhsk\") pod \"b83aea1b-8a6b-4d04-8f82-2293bc696179\" (UID: \"b83aea1b-8a6b-4d04-8f82-2293bc696179\") " Mar 18 20:54:04 crc kubenswrapper[4950]: I0318 20:54:04.893976 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b83aea1b-8a6b-4d04-8f82-2293bc696179-kube-api-access-wkhsk" (OuterVolumeSpecName: "kube-api-access-wkhsk") pod "b83aea1b-8a6b-4d04-8f82-2293bc696179" (UID: "b83aea1b-8a6b-4d04-8f82-2293bc696179"). InnerVolumeSpecName "kube-api-access-wkhsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:54:04 crc kubenswrapper[4950]: I0318 20:54:04.949009 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564454-jll6h" event={"ID":"b83aea1b-8a6b-4d04-8f82-2293bc696179","Type":"ContainerDied","Data":"fc680673a54a2820221628c5be0835cba2faed001772962a6b7632ed300a3175"} Mar 18 20:54:04 crc kubenswrapper[4950]: I0318 20:54:04.949045 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc680673a54a2820221628c5be0835cba2faed001772962a6b7632ed300a3175" Mar 18 20:54:04 crc kubenswrapper[4950]: I0318 20:54:04.949058 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564454-jll6h" Mar 18 20:54:04 crc kubenswrapper[4950]: I0318 20:54:04.989210 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkhsk\" (UniqueName: \"kubernetes.io/projected/b83aea1b-8a6b-4d04-8f82-2293bc696179-kube-api-access-wkhsk\") on node \"crc\" DevicePath \"\"" Mar 18 20:54:05 crc kubenswrapper[4950]: I0318 20:54:05.775369 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564448-cbwbn"] Mar 18 20:54:05 crc kubenswrapper[4950]: I0318 20:54:05.786096 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564448-cbwbn"] Mar 18 20:54:07 crc kubenswrapper[4950]: I0318 20:54:07.489962 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26ce5d7b-072e-407e-b8c5-4b5610711ee7" path="/var/lib/kubelet/pods/26ce5d7b-072e-407e-b8c5-4b5610711ee7/volumes" Mar 18 20:54:20 crc kubenswrapper[4950]: I0318 20:54:20.716315 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nxrsl"] Mar 18 20:54:20 crc kubenswrapper[4950]: E0318 20:54:20.717699 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b83aea1b-8a6b-4d04-8f82-2293bc696179" containerName="oc" Mar 18 20:54:20 crc kubenswrapper[4950]: I0318 20:54:20.717721 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="b83aea1b-8a6b-4d04-8f82-2293bc696179" containerName="oc" Mar 18 20:54:20 crc kubenswrapper[4950]: I0318 20:54:20.718016 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="b83aea1b-8a6b-4d04-8f82-2293bc696179" containerName="oc" Mar 18 20:54:20 crc kubenswrapper[4950]: I0318 20:54:20.719613 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nxrsl" Mar 18 20:54:20 crc kubenswrapper[4950]: I0318 20:54:20.751374 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nxrsl"] Mar 18 20:54:20 crc kubenswrapper[4950]: I0318 20:54:20.793866 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxzfs\" (UniqueName: \"kubernetes.io/projected/81564706-d904-4cca-905f-1169d42c46fc-kube-api-access-sxzfs\") pod \"certified-operators-nxrsl\" (UID: \"81564706-d904-4cca-905f-1169d42c46fc\") " pod="openshift-marketplace/certified-operators-nxrsl" Mar 18 20:54:20 crc kubenswrapper[4950]: I0318 20:54:20.793941 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81564706-d904-4cca-905f-1169d42c46fc-utilities\") pod \"certified-operators-nxrsl\" (UID: \"81564706-d904-4cca-905f-1169d42c46fc\") " pod="openshift-marketplace/certified-operators-nxrsl" Mar 18 20:54:20 crc kubenswrapper[4950]: I0318 20:54:20.794051 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81564706-d904-4cca-905f-1169d42c46fc-catalog-content\") pod \"certified-operators-nxrsl\" (UID: \"81564706-d904-4cca-905f-1169d42c46fc\") " pod="openshift-marketplace/certified-operators-nxrsl" Mar 18 20:54:20 crc kubenswrapper[4950]: I0318 20:54:20.896207 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxzfs\" (UniqueName: \"kubernetes.io/projected/81564706-d904-4cca-905f-1169d42c46fc-kube-api-access-sxzfs\") pod \"certified-operators-nxrsl\" (UID: \"81564706-d904-4cca-905f-1169d42c46fc\") " pod="openshift-marketplace/certified-operators-nxrsl" Mar 18 20:54:20 crc kubenswrapper[4950]: I0318 20:54:20.896257 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81564706-d904-4cca-905f-1169d42c46fc-utilities\") pod \"certified-operators-nxrsl\" (UID: \"81564706-d904-4cca-905f-1169d42c46fc\") " pod="openshift-marketplace/certified-operators-nxrsl" Mar 18 20:54:20 crc kubenswrapper[4950]: I0318 20:54:20.896288 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81564706-d904-4cca-905f-1169d42c46fc-catalog-content\") pod \"certified-operators-nxrsl\" (UID: \"81564706-d904-4cca-905f-1169d42c46fc\") " pod="openshift-marketplace/certified-operators-nxrsl" Mar 18 20:54:20 crc kubenswrapper[4950]: I0318 20:54:20.896956 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81564706-d904-4cca-905f-1169d42c46fc-utilities\") pod \"certified-operators-nxrsl\" (UID: \"81564706-d904-4cca-905f-1169d42c46fc\") " pod="openshift-marketplace/certified-operators-nxrsl" Mar 18 20:54:20 crc kubenswrapper[4950]: I0318 20:54:20.896985 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81564706-d904-4cca-905f-1169d42c46fc-catalog-content\") pod \"certified-operators-nxrsl\" (UID: \"81564706-d904-4cca-905f-1169d42c46fc\") " pod="openshift-marketplace/certified-operators-nxrsl" Mar 18 20:54:20 crc kubenswrapper[4950]: I0318 20:54:20.920919 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxzfs\" (UniqueName: \"kubernetes.io/projected/81564706-d904-4cca-905f-1169d42c46fc-kube-api-access-sxzfs\") pod \"certified-operators-nxrsl\" (UID: \"81564706-d904-4cca-905f-1169d42c46fc\") " pod="openshift-marketplace/certified-operators-nxrsl" Mar 18 20:54:21 crc kubenswrapper[4950]: I0318 20:54:21.041680 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nxrsl" Mar 18 20:54:21 crc kubenswrapper[4950]: I0318 20:54:21.688177 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nxrsl"] Mar 18 20:54:22 crc kubenswrapper[4950]: I0318 20:54:22.100967 4950 generic.go:334] "Generic (PLEG): container finished" podID="81564706-d904-4cca-905f-1169d42c46fc" containerID="af5be9add612244bb835207ff50e169710fd5b8350c603e07fca7c15a2059467" exitCode=0 Mar 18 20:54:22 crc kubenswrapper[4950]: I0318 20:54:22.101575 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nxrsl" event={"ID":"81564706-d904-4cca-905f-1169d42c46fc","Type":"ContainerDied","Data":"af5be9add612244bb835207ff50e169710fd5b8350c603e07fca7c15a2059467"} Mar 18 20:54:22 crc kubenswrapper[4950]: I0318 20:54:22.102377 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nxrsl" event={"ID":"81564706-d904-4cca-905f-1169d42c46fc","Type":"ContainerStarted","Data":"286973345bfc5987f2bebb2cc4297e9cc0f62ed8aa86df2b5702631f017994c9"} Mar 18 20:54:23 crc kubenswrapper[4950]: I0318 20:54:23.111022 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nxrsl" event={"ID":"81564706-d904-4cca-905f-1169d42c46fc","Type":"ContainerStarted","Data":"de454f70da5fb3eaae1d4619d47f7035912417c6dab3afc4c68ffdc8c81f9482"} Mar 18 20:54:25 crc kubenswrapper[4950]: I0318 20:54:25.130651 4950 generic.go:334] "Generic (PLEG): container finished" podID="81564706-d904-4cca-905f-1169d42c46fc" containerID="de454f70da5fb3eaae1d4619d47f7035912417c6dab3afc4c68ffdc8c81f9482" exitCode=0 Mar 18 20:54:25 crc kubenswrapper[4950]: I0318 20:54:25.130730 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nxrsl" event={"ID":"81564706-d904-4cca-905f-1169d42c46fc","Type":"ContainerDied","Data":"de454f70da5fb3eaae1d4619d47f7035912417c6dab3afc4c68ffdc8c81f9482"} Mar 18 20:54:26 crc kubenswrapper[4950]: I0318 20:54:26.140020 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nxrsl" event={"ID":"81564706-d904-4cca-905f-1169d42c46fc","Type":"ContainerStarted","Data":"adfbf2423bd52b7fe7927edebc86cab44709d4d50c00fb3871acf81f1218627e"} Mar 18 20:54:26 crc kubenswrapper[4950]: I0318 20:54:26.162895 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nxrsl" podStartSLOduration=2.4809103 podStartE2EDuration="6.162876899s" podCreationTimestamp="2026-03-18 20:54:20 +0000 UTC" firstStartedPulling="2026-03-18 20:54:22.103888927 +0000 UTC m=+2875.344730795" lastFinishedPulling="2026-03-18 20:54:25.785855516 +0000 UTC m=+2879.026697394" observedRunningTime="2026-03-18 20:54:26.160202579 +0000 UTC m=+2879.401044447" watchObservedRunningTime="2026-03-18 20:54:26.162876899 +0000 UTC m=+2879.403718777" Mar 18 20:54:31 crc kubenswrapper[4950]: I0318 20:54:31.042546 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nxrsl" Mar 18 20:54:31 crc kubenswrapper[4950]: I0318 20:54:31.043091 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nxrsl" Mar 18 20:54:31 crc kubenswrapper[4950]: I0318 20:54:31.092576 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nxrsl" Mar 18 20:54:31 crc kubenswrapper[4950]: I0318 20:54:31.240821 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nxrsl" Mar 18 20:54:31 crc kubenswrapper[4950]: I0318 20:54:31.327649 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nxrsl"] Mar 18 20:54:33 crc kubenswrapper[4950]: I0318 20:54:33.203739 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nxrsl" podUID="81564706-d904-4cca-905f-1169d42c46fc" containerName="registry-server" containerID="cri-o://adfbf2423bd52b7fe7927edebc86cab44709d4d50c00fb3871acf81f1218627e" gracePeriod=2 Mar 18 20:54:33 crc kubenswrapper[4950]: I0318 20:54:33.708566 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nxrsl" Mar 18 20:54:33 crc kubenswrapper[4950]: I0318 20:54:33.847609 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81564706-d904-4cca-905f-1169d42c46fc-utilities\") pod \"81564706-d904-4cca-905f-1169d42c46fc\" (UID: \"81564706-d904-4cca-905f-1169d42c46fc\") " Mar 18 20:54:33 crc kubenswrapper[4950]: I0318 20:54:33.847716 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81564706-d904-4cca-905f-1169d42c46fc-catalog-content\") pod \"81564706-d904-4cca-905f-1169d42c46fc\" (UID: \"81564706-d904-4cca-905f-1169d42c46fc\") " Mar 18 20:54:33 crc kubenswrapper[4950]: I0318 20:54:33.849154 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81564706-d904-4cca-905f-1169d42c46fc-utilities" (OuterVolumeSpecName: "utilities") pod "81564706-d904-4cca-905f-1169d42c46fc" (UID: "81564706-d904-4cca-905f-1169d42c46fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:54:33 crc kubenswrapper[4950]: I0318 20:54:33.853767 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxzfs\" (UniqueName: \"kubernetes.io/projected/81564706-d904-4cca-905f-1169d42c46fc-kube-api-access-sxzfs\") pod \"81564706-d904-4cca-905f-1169d42c46fc\" (UID: \"81564706-d904-4cca-905f-1169d42c46fc\") " Mar 18 20:54:33 crc kubenswrapper[4950]: I0318 20:54:33.854697 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81564706-d904-4cca-905f-1169d42c46fc-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:54:33 crc kubenswrapper[4950]: I0318 20:54:33.860423 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81564706-d904-4cca-905f-1169d42c46fc-kube-api-access-sxzfs" (OuterVolumeSpecName: "kube-api-access-sxzfs") pod "81564706-d904-4cca-905f-1169d42c46fc" (UID: "81564706-d904-4cca-905f-1169d42c46fc"). InnerVolumeSpecName "kube-api-access-sxzfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:54:33 crc kubenswrapper[4950]: I0318 20:54:33.901730 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81564706-d904-4cca-905f-1169d42c46fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "81564706-d904-4cca-905f-1169d42c46fc" (UID: "81564706-d904-4cca-905f-1169d42c46fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:54:33 crc kubenswrapper[4950]: I0318 20:54:33.956766 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81564706-d904-4cca-905f-1169d42c46fc-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:54:33 crc kubenswrapper[4950]: I0318 20:54:33.956807 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxzfs\" (UniqueName: \"kubernetes.io/projected/81564706-d904-4cca-905f-1169d42c46fc-kube-api-access-sxzfs\") on node \"crc\" DevicePath \"\"" Mar 18 20:54:34 crc kubenswrapper[4950]: I0318 20:54:34.212389 4950 generic.go:334] "Generic (PLEG): container finished" podID="81564706-d904-4cca-905f-1169d42c46fc" containerID="adfbf2423bd52b7fe7927edebc86cab44709d4d50c00fb3871acf81f1218627e" exitCode=0 Mar 18 20:54:34 crc kubenswrapper[4950]: I0318 20:54:34.212444 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nxrsl" event={"ID":"81564706-d904-4cca-905f-1169d42c46fc","Type":"ContainerDied","Data":"adfbf2423bd52b7fe7927edebc86cab44709d4d50c00fb3871acf81f1218627e"} Mar 18 20:54:34 crc kubenswrapper[4950]: I0318 20:54:34.212470 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nxrsl" event={"ID":"81564706-d904-4cca-905f-1169d42c46fc","Type":"ContainerDied","Data":"286973345bfc5987f2bebb2cc4297e9cc0f62ed8aa86df2b5702631f017994c9"} Mar 18 20:54:34 crc kubenswrapper[4950]: I0318 20:54:34.212477 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nxrsl" Mar 18 20:54:34 crc kubenswrapper[4950]: I0318 20:54:34.212485 4950 scope.go:117] "RemoveContainer" containerID="adfbf2423bd52b7fe7927edebc86cab44709d4d50c00fb3871acf81f1218627e" Mar 18 20:54:34 crc kubenswrapper[4950]: I0318 20:54:34.234001 4950 scope.go:117] "RemoveContainer" containerID="de454f70da5fb3eaae1d4619d47f7035912417c6dab3afc4c68ffdc8c81f9482" Mar 18 20:54:34 crc kubenswrapper[4950]: I0318 20:54:34.256176 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nxrsl"] Mar 18 20:54:34 crc kubenswrapper[4950]: I0318 20:54:34.266569 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nxrsl"] Mar 18 20:54:34 crc kubenswrapper[4950]: I0318 20:54:34.269304 4950 scope.go:117] "RemoveContainer" containerID="af5be9add612244bb835207ff50e169710fd5b8350c603e07fca7c15a2059467" Mar 18 20:54:34 crc kubenswrapper[4950]: I0318 20:54:34.308166 4950 scope.go:117] "RemoveContainer" containerID="adfbf2423bd52b7fe7927edebc86cab44709d4d50c00fb3871acf81f1218627e" Mar 18 20:54:34 crc kubenswrapper[4950]: E0318 20:54:34.308590 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adfbf2423bd52b7fe7927edebc86cab44709d4d50c00fb3871acf81f1218627e\": container with ID starting with adfbf2423bd52b7fe7927edebc86cab44709d4d50c00fb3871acf81f1218627e not found: ID does not exist" containerID="adfbf2423bd52b7fe7927edebc86cab44709d4d50c00fb3871acf81f1218627e" Mar 18 20:54:34 crc kubenswrapper[4950]: I0318 20:54:34.308738 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adfbf2423bd52b7fe7927edebc86cab44709d4d50c00fb3871acf81f1218627e"} err="failed to get container status \"adfbf2423bd52b7fe7927edebc86cab44709d4d50c00fb3871acf81f1218627e\": rpc error: code = NotFound desc = could not find container \"adfbf2423bd52b7fe7927edebc86cab44709d4d50c00fb3871acf81f1218627e\": container with ID starting with adfbf2423bd52b7fe7927edebc86cab44709d4d50c00fb3871acf81f1218627e not found: ID does not exist" Mar 18 20:54:34 crc kubenswrapper[4950]: I0318 20:54:34.308890 4950 scope.go:117] "RemoveContainer" containerID="de454f70da5fb3eaae1d4619d47f7035912417c6dab3afc4c68ffdc8c81f9482" Mar 18 20:54:34 crc kubenswrapper[4950]: E0318 20:54:34.309562 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de454f70da5fb3eaae1d4619d47f7035912417c6dab3afc4c68ffdc8c81f9482\": container with ID starting with de454f70da5fb3eaae1d4619d47f7035912417c6dab3afc4c68ffdc8c81f9482 not found: ID does not exist" containerID="de454f70da5fb3eaae1d4619d47f7035912417c6dab3afc4c68ffdc8c81f9482" Mar 18 20:54:34 crc kubenswrapper[4950]: I0318 20:54:34.309593 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de454f70da5fb3eaae1d4619d47f7035912417c6dab3afc4c68ffdc8c81f9482"} err="failed to get container status \"de454f70da5fb3eaae1d4619d47f7035912417c6dab3afc4c68ffdc8c81f9482\": rpc error: code = NotFound desc = could not find container \"de454f70da5fb3eaae1d4619d47f7035912417c6dab3afc4c68ffdc8c81f9482\": container with ID starting with de454f70da5fb3eaae1d4619d47f7035912417c6dab3afc4c68ffdc8c81f9482 not found: ID does not exist" Mar 18 20:54:34 crc kubenswrapper[4950]: I0318 20:54:34.309615 4950 scope.go:117] "RemoveContainer" containerID="af5be9add612244bb835207ff50e169710fd5b8350c603e07fca7c15a2059467" Mar 18 20:54:34 crc kubenswrapper[4950]: E0318 20:54:34.310101 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af5be9add612244bb835207ff50e169710fd5b8350c603e07fca7c15a2059467\": container with ID starting with af5be9add612244bb835207ff50e169710fd5b8350c603e07fca7c15a2059467 not found: ID does not exist" containerID="af5be9add612244bb835207ff50e169710fd5b8350c603e07fca7c15a2059467" Mar 18 20:54:34 crc kubenswrapper[4950]: I0318 20:54:34.310158 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af5be9add612244bb835207ff50e169710fd5b8350c603e07fca7c15a2059467"} err="failed to get container status \"af5be9add612244bb835207ff50e169710fd5b8350c603e07fca7c15a2059467\": rpc error: code = NotFound desc = could not find container \"af5be9add612244bb835207ff50e169710fd5b8350c603e07fca7c15a2059467\": container with ID starting with af5be9add612244bb835207ff50e169710fd5b8350c603e07fca7c15a2059467 not found: ID does not exist" Mar 18 20:54:35 crc kubenswrapper[4950]: I0318 20:54:35.492600 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81564706-d904-4cca-905f-1169d42c46fc" path="/var/lib/kubelet/pods/81564706-d904-4cca-905f-1169d42c46fc/volumes" Mar 18 20:54:50 crc kubenswrapper[4950]: I0318 20:54:50.490232 4950 scope.go:117] "RemoveContainer" containerID="2aae3f0a0a4ae8919c59beacb63d556d7f9973ea9021d03f673cc338f8e03877" Mar 18 20:55:09 crc kubenswrapper[4950]: I0318 20:55:09.879848 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lb6cj"] Mar 18 20:55:09 crc kubenswrapper[4950]: E0318 20:55:09.881015 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81564706-d904-4cca-905f-1169d42c46fc" containerName="registry-server" Mar 18 20:55:09 crc kubenswrapper[4950]: I0318 20:55:09.881153 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="81564706-d904-4cca-905f-1169d42c46fc" containerName="registry-server" Mar 18 20:55:09 crc kubenswrapper[4950]: E0318 20:55:09.881192 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81564706-d904-4cca-905f-1169d42c46fc" containerName="extract-utilities" Mar 18 20:55:09 crc kubenswrapper[4950]: I0318 20:55:09.881204 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="81564706-d904-4cca-905f-1169d42c46fc" containerName="extract-utilities" Mar 18 20:55:09 crc kubenswrapper[4950]: E0318 20:55:09.881232 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81564706-d904-4cca-905f-1169d42c46fc" containerName="extract-content" Mar 18 20:55:09 crc kubenswrapper[4950]: I0318 20:55:09.881246 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="81564706-d904-4cca-905f-1169d42c46fc" containerName="extract-content" Mar 18 20:55:09 crc kubenswrapper[4950]: I0318 20:55:09.881563 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="81564706-d904-4cca-905f-1169d42c46fc" containerName="registry-server" Mar 18 20:55:09 crc kubenswrapper[4950]: I0318 20:55:09.883731 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lb6cj" Mar 18 20:55:09 crc kubenswrapper[4950]: I0318 20:55:09.900617 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lb6cj"] Mar 18 20:55:09 crc kubenswrapper[4950]: I0318 20:55:09.991214 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm2kj\" (UniqueName: \"kubernetes.io/projected/fecf2897-240f-4ffb-96c5-e3749ef043df-kube-api-access-pm2kj\") pod \"community-operators-lb6cj\" (UID: \"fecf2897-240f-4ffb-96c5-e3749ef043df\") " pod="openshift-marketplace/community-operators-lb6cj" Mar 18 20:55:09 crc kubenswrapper[4950]: I0318 20:55:09.991293 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fecf2897-240f-4ffb-96c5-e3749ef043df-utilities\") pod \"community-operators-lb6cj\" (UID: \"fecf2897-240f-4ffb-96c5-e3749ef043df\") " pod="openshift-marketplace/community-operators-lb6cj" Mar 18 20:55:09 crc kubenswrapper[4950]: I0318 20:55:09.991338 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fecf2897-240f-4ffb-96c5-e3749ef043df-catalog-content\") pod \"community-operators-lb6cj\" (UID: \"fecf2897-240f-4ffb-96c5-e3749ef043df\") " pod="openshift-marketplace/community-operators-lb6cj" Mar 18 20:55:10 crc kubenswrapper[4950]: I0318 20:55:10.093706 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm2kj\" (UniqueName: \"kubernetes.io/projected/fecf2897-240f-4ffb-96c5-e3749ef043df-kube-api-access-pm2kj\") pod \"community-operators-lb6cj\" (UID: \"fecf2897-240f-4ffb-96c5-e3749ef043df\") " pod="openshift-marketplace/community-operators-lb6cj" Mar 18 20:55:10 crc kubenswrapper[4950]: I0318 20:55:10.093789 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fecf2897-240f-4ffb-96c5-e3749ef043df-utilities\") pod \"community-operators-lb6cj\" (UID: \"fecf2897-240f-4ffb-96c5-e3749ef043df\") " pod="openshift-marketplace/community-operators-lb6cj" Mar 18 20:55:10 crc kubenswrapper[4950]: I0318 20:55:10.093834 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fecf2897-240f-4ffb-96c5-e3749ef043df-catalog-content\") pod \"community-operators-lb6cj\" (UID: \"fecf2897-240f-4ffb-96c5-e3749ef043df\") " pod="openshift-marketplace/community-operators-lb6cj" Mar 18 20:55:10 crc kubenswrapper[4950]: I0318 20:55:10.094443 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fecf2897-240f-4ffb-96c5-e3749ef043df-utilities\") pod \"community-operators-lb6cj\" (UID: \"fecf2897-240f-4ffb-96c5-e3749ef043df\") " pod="openshift-marketplace/community-operators-lb6cj" Mar 18 20:55:10 crc kubenswrapper[4950]: I0318 20:55:10.094505 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fecf2897-240f-4ffb-96c5-e3749ef043df-catalog-content\") pod \"community-operators-lb6cj\" (UID: \"fecf2897-240f-4ffb-96c5-e3749ef043df\") " pod="openshift-marketplace/community-operators-lb6cj" Mar 18 20:55:10 crc kubenswrapper[4950]: I0318 20:55:10.120266 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm2kj\" (UniqueName: \"kubernetes.io/projected/fecf2897-240f-4ffb-96c5-e3749ef043df-kube-api-access-pm2kj\") pod \"community-operators-lb6cj\" (UID: \"fecf2897-240f-4ffb-96c5-e3749ef043df\") " pod="openshift-marketplace/community-operators-lb6cj" Mar 18 20:55:10 crc kubenswrapper[4950]: I0318 20:55:10.228138 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lb6cj" Mar 18 20:55:10 crc kubenswrapper[4950]: I0318 20:55:10.538630 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lb6cj"] Mar 18 20:55:10 crc kubenswrapper[4950]: I0318 20:55:10.590644 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lb6cj" event={"ID":"fecf2897-240f-4ffb-96c5-e3749ef043df","Type":"ContainerStarted","Data":"497153e07d461b8a7fb01dff3b12269971cd4436776d9af2ba0c928e03683c72"} Mar 18 20:55:11 crc kubenswrapper[4950]: I0318 20:55:11.604255 4950 generic.go:334] "Generic (PLEG): container finished" podID="fecf2897-240f-4ffb-96c5-e3749ef043df" containerID="a076362a1416f75d957e0ae9847f0f9ed2374d8e371dde876e1487ca57bfd526" exitCode=0 Mar 18 20:55:11 crc kubenswrapper[4950]: I0318 20:55:11.604325 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lb6cj" event={"ID":"fecf2897-240f-4ffb-96c5-e3749ef043df","Type":"ContainerDied","Data":"a076362a1416f75d957e0ae9847f0f9ed2374d8e371dde876e1487ca57bfd526"} Mar 18 20:55:12 crc kubenswrapper[4950]: I0318 20:55:12.613084 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lb6cj" event={"ID":"fecf2897-240f-4ffb-96c5-e3749ef043df","Type":"ContainerStarted","Data":"1bc3f32715fef4f174bd2828503be462e84308917a39fe55aeec225173522514"} Mar 18 20:55:14 crc kubenswrapper[4950]: E0318 20:55:14.173161 4950 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfecf2897_240f_4ffb_96c5_e3749ef043df.slice/crio-1bc3f32715fef4f174bd2828503be462e84308917a39fe55aeec225173522514.scope\": RecentStats: unable to find data in memory cache]" Mar 18 20:55:14 crc kubenswrapper[4950]: I0318 20:55:14.634748 4950 generic.go:334] "Generic (PLEG): container finished" podID="fecf2897-240f-4ffb-96c5-e3749ef043df" containerID="1bc3f32715fef4f174bd2828503be462e84308917a39fe55aeec225173522514" exitCode=0 Mar 18 20:55:14 crc kubenswrapper[4950]: I0318 20:55:14.634796 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lb6cj" event={"ID":"fecf2897-240f-4ffb-96c5-e3749ef043df","Type":"ContainerDied","Data":"1bc3f32715fef4f174bd2828503be462e84308917a39fe55aeec225173522514"} Mar 18 20:55:15 crc kubenswrapper[4950]: I0318 20:55:15.646955 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lb6cj" event={"ID":"fecf2897-240f-4ffb-96c5-e3749ef043df","Type":"ContainerStarted","Data":"169a9e6e790894bce3b7aa987111e9c0fcc0e1b3e1972003906861f3bb7b90cf"} Mar 18 20:55:15 crc kubenswrapper[4950]: I0318 20:55:15.678055 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-698nf"] Mar 18 20:55:15 crc kubenswrapper[4950]: I0318 20:55:15.680269 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-698nf" Mar 18 20:55:15 crc kubenswrapper[4950]: I0318 20:55:15.683525 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lb6cj" podStartSLOduration=3.252616402 podStartE2EDuration="6.683501117s" podCreationTimestamp="2026-03-18 20:55:09 +0000 UTC" firstStartedPulling="2026-03-18 20:55:11.606604977 +0000 UTC m=+2924.847446855" lastFinishedPulling="2026-03-18 20:55:15.037489702 +0000 UTC m=+2928.278331570" observedRunningTime="2026-03-18 20:55:15.672653382 +0000 UTC m=+2928.913495250" watchObservedRunningTime="2026-03-18 20:55:15.683501117 +0000 UTC m=+2928.924342995" Mar 18 20:55:15 crc kubenswrapper[4950]: I0318 20:55:15.726291 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-698nf"] Mar 18 20:55:15 crc kubenswrapper[4950]: I0318 20:55:15.848005 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46a7d322-9043-4313-b014-2dd5096abd03-utilities\") pod \"redhat-marketplace-698nf\" (UID: \"46a7d322-9043-4313-b014-2dd5096abd03\") " pod="openshift-marketplace/redhat-marketplace-698nf" Mar 18 20:55:15 crc kubenswrapper[4950]: I0318 20:55:15.848136 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzq44\" (UniqueName: \"kubernetes.io/projected/46a7d322-9043-4313-b014-2dd5096abd03-kube-api-access-zzq44\") pod \"redhat-marketplace-698nf\" (UID: \"46a7d322-9043-4313-b014-2dd5096abd03\") " pod="openshift-marketplace/redhat-marketplace-698nf" Mar 18 20:55:15 crc kubenswrapper[4950]: I0318 20:55:15.848218 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46a7d322-9043-4313-b014-2dd5096abd03-catalog-content\") pod \"redhat-marketplace-698nf\" (UID: \"46a7d322-9043-4313-b014-2dd5096abd03\") " pod="openshift-marketplace/redhat-marketplace-698nf" Mar 18 20:55:15 crc kubenswrapper[4950]: I0318 20:55:15.949613 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46a7d322-9043-4313-b014-2dd5096abd03-catalog-content\") pod \"redhat-marketplace-698nf\" (UID: \"46a7d322-9043-4313-b014-2dd5096abd03\") " pod="openshift-marketplace/redhat-marketplace-698nf" Mar 18 20:55:15 crc kubenswrapper[4950]: I0318 20:55:15.949743 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46a7d322-9043-4313-b014-2dd5096abd03-utilities\") pod \"redhat-marketplace-698nf\" (UID: \"46a7d322-9043-4313-b014-2dd5096abd03\") " pod="openshift-marketplace/redhat-marketplace-698nf" Mar 18 20:55:15 crc kubenswrapper[4950]: I0318 20:55:15.949784 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzq44\" (UniqueName: \"kubernetes.io/projected/46a7d322-9043-4313-b014-2dd5096abd03-kube-api-access-zzq44\") pod \"redhat-marketplace-698nf\" (UID: \"46a7d322-9043-4313-b014-2dd5096abd03\") " pod="openshift-marketplace/redhat-marketplace-698nf" Mar 18 20:55:15 crc kubenswrapper[4950]: I0318 20:55:15.950216 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46a7d322-9043-4313-b014-2dd5096abd03-catalog-content\") pod \"redhat-marketplace-698nf\" (UID: \"46a7d322-9043-4313-b014-2dd5096abd03\") " pod="openshift-marketplace/redhat-marketplace-698nf" Mar 18 20:55:15 crc kubenswrapper[4950]: I0318 20:55:15.950261 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46a7d322-9043-4313-b014-2dd5096abd03-utilities\") pod \"redhat-marketplace-698nf\" (UID: \"46a7d322-9043-4313-b014-2dd5096abd03\") " pod="openshift-marketplace/redhat-marketplace-698nf" Mar 18 20:55:15 crc kubenswrapper[4950]: I0318 20:55:15.974122 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzq44\" (UniqueName: \"kubernetes.io/projected/46a7d322-9043-4313-b014-2dd5096abd03-kube-api-access-zzq44\") pod \"redhat-marketplace-698nf\" (UID: \"46a7d322-9043-4313-b014-2dd5096abd03\") " pod="openshift-marketplace/redhat-marketplace-698nf" Mar 18 20:55:15 crc kubenswrapper[4950]: I0318 20:55:15.999724 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-698nf" Mar 18 20:55:16 crc kubenswrapper[4950]: I0318 20:55:16.537845 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-698nf"] Mar 18 20:55:16 crc kubenswrapper[4950]: W0318 20:55:16.549554 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46a7d322_9043_4313_b014_2dd5096abd03.slice/crio-97c31777a49b356ec3642a21dc3e4c30ae249f6ab63fa8e335dabc4edd998490 WatchSource:0}: Error finding container 97c31777a49b356ec3642a21dc3e4c30ae249f6ab63fa8e335dabc4edd998490: Status 404 returned error can't find the container with id 97c31777a49b356ec3642a21dc3e4c30ae249f6ab63fa8e335dabc4edd998490 Mar 18 20:55:16 crc kubenswrapper[4950]: I0318 20:55:16.656764 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-698nf" event={"ID":"46a7d322-9043-4313-b014-2dd5096abd03","Type":"ContainerStarted","Data":"97c31777a49b356ec3642a21dc3e4c30ae249f6ab63fa8e335dabc4edd998490"} Mar 18 20:55:17 crc kubenswrapper[4950]: I0318 20:55:17.671897 4950 generic.go:334] "Generic (PLEG): container finished" podID="46a7d322-9043-4313-b014-2dd5096abd03" containerID="6d55d1df48f6ecbb44041c7f0fcdf1471b0e22de1e43be8d60692eee1f8261de" exitCode=0 Mar 18 20:55:17 crc kubenswrapper[4950]: I0318 20:55:17.672157 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-698nf" event={"ID":"46a7d322-9043-4313-b014-2dd5096abd03","Type":"ContainerDied","Data":"6d55d1df48f6ecbb44041c7f0fcdf1471b0e22de1e43be8d60692eee1f8261de"} Mar 18 20:55:18 crc kubenswrapper[4950]: I0318 20:55:18.684891 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-698nf" event={"ID":"46a7d322-9043-4313-b014-2dd5096abd03","Type":"ContainerStarted","Data":"6b6b278a802c85e6b8a7fc37bc6df2c7d0b09240254dd633a76feeda63668014"} Mar 18 20:55:19 crc kubenswrapper[4950]: I0318 20:55:19.709248 4950 generic.go:334] "Generic (PLEG): container finished" podID="46a7d322-9043-4313-b014-2dd5096abd03" containerID="6b6b278a802c85e6b8a7fc37bc6df2c7d0b09240254dd633a76feeda63668014" exitCode=0 Mar 18 20:55:19 crc kubenswrapper[4950]: I0318 20:55:19.709374 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-698nf" event={"ID":"46a7d322-9043-4313-b014-2dd5096abd03","Type":"ContainerDied","Data":"6b6b278a802c85e6b8a7fc37bc6df2c7d0b09240254dd633a76feeda63668014"} Mar 18 20:55:20 crc kubenswrapper[4950]: I0318 20:55:20.229608 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lb6cj" Mar 18 20:55:20 crc kubenswrapper[4950]: I0318 20:55:20.230174 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lb6cj" Mar 18 20:55:20 crc kubenswrapper[4950]: I0318 20:55:20.295891 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lb6cj" Mar 18 20:55:20 crc kubenswrapper[4950]: I0318 20:55:20.719192 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-698nf" event={"ID":"46a7d322-9043-4313-b014-2dd5096abd03","Type":"ContainerStarted","Data":"e15246fb5178b7d8c64f0c9b38e8f214cb273b435057603287381228cbcd7085"} Mar 18 20:55:20 crc kubenswrapper[4950]: I0318 20:55:20.743071 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-698nf" podStartSLOduration=3.283176386 podStartE2EDuration="5.743055512s" podCreationTimestamp="2026-03-18 20:55:15 +0000 UTC" firstStartedPulling="2026-03-18 20:55:17.676477773 +0000 UTC m=+2930.917319641" lastFinishedPulling="2026-03-18 20:55:20.136356899 +0000 UTC m=+2933.377198767" observedRunningTime="2026-03-18 20:55:20.739836827 +0000 UTC m=+2933.980678705" watchObservedRunningTime="2026-03-18 20:55:20.743055512 +0000 UTC m=+2933.983897380" Mar 18 20:55:20 crc kubenswrapper[4950]: I0318 20:55:20.774254 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lb6cj" Mar 18 20:55:22 crc kubenswrapper[4950]: I0318 20:55:22.660766 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lb6cj"] Mar 18 20:55:22 crc kubenswrapper[4950]: I0318 20:55:22.736853 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lb6cj" podUID="fecf2897-240f-4ffb-96c5-e3749ef043df" containerName="registry-server" containerID="cri-o://169a9e6e790894bce3b7aa987111e9c0fcc0e1b3e1972003906861f3bb7b90cf" gracePeriod=2 Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.180443 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lb6cj" Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.299226 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fecf2897-240f-4ffb-96c5-e3749ef043df-utilities\") pod \"fecf2897-240f-4ffb-96c5-e3749ef043df\" (UID: \"fecf2897-240f-4ffb-96c5-e3749ef043df\") " Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.299353 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pm2kj\" (UniqueName: \"kubernetes.io/projected/fecf2897-240f-4ffb-96c5-e3749ef043df-kube-api-access-pm2kj\") pod \"fecf2897-240f-4ffb-96c5-e3749ef043df\" (UID: \"fecf2897-240f-4ffb-96c5-e3749ef043df\") " Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.299422 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fecf2897-240f-4ffb-96c5-e3749ef043df-catalog-content\") pod \"fecf2897-240f-4ffb-96c5-e3749ef043df\" (UID: \"fecf2897-240f-4ffb-96c5-e3749ef043df\") " Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.300213 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fecf2897-240f-4ffb-96c5-e3749ef043df-utilities" (OuterVolumeSpecName: "utilities") pod "fecf2897-240f-4ffb-96c5-e3749ef043df" (UID: "fecf2897-240f-4ffb-96c5-e3749ef043df"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.304632 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fecf2897-240f-4ffb-96c5-e3749ef043df-kube-api-access-pm2kj" (OuterVolumeSpecName: "kube-api-access-pm2kj") pod "fecf2897-240f-4ffb-96c5-e3749ef043df" (UID: "fecf2897-240f-4ffb-96c5-e3749ef043df"). InnerVolumeSpecName "kube-api-access-pm2kj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.367073 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fecf2897-240f-4ffb-96c5-e3749ef043df-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fecf2897-240f-4ffb-96c5-e3749ef043df" (UID: "fecf2897-240f-4ffb-96c5-e3749ef043df"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.401060 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pm2kj\" (UniqueName: \"kubernetes.io/projected/fecf2897-240f-4ffb-96c5-e3749ef043df-kube-api-access-pm2kj\") on node \"crc\" DevicePath \"\"" Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.401095 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fecf2897-240f-4ffb-96c5-e3749ef043df-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.401106 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fecf2897-240f-4ffb-96c5-e3749ef043df-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.746113 4950 generic.go:334] "Generic (PLEG): container finished" podID="fecf2897-240f-4ffb-96c5-e3749ef043df" containerID="169a9e6e790894bce3b7aa987111e9c0fcc0e1b3e1972003906861f3bb7b90cf" exitCode=0 Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.746171 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lb6cj" event={"ID":"fecf2897-240f-4ffb-96c5-e3749ef043df","Type":"ContainerDied","Data":"169a9e6e790894bce3b7aa987111e9c0fcc0e1b3e1972003906861f3bb7b90cf"} Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.746185 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lb6cj" Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.746235 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lb6cj" event={"ID":"fecf2897-240f-4ffb-96c5-e3749ef043df","Type":"ContainerDied","Data":"497153e07d461b8a7fb01dff3b12269971cd4436776d9af2ba0c928e03683c72"} Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.746265 4950 scope.go:117] "RemoveContainer" containerID="169a9e6e790894bce3b7aa987111e9c0fcc0e1b3e1972003906861f3bb7b90cf" Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.768953 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lb6cj"] Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.775804 4950 scope.go:117] "RemoveContainer" containerID="1bc3f32715fef4f174bd2828503be462e84308917a39fe55aeec225173522514" Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.782399 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lb6cj"] Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.802965 4950 scope.go:117] "RemoveContainer" containerID="a076362a1416f75d957e0ae9847f0f9ed2374d8e371dde876e1487ca57bfd526" Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.852283 4950 scope.go:117] "RemoveContainer" containerID="169a9e6e790894bce3b7aa987111e9c0fcc0e1b3e1972003906861f3bb7b90cf" Mar 18 20:55:23 crc kubenswrapper[4950]: E0318 20:55:23.855271 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"169a9e6e790894bce3b7aa987111e9c0fcc0e1b3e1972003906861f3bb7b90cf\": container with ID starting with 169a9e6e790894bce3b7aa987111e9c0fcc0e1b3e1972003906861f3bb7b90cf not found: ID does not exist" containerID="169a9e6e790894bce3b7aa987111e9c0fcc0e1b3e1972003906861f3bb7b90cf" Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.855315 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"169a9e6e790894bce3b7aa987111e9c0fcc0e1b3e1972003906861f3bb7b90cf"} err="failed to get container status \"169a9e6e790894bce3b7aa987111e9c0fcc0e1b3e1972003906861f3bb7b90cf\": rpc error: code = NotFound desc = could not find container \"169a9e6e790894bce3b7aa987111e9c0fcc0e1b3e1972003906861f3bb7b90cf\": container with ID starting with 169a9e6e790894bce3b7aa987111e9c0fcc0e1b3e1972003906861f3bb7b90cf not found: ID does not exist" Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.855342 4950 scope.go:117] "RemoveContainer" containerID="1bc3f32715fef4f174bd2828503be462e84308917a39fe55aeec225173522514" Mar 18 20:55:23 crc kubenswrapper[4950]: E0318 20:55:23.856121 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bc3f32715fef4f174bd2828503be462e84308917a39fe55aeec225173522514\": container with ID starting with 1bc3f32715fef4f174bd2828503be462e84308917a39fe55aeec225173522514 not found: ID does not exist" containerID="1bc3f32715fef4f174bd2828503be462e84308917a39fe55aeec225173522514" Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.856148 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bc3f32715fef4f174bd2828503be462e84308917a39fe55aeec225173522514"} err="failed to get container status \"1bc3f32715fef4f174bd2828503be462e84308917a39fe55aeec225173522514\": rpc error: code = NotFound desc = could not find container \"1bc3f32715fef4f174bd2828503be462e84308917a39fe55aeec225173522514\": container with ID starting with 1bc3f32715fef4f174bd2828503be462e84308917a39fe55aeec225173522514 not found: ID does not exist" Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.856160 4950 scope.go:117] "RemoveContainer" containerID="a076362a1416f75d957e0ae9847f0f9ed2374d8e371dde876e1487ca57bfd526" Mar 18 20:55:23 crc kubenswrapper[4950]: E0318 20:55:23.856382 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a076362a1416f75d957e0ae9847f0f9ed2374d8e371dde876e1487ca57bfd526\": container with ID starting with a076362a1416f75d957e0ae9847f0f9ed2374d8e371dde876e1487ca57bfd526 not found: ID does not exist" containerID="a076362a1416f75d957e0ae9847f0f9ed2374d8e371dde876e1487ca57bfd526" Mar 18 20:55:23 crc kubenswrapper[4950]: I0318 20:55:23.856465 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a076362a1416f75d957e0ae9847f0f9ed2374d8e371dde876e1487ca57bfd526"} err="failed to get container status \"a076362a1416f75d957e0ae9847f0f9ed2374d8e371dde876e1487ca57bfd526\": rpc error: code = NotFound desc = could not find container \"a076362a1416f75d957e0ae9847f0f9ed2374d8e371dde876e1487ca57bfd526\": container with ID starting with a076362a1416f75d957e0ae9847f0f9ed2374d8e371dde876e1487ca57bfd526 not found: ID does not exist" Mar 18 20:55:25 crc kubenswrapper[4950]: I0318 20:55:25.490838 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fecf2897-240f-4ffb-96c5-e3749ef043df" path="/var/lib/kubelet/pods/fecf2897-240f-4ffb-96c5-e3749ef043df/volumes" Mar 18 20:55:26 crc kubenswrapper[4950]: I0318 20:55:26.000393 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-698nf" Mar 18 20:55:26 crc kubenswrapper[4950]: I0318 20:55:26.000639 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-698nf" Mar 18 20:55:26 crc kubenswrapper[4950]: I0318 20:55:26.052789 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-698nf" Mar 18 20:55:26 crc kubenswrapper[4950]: I0318 20:55:26.825882 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-698nf" Mar 18 20:55:27 crc kubenswrapper[4950]: I0318 20:55:27.658692 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-698nf"] Mar 18 20:55:28 crc kubenswrapper[4950]: I0318 20:55:28.793108 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-698nf" podUID="46a7d322-9043-4313-b014-2dd5096abd03" containerName="registry-server" containerID="cri-o://e15246fb5178b7d8c64f0c9b38e8f214cb273b435057603287381228cbcd7085" gracePeriod=2 Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.280795 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-698nf" Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.317974 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46a7d322-9043-4313-b014-2dd5096abd03-utilities\") pod \"46a7d322-9043-4313-b014-2dd5096abd03\" (UID: \"46a7d322-9043-4313-b014-2dd5096abd03\") " Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.318069 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzq44\" (UniqueName: \"kubernetes.io/projected/46a7d322-9043-4313-b014-2dd5096abd03-kube-api-access-zzq44\") pod \"46a7d322-9043-4313-b014-2dd5096abd03\" (UID: \"46a7d322-9043-4313-b014-2dd5096abd03\") " Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.318297 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46a7d322-9043-4313-b014-2dd5096abd03-catalog-content\") pod \"46a7d322-9043-4313-b014-2dd5096abd03\" (UID: \"46a7d322-9043-4313-b014-2dd5096abd03\") " Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.318948 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46a7d322-9043-4313-b014-2dd5096abd03-utilities" (OuterVolumeSpecName: "utilities") pod "46a7d322-9043-4313-b014-2dd5096abd03" (UID: "46a7d322-9043-4313-b014-2dd5096abd03"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.323771 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46a7d322-9043-4313-b014-2dd5096abd03-kube-api-access-zzq44" (OuterVolumeSpecName: "kube-api-access-zzq44") pod "46a7d322-9043-4313-b014-2dd5096abd03" (UID: "46a7d322-9043-4313-b014-2dd5096abd03"). InnerVolumeSpecName "kube-api-access-zzq44". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.346370 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46a7d322-9043-4313-b014-2dd5096abd03-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "46a7d322-9043-4313-b014-2dd5096abd03" (UID: "46a7d322-9043-4313-b014-2dd5096abd03"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.419979 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46a7d322-9043-4313-b014-2dd5096abd03-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.420010 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46a7d322-9043-4313-b014-2dd5096abd03-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.420020 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzq44\" (UniqueName: \"kubernetes.io/projected/46a7d322-9043-4313-b014-2dd5096abd03-kube-api-access-zzq44\") on node \"crc\" DevicePath \"\"" Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.805975 4950 generic.go:334] "Generic (PLEG): container finished" podID="46a7d322-9043-4313-b014-2dd5096abd03" containerID="e15246fb5178b7d8c64f0c9b38e8f214cb273b435057603287381228cbcd7085" exitCode=0 Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.806043 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-698nf" Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.806048 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-698nf" event={"ID":"46a7d322-9043-4313-b014-2dd5096abd03","Type":"ContainerDied","Data":"e15246fb5178b7d8c64f0c9b38e8f214cb273b435057603287381228cbcd7085"} Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.806474 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-698nf" event={"ID":"46a7d322-9043-4313-b014-2dd5096abd03","Type":"ContainerDied","Data":"97c31777a49b356ec3642a21dc3e4c30ae249f6ab63fa8e335dabc4edd998490"} Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.806500 4950 scope.go:117] "RemoveContainer" containerID="e15246fb5178b7d8c64f0c9b38e8f214cb273b435057603287381228cbcd7085" Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.836274 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-698nf"] Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.844245 4950 scope.go:117] "RemoveContainer" containerID="6b6b278a802c85e6b8a7fc37bc6df2c7d0b09240254dd633a76feeda63668014" Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.847173 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-698nf"] Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.869429 4950 scope.go:117] "RemoveContainer" containerID="6d55d1df48f6ecbb44041c7f0fcdf1471b0e22de1e43be8d60692eee1f8261de" Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.898513 4950 scope.go:117] "RemoveContainer" containerID="e15246fb5178b7d8c64f0c9b38e8f214cb273b435057603287381228cbcd7085" Mar 18 20:55:29 crc kubenswrapper[4950]: E0318 20:55:29.899358 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e15246fb5178b7d8c64f0c9b38e8f214cb273b435057603287381228cbcd7085\": container with ID starting with e15246fb5178b7d8c64f0c9b38e8f214cb273b435057603287381228cbcd7085 not found: ID does not exist" containerID="e15246fb5178b7d8c64f0c9b38e8f214cb273b435057603287381228cbcd7085" Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.899398 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e15246fb5178b7d8c64f0c9b38e8f214cb273b435057603287381228cbcd7085"} err="failed to get container status \"e15246fb5178b7d8c64f0c9b38e8f214cb273b435057603287381228cbcd7085\": rpc error: code = NotFound desc = could not find container \"e15246fb5178b7d8c64f0c9b38e8f214cb273b435057603287381228cbcd7085\": container with ID starting with e15246fb5178b7d8c64f0c9b38e8f214cb273b435057603287381228cbcd7085 not found: ID does not exist" Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.899445 4950 scope.go:117] "RemoveContainer" containerID="6b6b278a802c85e6b8a7fc37bc6df2c7d0b09240254dd633a76feeda63668014" Mar 18 20:55:29 crc kubenswrapper[4950]: E0318 20:55:29.899803 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b6b278a802c85e6b8a7fc37bc6df2c7d0b09240254dd633a76feeda63668014\": container with ID starting with 6b6b278a802c85e6b8a7fc37bc6df2c7d0b09240254dd633a76feeda63668014 not found: ID does not exist" containerID="6b6b278a802c85e6b8a7fc37bc6df2c7d0b09240254dd633a76feeda63668014" Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.899842 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b6b278a802c85e6b8a7fc37bc6df2c7d0b09240254dd633a76feeda63668014"} err="failed to get container status \"6b6b278a802c85e6b8a7fc37bc6df2c7d0b09240254dd633a76feeda63668014\": rpc error: code = NotFound desc = could not find container \"6b6b278a802c85e6b8a7fc37bc6df2c7d0b09240254dd633a76feeda63668014\": container with ID starting with 6b6b278a802c85e6b8a7fc37bc6df2c7d0b09240254dd633a76feeda63668014 not found: ID does not exist" Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.899871 4950 scope.go:117] "RemoveContainer" containerID="6d55d1df48f6ecbb44041c7f0fcdf1471b0e22de1e43be8d60692eee1f8261de" Mar 18 20:55:29 crc kubenswrapper[4950]: E0318 20:55:29.900205 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d55d1df48f6ecbb44041c7f0fcdf1471b0e22de1e43be8d60692eee1f8261de\": container with ID starting with 6d55d1df48f6ecbb44041c7f0fcdf1471b0e22de1e43be8d60692eee1f8261de not found: ID does not exist" containerID="6d55d1df48f6ecbb44041c7f0fcdf1471b0e22de1e43be8d60692eee1f8261de" Mar 18 20:55:29 crc kubenswrapper[4950]: I0318 20:55:29.900234 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d55d1df48f6ecbb44041c7f0fcdf1471b0e22de1e43be8d60692eee1f8261de"} err="failed to get container status \"6d55d1df48f6ecbb44041c7f0fcdf1471b0e22de1e43be8d60692eee1f8261de\": rpc error: code = NotFound desc = could not find container \"6d55d1df48f6ecbb44041c7f0fcdf1471b0e22de1e43be8d60692eee1f8261de\": container with ID starting with 6d55d1df48f6ecbb44041c7f0fcdf1471b0e22de1e43be8d60692eee1f8261de not found: ID does not exist" Mar 18 20:55:31 crc kubenswrapper[4950]: I0318 20:55:31.494824 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46a7d322-9043-4313-b014-2dd5096abd03" path="/var/lib/kubelet/pods/46a7d322-9043-4313-b014-2dd5096abd03/volumes" Mar 18 20:55:33 crc kubenswrapper[4950]: I0318 20:55:33.836455 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:55:33 crc kubenswrapper[4950]: I0318 20:55:33.836725 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:56:00 crc kubenswrapper[4950]: I0318 20:56:00.163811 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564456-mpkd8"] Mar 18 20:56:00 crc kubenswrapper[4950]: E0318 20:56:00.165226 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fecf2897-240f-4ffb-96c5-e3749ef043df" containerName="registry-server" Mar 18 20:56:00 crc kubenswrapper[4950]: I0318 20:56:00.165255 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="fecf2897-240f-4ffb-96c5-e3749ef043df" containerName="registry-server" Mar 18 20:56:00 crc kubenswrapper[4950]: E0318 20:56:00.165273 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fecf2897-240f-4ffb-96c5-e3749ef043df" containerName="extract-content" Mar 18 20:56:00 crc kubenswrapper[4950]: I0318 20:56:00.165286 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="fecf2897-240f-4ffb-96c5-e3749ef043df" containerName="extract-content" Mar 18 20:56:00 crc kubenswrapper[4950]: E0318 20:56:00.165310 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46a7d322-9043-4313-b014-2dd5096abd03" containerName="registry-server" Mar 18 20:56:00 crc kubenswrapper[4950]: I0318 20:56:00.165323 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="46a7d322-9043-4313-b014-2dd5096abd03" containerName="registry-server" Mar 18 20:56:00 crc kubenswrapper[4950]: E0318 20:56:00.165360 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fecf2897-240f-4ffb-96c5-e3749ef043df" containerName="extract-utilities" Mar 18 20:56:00 crc kubenswrapper[4950]: I0318 20:56:00.165375 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="fecf2897-240f-4ffb-96c5-e3749ef043df" containerName="extract-utilities" Mar 18 20:56:00 crc kubenswrapper[4950]: E0318 20:56:00.165406 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46a7d322-9043-4313-b014-2dd5096abd03" containerName="extract-content" Mar 18 20:56:00 crc kubenswrapper[4950]: I0318 20:56:00.165552 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="46a7d322-9043-4313-b014-2dd5096abd03" containerName="extract-content" Mar 18 20:56:00 crc kubenswrapper[4950]: E0318 20:56:00.165576 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46a7d322-9043-4313-b014-2dd5096abd03" containerName="extract-utilities" Mar 18 20:56:00 crc kubenswrapper[4950]: I0318 20:56:00.165592 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="46a7d322-9043-4313-b014-2dd5096abd03" containerName="extract-utilities" Mar 18 20:56:00 crc kubenswrapper[4950]: I0318 20:56:00.165995 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="46a7d322-9043-4313-b014-2dd5096abd03" containerName="registry-server" Mar 18 20:56:00 crc kubenswrapper[4950]: I0318 20:56:00.166049 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="fecf2897-240f-4ffb-96c5-e3749ef043df" containerName="registry-server" Mar 18 20:56:00 crc kubenswrapper[4950]: I0318 20:56:00.167260 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564456-mpkd8" Mar 18 20:56:00 crc kubenswrapper[4950]: I0318 20:56:00.171125 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:56:00 crc kubenswrapper[4950]: I0318 20:56:00.171436 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:56:00 crc kubenswrapper[4950]: I0318 20:56:00.171660 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:56:00 crc kubenswrapper[4950]: I0318 20:56:00.174923 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564456-mpkd8"] Mar 18 20:56:00 crc kubenswrapper[4950]: I0318 20:56:00.216352 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtqns\" (UniqueName: \"kubernetes.io/projected/f9960f16-254e-4b51-8c28-5985808c6f84-kube-api-access-qtqns\") pod \"auto-csr-approver-29564456-mpkd8\" (UID: \"f9960f16-254e-4b51-8c28-5985808c6f84\") " pod="openshift-infra/auto-csr-approver-29564456-mpkd8" Mar 18 20:56:00 crc kubenswrapper[4950]: I0318 20:56:00.317359 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtqns\" (UniqueName: \"kubernetes.io/projected/f9960f16-254e-4b51-8c28-5985808c6f84-kube-api-access-qtqns\") pod \"auto-csr-approver-29564456-mpkd8\" (UID: \"f9960f16-254e-4b51-8c28-5985808c6f84\") " pod="openshift-infra/auto-csr-approver-29564456-mpkd8" Mar 18 20:56:00 crc kubenswrapper[4950]: I0318 20:56:00.349296 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtqns\" (UniqueName: \"kubernetes.io/projected/f9960f16-254e-4b51-8c28-5985808c6f84-kube-api-access-qtqns\") pod \"auto-csr-approver-29564456-mpkd8\" (UID: \"f9960f16-254e-4b51-8c28-5985808c6f84\") " pod="openshift-infra/auto-csr-approver-29564456-mpkd8" Mar 18 20:56:00 crc kubenswrapper[4950]: I0318 20:56:00.507569 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564456-mpkd8" Mar 18 20:56:00 crc kubenswrapper[4950]: I0318 20:56:00.962504 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564456-mpkd8"] Mar 18 20:56:01 crc kubenswrapper[4950]: I0318 20:56:01.074297 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564456-mpkd8" event={"ID":"f9960f16-254e-4b51-8c28-5985808c6f84","Type":"ContainerStarted","Data":"a6edf637f09cc1b7d948c32a3db5d6923bf2286de4b7eaec15245eca541d7f63"} Mar 18 20:56:03 crc kubenswrapper[4950]: I0318 20:56:03.092254 4950 generic.go:334] "Generic (PLEG): container finished" podID="f9960f16-254e-4b51-8c28-5985808c6f84" containerID="8d96716591a3c62d75372d015cbe66a76b30ba5103cd9171fdaf45bf213ca81c" exitCode=0 Mar 18 20:56:03 crc kubenswrapper[4950]: I0318 20:56:03.092715 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564456-mpkd8" event={"ID":"f9960f16-254e-4b51-8c28-5985808c6f84","Type":"ContainerDied","Data":"8d96716591a3c62d75372d015cbe66a76b30ba5103cd9171fdaf45bf213ca81c"} Mar 18 20:56:03 crc kubenswrapper[4950]: I0318 20:56:03.836478 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:56:03 crc kubenswrapper[4950]: I0318 20:56:03.836556 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:56:04 crc kubenswrapper[4950]: I0318 20:56:04.437326 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564456-mpkd8" Mar 18 20:56:04 crc kubenswrapper[4950]: I0318 20:56:04.590908 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtqns\" (UniqueName: \"kubernetes.io/projected/f9960f16-254e-4b51-8c28-5985808c6f84-kube-api-access-qtqns\") pod \"f9960f16-254e-4b51-8c28-5985808c6f84\" (UID: \"f9960f16-254e-4b51-8c28-5985808c6f84\") " Mar 18 20:56:04 crc kubenswrapper[4950]: I0318 20:56:04.599552 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9960f16-254e-4b51-8c28-5985808c6f84-kube-api-access-qtqns" (OuterVolumeSpecName: "kube-api-access-qtqns") pod "f9960f16-254e-4b51-8c28-5985808c6f84" (UID: "f9960f16-254e-4b51-8c28-5985808c6f84"). InnerVolumeSpecName "kube-api-access-qtqns". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:56:04 crc kubenswrapper[4950]: I0318 20:56:04.692968 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtqns\" (UniqueName: \"kubernetes.io/projected/f9960f16-254e-4b51-8c28-5985808c6f84-kube-api-access-qtqns\") on node \"crc\" DevicePath \"\"" Mar 18 20:56:05 crc kubenswrapper[4950]: I0318 20:56:05.113077 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564456-mpkd8" event={"ID":"f9960f16-254e-4b51-8c28-5985808c6f84","Type":"ContainerDied","Data":"a6edf637f09cc1b7d948c32a3db5d6923bf2286de4b7eaec15245eca541d7f63"} Mar 18 20:56:05 crc kubenswrapper[4950]: I0318 20:56:05.113688 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6edf637f09cc1b7d948c32a3db5d6923bf2286de4b7eaec15245eca541d7f63" Mar 18 20:56:05 crc kubenswrapper[4950]: I0318 20:56:05.113137 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564456-mpkd8" Mar 18 20:56:05 crc kubenswrapper[4950]: E0318 20:56:05.316442 4950 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9960f16_254e_4b51_8c28_5985808c6f84.slice/crio-a6edf637f09cc1b7d948c32a3db5d6923bf2286de4b7eaec15245eca541d7f63\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9960f16_254e_4b51_8c28_5985808c6f84.slice\": RecentStats: unable to find data in memory cache]" Mar 18 20:56:05 crc kubenswrapper[4950]: I0318 20:56:05.518468 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564450-xwsr8"] Mar 18 20:56:05 crc kubenswrapper[4950]: I0318 20:56:05.536942 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564450-xwsr8"] Mar 18 20:56:07 crc kubenswrapper[4950]: I0318 20:56:07.499526 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d9d3129-446e-40a8-873f-f427bdd96d25" path="/var/lib/kubelet/pods/9d9d3129-446e-40a8-873f-f427bdd96d25/volumes" Mar 18 20:56:19 crc kubenswrapper[4950]: I0318 20:56:19.221960 4950 generic.go:334] "Generic (PLEG): container finished" podID="22d4a06f-ff3c-4fbf-9eae-8be967f1febd" containerID="4c80b144d282d4fdab6d4b5821131aacdba68df40fdc5be615a802fb4ab42abf" exitCode=0 Mar 18 20:56:19 crc kubenswrapper[4950]: I0318 20:56:19.222053 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" event={"ID":"22d4a06f-ff3c-4fbf-9eae-8be967f1febd","Type":"ContainerDied","Data":"4c80b144d282d4fdab6d4b5821131aacdba68df40fdc5be615a802fb4ab42abf"} Mar 18 20:56:20 crc kubenswrapper[4950]: I0318 20:56:20.650186 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" Mar 18 20:56:20 crc kubenswrapper[4950]: I0318 20:56:20.697245 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-ssh-key-openstack-edpm-ipam\") pod \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " Mar 18 20:56:20 crc kubenswrapper[4950]: I0318 20:56:20.697282 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-inventory\") pod \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " Mar 18 20:56:20 crc kubenswrapper[4950]: I0318 20:56:20.697333 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-libvirt-secret-0\") pod \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " Mar 18 20:56:20 crc kubenswrapper[4950]: I0318 20:56:20.697378 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-ceph\") pod \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " Mar 18 20:56:20 crc kubenswrapper[4950]: I0318 20:56:20.697518 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-libvirt-combined-ca-bundle\") pod \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " Mar 18 20:56:20 crc kubenswrapper[4950]: I0318 20:56:20.697547 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkcwc\" (UniqueName: \"kubernetes.io/projected/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-kube-api-access-fkcwc\") pod \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\" (UID: \"22d4a06f-ff3c-4fbf-9eae-8be967f1febd\") " Mar 18 20:56:20 crc kubenswrapper[4950]: I0318 20:56:20.703639 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-ceph" (OuterVolumeSpecName: "ceph") pod "22d4a06f-ff3c-4fbf-9eae-8be967f1febd" (UID: "22d4a06f-ff3c-4fbf-9eae-8be967f1febd"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:56:20 crc kubenswrapper[4950]: I0318 20:56:20.703751 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "22d4a06f-ff3c-4fbf-9eae-8be967f1febd" (UID: "22d4a06f-ff3c-4fbf-9eae-8be967f1febd"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:56:20 crc kubenswrapper[4950]: I0318 20:56:20.704828 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-kube-api-access-fkcwc" (OuterVolumeSpecName: "kube-api-access-fkcwc") pod "22d4a06f-ff3c-4fbf-9eae-8be967f1febd" (UID: "22d4a06f-ff3c-4fbf-9eae-8be967f1febd"). InnerVolumeSpecName "kube-api-access-fkcwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:56:20 crc kubenswrapper[4950]: I0318 20:56:20.724909 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-inventory" (OuterVolumeSpecName: "inventory") pod "22d4a06f-ff3c-4fbf-9eae-8be967f1febd" (UID: "22d4a06f-ff3c-4fbf-9eae-8be967f1febd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:56:20 crc kubenswrapper[4950]: I0318 20:56:20.727734 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "22d4a06f-ff3c-4fbf-9eae-8be967f1febd" (UID: "22d4a06f-ff3c-4fbf-9eae-8be967f1febd"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:56:20 crc kubenswrapper[4950]: I0318 20:56:20.732256 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "22d4a06f-ff3c-4fbf-9eae-8be967f1febd" (UID: "22d4a06f-ff3c-4fbf-9eae-8be967f1febd"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:56:20 crc kubenswrapper[4950]: I0318 20:56:20.799907 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:56:20 crc kubenswrapper[4950]: I0318 20:56:20.799948 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:56:20 crc kubenswrapper[4950]: I0318 20:56:20.799963 4950 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Mar 18 20:56:20 crc kubenswrapper[4950]: I0318 20:56:20.799977 4950 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-ceph\") on node \"crc\" DevicePath \"\"" Mar 18 20:56:20 crc kubenswrapper[4950]: I0318 20:56:20.799990 4950 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:56:20 crc kubenswrapper[4950]: I0318 20:56:20.800005 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkcwc\" (UniqueName: \"kubernetes.io/projected/22d4a06f-ff3c-4fbf-9eae-8be967f1febd-kube-api-access-fkcwc\") on node \"crc\" DevicePath \"\"" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.244253 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" event={"ID":"22d4a06f-ff3c-4fbf-9eae-8be967f1febd","Type":"ContainerDied","Data":"ab99569365a39678fde7fc1ee6c6adba41ecbfec1bd62fa0d0f68c9b33a42121"} Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.244561 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab99569365a39678fde7fc1ee6c6adba41ecbfec1bd62fa0d0f68c9b33a42121" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.244327 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.379990 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx"] Mar 18 20:56:21 crc kubenswrapper[4950]: E0318 20:56:21.380474 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22d4a06f-ff3c-4fbf-9eae-8be967f1febd" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.380500 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="22d4a06f-ff3c-4fbf-9eae-8be967f1febd" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 18 20:56:21 crc kubenswrapper[4950]: E0318 20:56:21.380533 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9960f16-254e-4b51-8c28-5985808c6f84" containerName="oc" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.380542 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9960f16-254e-4b51-8c28-5985808c6f84" containerName="oc" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.380758 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9960f16-254e-4b51-8c28-5985808c6f84" containerName="oc" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.380788 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="22d4a06f-ff3c-4fbf-9eae-8be967f1febd" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.381614 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.387280 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.387589 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.387817 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.387384 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.388021 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.388141 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.388281 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5sn4v" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.387467 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.387514 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.408582 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx"] Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.414925 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.415074 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.415153 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lxrn\" (UniqueName: \"kubernetes.io/projected/006a96ed-5832-49f8-a39c-d433a2b1bfd4-kube-api-access-5lxrn\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.415257 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.415366 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.415503 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-2\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.415616 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.415692 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.415768 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-3\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.415847 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/006a96ed-5832-49f8-a39c-d433a2b1bfd4-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.415948 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.416045 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.416122 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-ssh-key-openstack-edpm-ipam\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.517331 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-3\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.517392 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/006a96ed-5832-49f8-a39c-d433a2b1bfd4-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.517459 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.517500 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.517526 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-ssh-key-openstack-edpm-ipam\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.517569 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.517588 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.517612 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lxrn\" (UniqueName: \"kubernetes.io/projected/006a96ed-5832-49f8-a39c-d433a2b1bfd4-kube-api-access-5lxrn\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.517634 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.517652 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.517670 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-2\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.517765 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.517795 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.519305 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.519591 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/006a96ed-5832-49f8-a39c-d433a2b1bfd4-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.523739 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.524705 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.524820 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-3\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.524893 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.525885 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-2\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.526094 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.528232 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.531010 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.536652 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.537141 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-ssh-key-openstack-edpm-ipam\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.539976 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lxrn\" (UniqueName: \"kubernetes.io/projected/006a96ed-5832-49f8-a39c-d433a2b1bfd4-kube-api-access-5lxrn\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:21 crc kubenswrapper[4950]: I0318 20:56:21.706322 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:56:22 crc kubenswrapper[4950]: I0318 20:56:22.269122 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx"] Mar 18 20:56:23 crc kubenswrapper[4950]: I0318 20:56:23.268200 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" event={"ID":"006a96ed-5832-49f8-a39c-d433a2b1bfd4","Type":"ContainerStarted","Data":"cebe513ebfb6fb5fdb94e3ce792835a1eaab609880b119a271282deebca92290"} Mar 18 20:56:23 crc kubenswrapper[4950]: I0318 20:56:23.268617 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" event={"ID":"006a96ed-5832-49f8-a39c-d433a2b1bfd4","Type":"ContainerStarted","Data":"6e1499e14466076f62f1567b334f334e25cf900c6bca997a3e6c30338143957b"} Mar 18 20:56:23 crc kubenswrapper[4950]: I0318 20:56:23.307426 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" podStartSLOduration=1.8029325539999999 podStartE2EDuration="2.307393061s" podCreationTimestamp="2026-03-18 20:56:21 +0000 UTC" firstStartedPulling="2026-03-18 20:56:22.283816081 +0000 UTC m=+2995.524657949" lastFinishedPulling="2026-03-18 20:56:22.788276588 +0000 UTC m=+2996.029118456" observedRunningTime="2026-03-18 20:56:23.292247123 +0000 UTC m=+2996.533088991" watchObservedRunningTime="2026-03-18 20:56:23.307393061 +0000 UTC m=+2996.548234929" Mar 18 20:56:33 crc kubenswrapper[4950]: I0318 20:56:33.837260 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:56:33 crc kubenswrapper[4950]: I0318 20:56:33.838142 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:56:33 crc kubenswrapper[4950]: I0318 20:56:33.838346 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 20:56:33 crc kubenswrapper[4950]: I0318 20:56:33.839238 4950 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"659b90655ee76a1c728bf80c079e29e36775b2f88753e6ed85575c3829a23583"} pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 20:56:33 crc kubenswrapper[4950]: I0318 20:56:33.839305 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" containerID="cri-o://659b90655ee76a1c728bf80c079e29e36775b2f88753e6ed85575c3829a23583" gracePeriod=600 Mar 18 20:56:34 crc kubenswrapper[4950]: I0318 20:56:34.363645 4950 generic.go:334] "Generic (PLEG): container finished" podID="4048b439-3266-46e7-9de0-22377efacc46" containerID="659b90655ee76a1c728bf80c079e29e36775b2f88753e6ed85575c3829a23583" exitCode=0 Mar 18 20:56:34 crc kubenswrapper[4950]: I0318 20:56:34.363873 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerDied","Data":"659b90655ee76a1c728bf80c079e29e36775b2f88753e6ed85575c3829a23583"} Mar 18 20:56:34 crc kubenswrapper[4950]: I0318 20:56:34.364476 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerStarted","Data":"afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1"} Mar 18 20:56:34 crc kubenswrapper[4950]: I0318 20:56:34.364587 4950 scope.go:117] "RemoveContainer" containerID="75400986b0fb2e7e88fce2d971c71312c8eab4a72027c13e9850e6c30d6fbf18" Mar 18 20:56:50 crc kubenswrapper[4950]: I0318 20:56:50.615398 4950 scope.go:117] "RemoveContainer" containerID="0aa0efd12d9ea9bbe4a547a35f9f99d1107d0539f9ef66bd867fa8cb4f44fc05" Mar 18 20:58:00 crc kubenswrapper[4950]: I0318 20:58:00.154864 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564458-fpv4w"] Mar 18 20:58:00 crc kubenswrapper[4950]: I0318 20:58:00.156466 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564458-fpv4w" Mar 18 20:58:00 crc kubenswrapper[4950]: I0318 20:58:00.160304 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 20:58:00 crc kubenswrapper[4950]: I0318 20:58:00.161338 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 20:58:00 crc kubenswrapper[4950]: I0318 20:58:00.164868 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 20:58:00 crc kubenswrapper[4950]: I0318 20:58:00.174115 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564458-fpv4w"] Mar 18 20:58:00 crc kubenswrapper[4950]: I0318 20:58:00.337298 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppkkf\" (UniqueName: \"kubernetes.io/projected/fdf6ffec-9a44-461d-b30c-d59e5ad41801-kube-api-access-ppkkf\") pod \"auto-csr-approver-29564458-fpv4w\" (UID: \"fdf6ffec-9a44-461d-b30c-d59e5ad41801\") " pod="openshift-infra/auto-csr-approver-29564458-fpv4w" Mar 18 20:58:00 crc kubenswrapper[4950]: I0318 20:58:00.439025 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppkkf\" (UniqueName: \"kubernetes.io/projected/fdf6ffec-9a44-461d-b30c-d59e5ad41801-kube-api-access-ppkkf\") pod \"auto-csr-approver-29564458-fpv4w\" (UID: \"fdf6ffec-9a44-461d-b30c-d59e5ad41801\") " pod="openshift-infra/auto-csr-approver-29564458-fpv4w" Mar 18 20:58:00 crc kubenswrapper[4950]: I0318 20:58:00.460880 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppkkf\" (UniqueName: \"kubernetes.io/projected/fdf6ffec-9a44-461d-b30c-d59e5ad41801-kube-api-access-ppkkf\") pod \"auto-csr-approver-29564458-fpv4w\" (UID: \"fdf6ffec-9a44-461d-b30c-d59e5ad41801\") " pod="openshift-infra/auto-csr-approver-29564458-fpv4w" Mar 18 20:58:00 crc kubenswrapper[4950]: I0318 20:58:00.480116 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564458-fpv4w" Mar 18 20:58:00 crc kubenswrapper[4950]: I0318 20:58:00.944726 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564458-fpv4w"] Mar 18 20:58:01 crc kubenswrapper[4950]: I0318 20:58:01.191303 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564458-fpv4w" event={"ID":"fdf6ffec-9a44-461d-b30c-d59e5ad41801","Type":"ContainerStarted","Data":"164a1b87f0a2b6a9860494227d9ac9489e5675bd58e56b38c6339a61d5ed277e"} Mar 18 20:58:03 crc kubenswrapper[4950]: I0318 20:58:03.209231 4950 generic.go:334] "Generic (PLEG): container finished" podID="fdf6ffec-9a44-461d-b30c-d59e5ad41801" containerID="af2bfc4c6a4a4d1c3d66471c2ce354b5460e43bab98d08892df2a6d3f639eabf" exitCode=0 Mar 18 20:58:03 crc kubenswrapper[4950]: I0318 20:58:03.209719 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564458-fpv4w" event={"ID":"fdf6ffec-9a44-461d-b30c-d59e5ad41801","Type":"ContainerDied","Data":"af2bfc4c6a4a4d1c3d66471c2ce354b5460e43bab98d08892df2a6d3f639eabf"} Mar 18 20:58:04 crc kubenswrapper[4950]: I0318 20:58:04.541979 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564458-fpv4w" Mar 18 20:58:04 crc kubenswrapper[4950]: I0318 20:58:04.626761 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppkkf\" (UniqueName: \"kubernetes.io/projected/fdf6ffec-9a44-461d-b30c-d59e5ad41801-kube-api-access-ppkkf\") pod \"fdf6ffec-9a44-461d-b30c-d59e5ad41801\" (UID: \"fdf6ffec-9a44-461d-b30c-d59e5ad41801\") " Mar 18 20:58:04 crc kubenswrapper[4950]: I0318 20:58:04.637649 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdf6ffec-9a44-461d-b30c-d59e5ad41801-kube-api-access-ppkkf" (OuterVolumeSpecName: "kube-api-access-ppkkf") pod "fdf6ffec-9a44-461d-b30c-d59e5ad41801" (UID: "fdf6ffec-9a44-461d-b30c-d59e5ad41801"). InnerVolumeSpecName "kube-api-access-ppkkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:58:04 crc kubenswrapper[4950]: I0318 20:58:04.728499 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppkkf\" (UniqueName: \"kubernetes.io/projected/fdf6ffec-9a44-461d-b30c-d59e5ad41801-kube-api-access-ppkkf\") on node \"crc\" DevicePath \"\"" Mar 18 20:58:05 crc kubenswrapper[4950]: I0318 20:58:05.224161 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564458-fpv4w" event={"ID":"fdf6ffec-9a44-461d-b30c-d59e5ad41801","Type":"ContainerDied","Data":"164a1b87f0a2b6a9860494227d9ac9489e5675bd58e56b38c6339a61d5ed277e"} Mar 18 20:58:05 crc kubenswrapper[4950]: I0318 20:58:05.224216 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="164a1b87f0a2b6a9860494227d9ac9489e5675bd58e56b38c6339a61d5ed277e" Mar 18 20:58:05 crc kubenswrapper[4950]: I0318 20:58:05.224217 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564458-fpv4w" Mar 18 20:58:05 crc kubenswrapper[4950]: I0318 20:58:05.626832 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564452-f997f"] Mar 18 20:58:05 crc kubenswrapper[4950]: I0318 20:58:05.633887 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564452-f997f"] Mar 18 20:58:07 crc kubenswrapper[4950]: I0318 20:58:07.490866 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d88011d-78a5-45a0-8cc9-939233b4cad4" path="/var/lib/kubelet/pods/3d88011d-78a5-45a0-8cc9-939233b4cad4/volumes" Mar 18 20:58:50 crc kubenswrapper[4950]: I0318 20:58:50.743930 4950 scope.go:117] "RemoveContainer" containerID="33e379d85361962e47f13d952ee22ca3a03d974efd14d056aac0043f264a18f4" Mar 18 20:59:03 crc kubenswrapper[4950]: I0318 20:59:03.836264 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:59:03 crc kubenswrapper[4950]: I0318 20:59:03.836880 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:59:14 crc kubenswrapper[4950]: I0318 20:59:14.829932 4950 generic.go:334] "Generic (PLEG): container finished" podID="006a96ed-5832-49f8-a39c-d433a2b1bfd4" containerID="cebe513ebfb6fb5fdb94e3ce792835a1eaab609880b119a271282deebca92290" exitCode=0 Mar 18 20:59:14 crc kubenswrapper[4950]: I0318 20:59:14.830591 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" event={"ID":"006a96ed-5832-49f8-a39c-d433a2b1bfd4","Type":"ContainerDied","Data":"cebe513ebfb6fb5fdb94e3ce792835a1eaab609880b119a271282deebca92290"} Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.273969 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.334734 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-extra-config-0\") pod \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.334779 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-migration-ssh-key-0\") pod \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.334823 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-migration-ssh-key-1\") pod \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.334854 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lxrn\" (UniqueName: \"kubernetes.io/projected/006a96ed-5832-49f8-a39c-d433a2b1bfd4-kube-api-access-5lxrn\") pod \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.334885 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/006a96ed-5832-49f8-a39c-d433a2b1bfd4-ceph-nova-0\") pod \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.334943 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-ceph\") pod \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.334968 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-inventory\") pod \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.335018 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-ssh-key-openstack-edpm-ipam\") pod \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.335072 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-custom-ceph-combined-ca-bundle\") pod \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.335102 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-2\") pod \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.335126 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-0\") pod \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.335165 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-1\") pod \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.335205 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-3\") pod \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\" (UID: \"006a96ed-5832-49f8-a39c-d433a2b1bfd4\") " Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.356960 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/006a96ed-5832-49f8-a39c-d433a2b1bfd4-kube-api-access-5lxrn" (OuterVolumeSpecName: "kube-api-access-5lxrn") pod "006a96ed-5832-49f8-a39c-d433a2b1bfd4" (UID: "006a96ed-5832-49f8-a39c-d433a2b1bfd4"). InnerVolumeSpecName "kube-api-access-5lxrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.357105 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "006a96ed-5832-49f8-a39c-d433a2b1bfd4" (UID: "006a96ed-5832-49f8-a39c-d433a2b1bfd4"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.360543 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-ceph" (OuterVolumeSpecName: "ceph") pod "006a96ed-5832-49f8-a39c-d433a2b1bfd4" (UID: "006a96ed-5832-49f8-a39c-d433a2b1bfd4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.379063 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-inventory" (OuterVolumeSpecName: "inventory") pod "006a96ed-5832-49f8-a39c-d433a2b1bfd4" (UID: "006a96ed-5832-49f8-a39c-d433a2b1bfd4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.385125 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "006a96ed-5832-49f8-a39c-d433a2b1bfd4" (UID: "006a96ed-5832-49f8-a39c-d433a2b1bfd4"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.393496 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "006a96ed-5832-49f8-a39c-d433a2b1bfd4" (UID: "006a96ed-5832-49f8-a39c-d433a2b1bfd4"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.401425 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "006a96ed-5832-49f8-a39c-d433a2b1bfd4" (UID: "006a96ed-5832-49f8-a39c-d433a2b1bfd4"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.403555 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "006a96ed-5832-49f8-a39c-d433a2b1bfd4" (UID: "006a96ed-5832-49f8-a39c-d433a2b1bfd4"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.404700 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "006a96ed-5832-49f8-a39c-d433a2b1bfd4" (UID: "006a96ed-5832-49f8-a39c-d433a2b1bfd4"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.406137 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/006a96ed-5832-49f8-a39c-d433a2b1bfd4-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "006a96ed-5832-49f8-a39c-d433a2b1bfd4" (UID: "006a96ed-5832-49f8-a39c-d433a2b1bfd4"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.406962 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "006a96ed-5832-49f8-a39c-d433a2b1bfd4" (UID: "006a96ed-5832-49f8-a39c-d433a2b1bfd4"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.422748 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "006a96ed-5832-49f8-a39c-d433a2b1bfd4" (UID: "006a96ed-5832-49f8-a39c-d433a2b1bfd4"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.423059 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "006a96ed-5832-49f8-a39c-d433a2b1bfd4" (UID: "006a96ed-5832-49f8-a39c-d433a2b1bfd4"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.439483 4950 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.439514 4950 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.439526 4950 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.439536 4950 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.439545 4950 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.439554 4950 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.439562 4950 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.439570 4950 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.439580 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lxrn\" (UniqueName: \"kubernetes.io/projected/006a96ed-5832-49f8-a39c-d433a2b1bfd4-kube-api-access-5lxrn\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.439590 4950 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/006a96ed-5832-49f8-a39c-d433a2b1bfd4-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.439599 4950 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-ceph\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.439609 4950 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-inventory\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.439619 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/006a96ed-5832-49f8-a39c-d433a2b1bfd4-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.853528 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" event={"ID":"006a96ed-5832-49f8-a39c-d433a2b1bfd4","Type":"ContainerDied","Data":"6e1499e14466076f62f1567b334f334e25cf900c6bca997a3e6c30338143957b"} Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.853968 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e1499e14466076f62f1567b334f334e25cf900c6bca997a3e6c30338143957b" Mar 18 20:59:16 crc kubenswrapper[4950]: I0318 20:59:16.853561 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.538406 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Mar 18 20:59:31 crc kubenswrapper[4950]: E0318 20:59:31.539229 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="006a96ed-5832-49f8-a39c-d433a2b1bfd4" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.539244 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="006a96ed-5832-49f8-a39c-d433a2b1bfd4" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Mar 18 20:59:31 crc kubenswrapper[4950]: E0318 20:59:31.539266 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdf6ffec-9a44-461d-b30c-d59e5ad41801" containerName="oc" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.539273 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdf6ffec-9a44-461d-b30c-d59e5ad41801" containerName="oc" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.539444 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="006a96ed-5832-49f8-a39c-d433a2b1bfd4" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.539464 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdf6ffec-9a44-461d-b30c-d59e5ad41801" containerName="oc" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.540383 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.552877 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.553092 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.555241 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.556712 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.569303 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.576285 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.606292 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.710958 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.711002 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-run\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.711028 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-sys\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.711043 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.711787 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.711903 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-dev\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.711978 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.712072 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.712165 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9afecfad-52c5-439c-a3bc-a56640e9a79a-config-data-custom\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.712234 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.712335 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9afecfad-52c5-439c-a3bc-a56640e9a79a-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.712422 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.712625 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-lib-modules\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.712713 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.712820 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9afecfad-52c5-439c-a3bc-a56640e9a79a-ceph\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.712902 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.712974 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-run\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.713044 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-sys\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.713131 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.713225 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.713298 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.713433 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.713542 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npcp9\" (UniqueName: \"kubernetes.io/projected/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-kube-api-access-npcp9\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.713645 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.713713 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-dev\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.713801 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.713925 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.714004 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6svm\" (UniqueName: \"kubernetes.io/projected/9afecfad-52c5-439c-a3bc-a56640e9a79a-kube-api-access-w6svm\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.714082 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.714158 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9afecfad-52c5-439c-a3bc-a56640e9a79a-config-data\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.714243 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9afecfad-52c5-439c-a3bc-a56640e9a79a-scripts\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.714321 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-etc-nvme\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.815763 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-sys\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.815800 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.815829 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.815850 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-dev\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.815870 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.815893 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.815916 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9afecfad-52c5-439c-a3bc-a56640e9a79a-config-data-custom\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.815932 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.815953 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9afecfad-52c5-439c-a3bc-a56640e9a79a-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.815967 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.815990 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-lib-modules\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816007 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816027 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9afecfad-52c5-439c-a3bc-a56640e9a79a-ceph\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816042 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816057 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-run\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816073 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-sys\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816094 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816112 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816125 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816152 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816177 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npcp9\" (UniqueName: \"kubernetes.io/projected/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-kube-api-access-npcp9\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816196 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816211 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-dev\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816231 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816269 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816285 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6svm\" (UniqueName: \"kubernetes.io/projected/9afecfad-52c5-439c-a3bc-a56640e9a79a-kube-api-access-w6svm\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816304 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816319 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9afecfad-52c5-439c-a3bc-a56640e9a79a-config-data\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816338 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9afecfad-52c5-439c-a3bc-a56640e9a79a-scripts\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816359 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-etc-nvme\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816385 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816400 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-run\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816498 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-run\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816533 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816554 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.816560 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-sys\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.817340 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-dev\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.817584 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.817732 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-dev\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.817739 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.817763 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.817782 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-run\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.817978 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.817959 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.817815 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.818031 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-lib-modules\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.818217 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.821524 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.821567 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-sys\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.825777 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.825835 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.826764 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9afecfad-52c5-439c-a3bc-a56640e9a79a-scripts\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.826838 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9afecfad-52c5-439c-a3bc-a56640e9a79a-etc-nvme\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.826869 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.832056 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.836148 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9afecfad-52c5-439c-a3bc-a56640e9a79a-ceph\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.836751 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9afecfad-52c5-439c-a3bc-a56640e9a79a-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.837223 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9afecfad-52c5-439c-a3bc-a56640e9a79a-config-data-custom\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.837334 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.837854 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.838118 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.838231 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9afecfad-52c5-439c-a3bc-a56640e9a79a-config-data\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.840332 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6svm\" (UniqueName: \"kubernetes.io/projected/9afecfad-52c5-439c-a3bc-a56640e9a79a-kube-api-access-w6svm\") pod \"cinder-backup-0\" (UID: \"9afecfad-52c5-439c-a3bc-a56640e9a79a\") " pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.842371 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npcp9\" (UniqueName: \"kubernetes.io/projected/e22cc0cd-ebf4-4f03-91f3-1db5f2eac881-kube-api-access-npcp9\") pod \"cinder-volume-volume1-0\" (UID: \"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881\") " pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.863709 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Mar 18 20:59:31 crc kubenswrapper[4950]: I0318 20:59:31.885966 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.316068 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-4qv4l"] Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.317314 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-4qv4l" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.323893 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e05258ae-28d6-4c4c-8bc0-c20e6f112198-operator-scripts\") pod \"manila-db-create-4qv4l\" (UID: \"e05258ae-28d6-4c4c-8bc0-c20e6f112198\") " pod="openstack/manila-db-create-4qv4l" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.324031 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvkp5\" (UniqueName: \"kubernetes.io/projected/e05258ae-28d6-4c4c-8bc0-c20e6f112198-kube-api-access-pvkp5\") pod \"manila-db-create-4qv4l\" (UID: \"e05258ae-28d6-4c4c-8bc0-c20e6f112198\") " pod="openstack/manila-db-create-4qv4l" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.337555 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-4qv4l"] Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.423202 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-ca5b-account-create-update-8l5r2"] Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.424661 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-ca5b-account-create-update-8l5r2" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.428209 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e05258ae-28d6-4c4c-8bc0-c20e6f112198-operator-scripts\") pod \"manila-db-create-4qv4l\" (UID: \"e05258ae-28d6-4c4c-8bc0-c20e6f112198\") " pod="openstack/manila-db-create-4qv4l" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.428333 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvkp5\" (UniqueName: \"kubernetes.io/projected/e05258ae-28d6-4c4c-8bc0-c20e6f112198-kube-api-access-pvkp5\") pod \"manila-db-create-4qv4l\" (UID: \"e05258ae-28d6-4c4c-8bc0-c20e6f112198\") " pod="openstack/manila-db-create-4qv4l" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.430159 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e05258ae-28d6-4c4c-8bc0-c20e6f112198-operator-scripts\") pod \"manila-db-create-4qv4l\" (UID: \"e05258ae-28d6-4c4c-8bc0-c20e6f112198\") " pod="openstack/manila-db-create-4qv4l" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.430388 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.438168 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.449621 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.457324 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.457488 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.457644 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.457679 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2vwp6" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.484865 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-ca5b-account-create-update-8l5r2"] Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.491673 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvkp5\" (UniqueName: \"kubernetes.io/projected/e05258ae-28d6-4c4c-8bc0-c20e6f112198-kube-api-access-pvkp5\") pod \"manila-db-create-4qv4l\" (UID: \"e05258ae-28d6-4c4c-8bc0-c20e6f112198\") " pod="openstack/manila-db-create-4qv4l" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.563601 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.566946 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppvkv\" (UniqueName: \"kubernetes.io/projected/4bf02c5a-528e-43e3-8f9d-faa5619091e8-kube-api-access-ppvkv\") pod \"manila-ca5b-account-create-update-8l5r2\" (UID: \"4bf02c5a-528e-43e3-8f9d-faa5619091e8\") " pod="openstack/manila-ca5b-account-create-update-8l5r2" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.567057 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bf02c5a-528e-43e3-8f9d-faa5619091e8-operator-scripts\") pod \"manila-ca5b-account-create-update-8l5r2\" (UID: \"4bf02c5a-528e-43e3-8f9d-faa5619091e8\") " pod="openstack/manila-ca5b-account-create-update-8l5r2" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.645717 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-659bf66fb7-qw9s8"] Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.647646 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-659bf66fb7-qw9s8" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.651072 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-4qv4l" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.662441 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-66tbp" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.663295 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.667784 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.667785 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.669493 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.669599 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.669648 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bf02c5a-528e-43e3-8f9d-faa5619091e8-operator-scripts\") pod \"manila-ca5b-account-create-update-8l5r2\" (UID: \"4bf02c5a-528e-43e3-8f9d-faa5619091e8\") " pod="openstack/manila-ca5b-account-create-update-8l5r2" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.669685 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.669734 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/42d532dd-095d-47c2-8a85-806c38a6ed36-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.669775 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42d532dd-095d-47c2-8a85-806c38a6ed36-logs\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.669823 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tptll\" (UniqueName: \"kubernetes.io/projected/42d532dd-095d-47c2-8a85-806c38a6ed36-kube-api-access-tptll\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.669890 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/42d532dd-095d-47c2-8a85-806c38a6ed36-ceph\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.669921 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-config-data\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.669965 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-scripts\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.669994 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppvkv\" (UniqueName: \"kubernetes.io/projected/4bf02c5a-528e-43e3-8f9d-faa5619091e8-kube-api-access-ppvkv\") pod \"manila-ca5b-account-create-update-8l5r2\" (UID: \"4bf02c5a-528e-43e3-8f9d-faa5619091e8\") " pod="openstack/manila-ca5b-account-create-update-8l5r2" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.671507 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-659bf66fb7-qw9s8"] Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.671624 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bf02c5a-528e-43e3-8f9d-faa5619091e8-operator-scripts\") pod \"manila-ca5b-account-create-update-8l5r2\" (UID: \"4bf02c5a-528e-43e3-8f9d-faa5619091e8\") " pod="openstack/manila-ca5b-account-create-update-8l5r2" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.707683 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppvkv\" (UniqueName: \"kubernetes.io/projected/4bf02c5a-528e-43e3-8f9d-faa5619091e8-kube-api-access-ppvkv\") pod \"manila-ca5b-account-create-update-8l5r2\" (UID: \"4bf02c5a-528e-43e3-8f9d-faa5619091e8\") " pod="openstack/manila-ca5b-account-create-update-8l5r2" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.711507 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.713233 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.719308 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.719511 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.756544 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.772558 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/42d532dd-095d-47c2-8a85-806c38a6ed36-ceph\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.772884 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-config-data\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.772912 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/746eaa13-6098-490f-9ada-f8338fe419be-horizon-secret-key\") pod \"horizon-659bf66fb7-qw9s8\" (UID: \"746eaa13-6098-490f-9ada-f8338fe419be\") " pod="openstack/horizon-659bf66fb7-qw9s8" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.772939 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-scripts\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.772958 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgzf5\" (UniqueName: \"kubernetes.io/projected/746eaa13-6098-490f-9ada-f8338fe419be-kube-api-access-sgzf5\") pod \"horizon-659bf66fb7-qw9s8\" (UID: \"746eaa13-6098-490f-9ada-f8338fe419be\") " pod="openstack/horizon-659bf66fb7-qw9s8" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.772983 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/746eaa13-6098-490f-9ada-f8338fe419be-config-data\") pod \"horizon-659bf66fb7-qw9s8\" (UID: \"746eaa13-6098-490f-9ada-f8338fe419be\") " pod="openstack/horizon-659bf66fb7-qw9s8" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.773007 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/746eaa13-6098-490f-9ada-f8338fe419be-scripts\") pod \"horizon-659bf66fb7-qw9s8\" (UID: \"746eaa13-6098-490f-9ada-f8338fe419be\") " pod="openstack/horizon-659bf66fb7-qw9s8" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.773033 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.773075 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.773108 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.773133 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/42d532dd-095d-47c2-8a85-806c38a6ed36-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.773164 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42d532dd-095d-47c2-8a85-806c38a6ed36-logs\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.773183 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/746eaa13-6098-490f-9ada-f8338fe419be-logs\") pod \"horizon-659bf66fb7-qw9s8\" (UID: \"746eaa13-6098-490f-9ada-f8338fe419be\") " pod="openstack/horizon-659bf66fb7-qw9s8" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.773199 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tptll\" (UniqueName: \"kubernetes.io/projected/42d532dd-095d-47c2-8a85-806c38a6ed36-kube-api-access-tptll\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.779730 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42d532dd-095d-47c2-8a85-806c38a6ed36-logs\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.779957 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/42d532dd-095d-47c2-8a85-806c38a6ed36-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.781712 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/42d532dd-095d-47c2-8a85-806c38a6ed36-ceph\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.781969 4950 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.783387 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.785986 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-scripts\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.795658 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.802476 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 18 20:59:32 crc kubenswrapper[4950]: E0318 20:59:32.803234 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config-data glance kube-api-access-tptll], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/glance-default-external-api-0" podUID="42d532dd-095d-47c2-8a85-806c38a6ed36" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.821838 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-ca5b-account-create-update-8l5r2" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.823604 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-config-data\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.847015 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tptll\" (UniqueName: \"kubernetes.io/projected/42d532dd-095d-47c2-8a85-806c38a6ed36-kube-api-access-tptll\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.854463 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 18 20:59:32 crc kubenswrapper[4950]: E0318 20:59:32.855200 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceph combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-nnlng logs scripts], unattached volumes=[], failed to process volumes=[ceph combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-nnlng logs scripts]: context canceled" pod="openstack/glance-default-internal-api-0" podUID="1f04bf5f-c70b-48bb-8eba-78f18fd9b60c" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.879302 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/746eaa13-6098-490f-9ada-f8338fe419be-logs\") pod \"horizon-659bf66fb7-qw9s8\" (UID: \"746eaa13-6098-490f-9ada-f8338fe419be\") " pod="openstack/horizon-659bf66fb7-qw9s8" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.879375 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnlng\" (UniqueName: \"kubernetes.io/projected/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-kube-api-access-nnlng\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.879450 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/746eaa13-6098-490f-9ada-f8338fe419be-horizon-secret-key\") pod \"horizon-659bf66fb7-qw9s8\" (UID: \"746eaa13-6098-490f-9ada-f8338fe419be\") " pod="openstack/horizon-659bf66fb7-qw9s8" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.879480 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgzf5\" (UniqueName: \"kubernetes.io/projected/746eaa13-6098-490f-9ada-f8338fe419be-kube-api-access-sgzf5\") pod \"horizon-659bf66fb7-qw9s8\" (UID: \"746eaa13-6098-490f-9ada-f8338fe419be\") " pod="openstack/horizon-659bf66fb7-qw9s8" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.879497 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-ceph\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.879518 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/746eaa13-6098-490f-9ada-f8338fe419be-config-data\") pod \"horizon-659bf66fb7-qw9s8\" (UID: \"746eaa13-6098-490f-9ada-f8338fe419be\") " pod="openstack/horizon-659bf66fb7-qw9s8" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.879538 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/746eaa13-6098-490f-9ada-f8338fe419be-scripts\") pod \"horizon-659bf66fb7-qw9s8\" (UID: \"746eaa13-6098-490f-9ada-f8338fe419be\") " pod="openstack/horizon-659bf66fb7-qw9s8" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.879562 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.879593 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.879609 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.879626 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.879652 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.879678 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-logs\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.879738 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.880235 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/746eaa13-6098-490f-9ada-f8338fe419be-logs\") pod \"horizon-659bf66fb7-qw9s8\" (UID: \"746eaa13-6098-490f-9ada-f8338fe419be\") " pod="openstack/horizon-659bf66fb7-qw9s8" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.882458 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/746eaa13-6098-490f-9ada-f8338fe419be-config-data\") pod \"horizon-659bf66fb7-qw9s8\" (UID: \"746eaa13-6098-490f-9ada-f8338fe419be\") " pod="openstack/horizon-659bf66fb7-qw9s8" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.882494 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/746eaa13-6098-490f-9ada-f8338fe419be-scripts\") pod \"horizon-659bf66fb7-qw9s8\" (UID: \"746eaa13-6098-490f-9ada-f8338fe419be\") " pod="openstack/horizon-659bf66fb7-qw9s8" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.888810 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.900391 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/746eaa13-6098-490f-9ada-f8338fe419be-horizon-secret-key\") pod \"horizon-659bf66fb7-qw9s8\" (UID: \"746eaa13-6098-490f-9ada-f8338fe419be\") " pod="openstack/horizon-659bf66fb7-qw9s8" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.910071 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-848988ddb5-9t8cx"] Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.911774 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-848988ddb5-9t8cx" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.923809 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgzf5\" (UniqueName: \"kubernetes.io/projected/746eaa13-6098-490f-9ada-f8338fe419be-kube-api-access-sgzf5\") pod \"horizon-659bf66fb7-qw9s8\" (UID: \"746eaa13-6098-490f-9ada-f8338fe419be\") " pod="openstack/horizon-659bf66fb7-qw9s8" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.926962 4950 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.979737 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-848988ddb5-9t8cx"] Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.982909 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-scripts\") pod \"horizon-848988ddb5-9t8cx\" (UID: \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\") " pod="openstack/horizon-848988ddb5-9t8cx" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.982946 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-ceph\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.982986 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.983003 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-config-data\") pod \"horizon-848988ddb5-9t8cx\" (UID: \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\") " pod="openstack/horizon-848988ddb5-9t8cx" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.983031 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.983048 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.983063 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.983095 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.983111 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-logs\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.983145 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.983164 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-logs\") pod \"horizon-848988ddb5-9t8cx\" (UID: \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\") " pod="openstack/horizon-848988ddb5-9t8cx" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.983197 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7tx5\" (UniqueName: \"kubernetes.io/projected/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-kube-api-access-r7tx5\") pod \"horizon-848988ddb5-9t8cx\" (UID: \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\") " pod="openstack/horizon-848988ddb5-9t8cx" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.983234 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnlng\" (UniqueName: \"kubernetes.io/projected/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-kube-api-access-nnlng\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.984162 4950 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.987724 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-logs\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.988695 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.992050 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-horizon-secret-key\") pod \"horizon-848988ddb5-9t8cx\" (UID: \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\") " pod="openstack/horizon-848988ddb5-9t8cx" Mar 18 20:59:32 crc kubenswrapper[4950]: I0318 20:59:32.994846 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-659bf66fb7-qw9s8" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.005646 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.009162 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.030125 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.056517 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.057451 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-ceph\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.070188 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnlng\" (UniqueName: \"kubernetes.io/projected/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-kube-api-access-nnlng\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.080813 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.081391 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"9afecfad-52c5-439c-a3bc-a56640e9a79a","Type":"ContainerStarted","Data":"df51d92c91eeef89f89c99afa2d20c6ce05af0df1a2c5df7a109b1bbb5172be6"} Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.081456 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.088633 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.094224 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-horizon-secret-key\") pod \"horizon-848988ddb5-9t8cx\" (UID: \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\") " pod="openstack/horizon-848988ddb5-9t8cx" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.094286 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-scripts\") pod \"horizon-848988ddb5-9t8cx\" (UID: \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\") " pod="openstack/horizon-848988ddb5-9t8cx" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.094334 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-config-data\") pod \"horizon-848988ddb5-9t8cx\" (UID: \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\") " pod="openstack/horizon-848988ddb5-9t8cx" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.094406 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-logs\") pod \"horizon-848988ddb5-9t8cx\" (UID: \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\") " pod="openstack/horizon-848988ddb5-9t8cx" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.094456 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7tx5\" (UniqueName: \"kubernetes.io/projected/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-kube-api-access-r7tx5\") pod \"horizon-848988ddb5-9t8cx\" (UID: \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\") " pod="openstack/horizon-848988ddb5-9t8cx" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.102113 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-horizon-secret-key\") pod \"horizon-848988ddb5-9t8cx\" (UID: \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\") " pod="openstack/horizon-848988ddb5-9t8cx" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.102519 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-logs\") pod \"horizon-848988ddb5-9t8cx\" (UID: \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\") " pod="openstack/horizon-848988ddb5-9t8cx" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.103052 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-scripts\") pod \"horizon-848988ddb5-9t8cx\" (UID: \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\") " pod="openstack/horizon-848988ddb5-9t8cx" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.104232 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-config-data\") pod \"horizon-848988ddb5-9t8cx\" (UID: \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\") " pod="openstack/horizon-848988ddb5-9t8cx" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.105439 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.121491 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.129147 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7tx5\" (UniqueName: \"kubernetes.io/projected/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-kube-api-access-r7tx5\") pod \"horizon-848988ddb5-9t8cx\" (UID: \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\") " pod="openstack/horizon-848988ddb5-9t8cx" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.131437 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-848988ddb5-9t8cx" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.150302 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.167198 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.196354 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"42d532dd-095d-47c2-8a85-806c38a6ed36\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.196504 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-httpd-run\") pod \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.196607 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42d532dd-095d-47c2-8a85-806c38a6ed36-logs\") pod \"42d532dd-095d-47c2-8a85-806c38a6ed36\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.196685 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.196749 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-combined-ca-bundle\") pod \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.196829 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-config-data\") pod \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.196904 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-logs\") pod \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.196992 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnlng\" (UniqueName: \"kubernetes.io/projected/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-kube-api-access-nnlng\") pod \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.197056 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-ceph\") pod \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.197117 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tptll\" (UniqueName: \"kubernetes.io/projected/42d532dd-095d-47c2-8a85-806c38a6ed36-kube-api-access-tptll\") pod \"42d532dd-095d-47c2-8a85-806c38a6ed36\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.197193 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-config-data\") pod \"42d532dd-095d-47c2-8a85-806c38a6ed36\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.197268 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-scripts\") pod \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.197341 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/42d532dd-095d-47c2-8a85-806c38a6ed36-ceph\") pod \"42d532dd-095d-47c2-8a85-806c38a6ed36\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.197410 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-combined-ca-bundle\") pod \"42d532dd-095d-47c2-8a85-806c38a6ed36\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.197505 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-internal-tls-certs\") pod \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\" (UID: \"1f04bf5f-c70b-48bb-8eba-78f18fd9b60c\") " Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.197600 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-scripts\") pod \"42d532dd-095d-47c2-8a85-806c38a6ed36\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.197682 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/42d532dd-095d-47c2-8a85-806c38a6ed36-httpd-run\") pod \"42d532dd-095d-47c2-8a85-806c38a6ed36\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.197774 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-public-tls-certs\") pod \"42d532dd-095d-47c2-8a85-806c38a6ed36\" (UID: \"42d532dd-095d-47c2-8a85-806c38a6ed36\") " Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.204232 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-logs" (OuterVolumeSpecName: "logs") pod "1f04bf5f-c70b-48bb-8eba-78f18fd9b60c" (UID: "1f04bf5f-c70b-48bb-8eba-78f18fd9b60c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.204363 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-config-data" (OuterVolumeSpecName: "config-data") pod "1f04bf5f-c70b-48bb-8eba-78f18fd9b60c" (UID: "1f04bf5f-c70b-48bb-8eba-78f18fd9b60c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.204891 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1f04bf5f-c70b-48bb-8eba-78f18fd9b60c" (UID: "1f04bf5f-c70b-48bb-8eba-78f18fd9b60c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.205093 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42d532dd-095d-47c2-8a85-806c38a6ed36-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "42d532dd-095d-47c2-8a85-806c38a6ed36" (UID: "42d532dd-095d-47c2-8a85-806c38a6ed36"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.205523 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42d532dd-095d-47c2-8a85-806c38a6ed36-logs" (OuterVolumeSpecName: "logs") pod "42d532dd-095d-47c2-8a85-806c38a6ed36" (UID: "42d532dd-095d-47c2-8a85-806c38a6ed36"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.208871 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42d532dd-095d-47c2-8a85-806c38a6ed36" (UID: "42d532dd-095d-47c2-8a85-806c38a6ed36"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.213996 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1f04bf5f-c70b-48bb-8eba-78f18fd9b60c" (UID: "1f04bf5f-c70b-48bb-8eba-78f18fd9b60c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.214105 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "1f04bf5f-c70b-48bb-8eba-78f18fd9b60c" (UID: "1f04bf5f-c70b-48bb-8eba-78f18fd9b60c"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.214112 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1f04bf5f-c70b-48bb-8eba-78f18fd9b60c" (UID: "1f04bf5f-c70b-48bb-8eba-78f18fd9b60c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.214166 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-config-data" (OuterVolumeSpecName: "config-data") pod "42d532dd-095d-47c2-8a85-806c38a6ed36" (UID: "42d532dd-095d-47c2-8a85-806c38a6ed36"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.214128 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "42d532dd-095d-47c2-8a85-806c38a6ed36" (UID: "42d532dd-095d-47c2-8a85-806c38a6ed36"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.214308 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-ceph" (OuterVolumeSpecName: "ceph") pod "1f04bf5f-c70b-48bb-8eba-78f18fd9b60c" (UID: "1f04bf5f-c70b-48bb-8eba-78f18fd9b60c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.215083 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "42d532dd-095d-47c2-8a85-806c38a6ed36" (UID: "42d532dd-095d-47c2-8a85-806c38a6ed36"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.216816 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-scripts" (OuterVolumeSpecName: "scripts") pod "1f04bf5f-c70b-48bb-8eba-78f18fd9b60c" (UID: "1f04bf5f-c70b-48bb-8eba-78f18fd9b60c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.216939 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-kube-api-access-nnlng" (OuterVolumeSpecName: "kube-api-access-nnlng") pod "1f04bf5f-c70b-48bb-8eba-78f18fd9b60c" (UID: "1f04bf5f-c70b-48bb-8eba-78f18fd9b60c"). InnerVolumeSpecName "kube-api-access-nnlng". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.217468 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-scripts" (OuterVolumeSpecName: "scripts") pod "42d532dd-095d-47c2-8a85-806c38a6ed36" (UID: "42d532dd-095d-47c2-8a85-806c38a6ed36"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.217632 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42d532dd-095d-47c2-8a85-806c38a6ed36-kube-api-access-tptll" (OuterVolumeSpecName: "kube-api-access-tptll") pod "42d532dd-095d-47c2-8a85-806c38a6ed36" (UID: "42d532dd-095d-47c2-8a85-806c38a6ed36"). InnerVolumeSpecName "kube-api-access-tptll". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.219245 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42d532dd-095d-47c2-8a85-806c38a6ed36-ceph" (OuterVolumeSpecName: "ceph") pod "42d532dd-095d-47c2-8a85-806c38a6ed36" (UID: "42d532dd-095d-47c2-8a85-806c38a6ed36"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.300400 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.301200 4950 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/42d532dd-095d-47c2-8a85-806c38a6ed36-ceph\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.301287 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.301361 4950 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.301435 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.301489 4950 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/42d532dd-095d-47c2-8a85-806c38a6ed36-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.301540 4950 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.301633 4950 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.301743 4950 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.301803 4950 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42d532dd-095d-47c2-8a85-806c38a6ed36-logs\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.301905 4950 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.301988 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.302043 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.302099 4950 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-logs\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.302155 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnlng\" (UniqueName: \"kubernetes.io/projected/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-kube-api-access-nnlng\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.302302 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tptll\" (UniqueName: \"kubernetes.io/projected/42d532dd-095d-47c2-8a85-806c38a6ed36-kube-api-access-tptll\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.303095 4950 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c-ceph\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.303169 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42d532dd-095d-47c2-8a85-806c38a6ed36-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.341941 4950 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.384246 4950 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.406028 4950 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.406056 4950 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.461784 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-4qv4l"] Mar 18 20:59:33 crc kubenswrapper[4950]: W0318 20:59:33.472885 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode05258ae_28d6_4c4c_8bc0_c20e6f112198.slice/crio-9eeb864b7c02bf07e7df6c7ae1f89d87798d453ef4998a366c8e2ba2fb9615ea WatchSource:0}: Error finding container 9eeb864b7c02bf07e7df6c7ae1f89d87798d453ef4998a366c8e2ba2fb9615ea: Status 404 returned error can't find the container with id 9eeb864b7c02bf07e7df6c7ae1f89d87798d453ef4998a366c8e2ba2fb9615ea Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.647022 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-ca5b-account-create-update-8l5r2"] Mar 18 20:59:33 crc kubenswrapper[4950]: W0318 20:59:33.703742 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4bf02c5a_528e_43e3_8f9d_faa5619091e8.slice/crio-c40d0a03711f375d680ef1c42f2a1968feb4fd34ed7c23e0cb903f0c6d3b31a6 WatchSource:0}: Error finding container c40d0a03711f375d680ef1c42f2a1968feb4fd34ed7c23e0cb903f0c6d3b31a6: Status 404 returned error can't find the container with id c40d0a03711f375d680ef1c42f2a1968feb4fd34ed7c23e0cb903f0c6d3b31a6 Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.797198 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-848988ddb5-9t8cx"] Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.815841 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-659bf66fb7-qw9s8"] Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.843144 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 20:59:33 crc kubenswrapper[4950]: I0318 20:59:33.843201 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.100602 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-ca5b-account-create-update-8l5r2" event={"ID":"4bf02c5a-528e-43e3-8f9d-faa5619091e8","Type":"ContainerStarted","Data":"93b98e9080b2a203c6474765a7e3afd36f0f3a668d12adcbcd4e48111df5e0b4"} Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.100939 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-ca5b-account-create-update-8l5r2" event={"ID":"4bf02c5a-528e-43e3-8f9d-faa5619091e8","Type":"ContainerStarted","Data":"c40d0a03711f375d680ef1c42f2a1968feb4fd34ed7c23e0cb903f0c6d3b31a6"} Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.108082 4950 generic.go:334] "Generic (PLEG): container finished" podID="e05258ae-28d6-4c4c-8bc0-c20e6f112198" containerID="4f8a0098d6fda368230ff5bb863ce9e72a1a14336e6d63b90fa918bb309607fa" exitCode=0 Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.108199 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-4qv4l" event={"ID":"e05258ae-28d6-4c4c-8bc0-c20e6f112198","Type":"ContainerDied","Data":"4f8a0098d6fda368230ff5bb863ce9e72a1a14336e6d63b90fa918bb309607fa"} Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.108221 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-4qv4l" event={"ID":"e05258ae-28d6-4c4c-8bc0-c20e6f112198","Type":"ContainerStarted","Data":"9eeb864b7c02bf07e7df6c7ae1f89d87798d453ef4998a366c8e2ba2fb9615ea"} Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.112145 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-659bf66fb7-qw9s8" event={"ID":"746eaa13-6098-490f-9ada-f8338fe419be","Type":"ContainerStarted","Data":"88a48b23d683ef643ac9af0eb09e936825ca0a4a82641d233b8bd1fe37cb3979"} Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.114786 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-848988ddb5-9t8cx" event={"ID":"d592b2fb-1e76-410f-8b0d-a48bc85b8b58","Type":"ContainerStarted","Data":"48b0938e7e2dcfc63891667ff8ebb4b9e1a5ddd5aa6417abb99fde5118c0b9e2"} Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.118211 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.119031 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881","Type":"ContainerStarted","Data":"9f518f7d5509cf3072f9d4cb395cb983efa472b93f8514dcc307f7365f3c152a"} Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.119089 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.125193 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-ca5b-account-create-update-8l5r2" podStartSLOduration=2.122764572 podStartE2EDuration="2.122764572s" podCreationTimestamp="2026-03-18 20:59:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:59:34.117495323 +0000 UTC m=+3187.358337191" watchObservedRunningTime="2026-03-18 20:59:34.122764572 +0000 UTC m=+3187.363606440" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.264549 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.279861 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.298487 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.300077 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.312943 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.326470 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.332471 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.335903 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.337244 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.375001 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.375048 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.376283 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2vwp6" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.376412 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.376531 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.376642 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.386622 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.461455 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db3cf9ac-1719-4063-adae-b7bf72e3086e-logs\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.461836 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.461946 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db3cf9ac-1719-4063-adae-b7bf72e3086e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.462024 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.462097 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.462183 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qxkm\" (UniqueName: \"kubernetes.io/projected/db3cf9ac-1719-4063-adae-b7bf72e3086e-kube-api-access-6qxkm\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.462265 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-logs\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.462339 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95smr\" (UniqueName: \"kubernetes.io/projected/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-kube-api-access-95smr\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.462411 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.462489 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.462563 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.462626 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-config-data\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.462695 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-ceph\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.462771 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.462848 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-scripts\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.462914 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/db3cf9ac-1719-4063-adae-b7bf72e3086e-ceph\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.462977 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.463037 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.569355 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qxkm\" (UniqueName: \"kubernetes.io/projected/db3cf9ac-1719-4063-adae-b7bf72e3086e-kube-api-access-6qxkm\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.569544 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-logs\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.569603 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95smr\" (UniqueName: \"kubernetes.io/projected/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-kube-api-access-95smr\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.569625 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.569659 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.569690 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.569727 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-config-data\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.569759 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-ceph\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.569773 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.569822 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-scripts\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.569838 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/db3cf9ac-1719-4063-adae-b7bf72e3086e-ceph\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.569856 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.569894 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.569934 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db3cf9ac-1719-4063-adae-b7bf72e3086e-logs\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.569996 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.570103 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db3cf9ac-1719-4063-adae-b7bf72e3086e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.570165 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.570204 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.575139 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-logs\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.578972 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.579207 4950 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.581185 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db3cf9ac-1719-4063-adae-b7bf72e3086e-logs\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.581783 4950 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.588238 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db3cf9ac-1719-4063-adae-b7bf72e3086e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.604756 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.606866 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.607282 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qxkm\" (UniqueName: \"kubernetes.io/projected/db3cf9ac-1719-4063-adae-b7bf72e3086e-kube-api-access-6qxkm\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.608619 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.609165 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.609559 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-scripts\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.609703 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.609890 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95smr\" (UniqueName: \"kubernetes.io/projected/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-kube-api-access-95smr\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.611388 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.611866 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-config-data\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.612739 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/db3cf9ac-1719-4063-adae-b7bf72e3086e-ceph\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.617102 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-ceph\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.672028 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.686366 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.808883 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 18 20:59:34 crc kubenswrapper[4950]: I0318 20:59:34.857795 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 18 20:59:35 crc kubenswrapper[4950]: I0318 20:59:35.153156 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881","Type":"ContainerStarted","Data":"9c8813b5208184d81fc2f5fb0f0fa008d15560d6565e561c74d7e6bde6b9b1d7"} Mar 18 20:59:35 crc kubenswrapper[4950]: I0318 20:59:35.172658 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"9afecfad-52c5-439c-a3bc-a56640e9a79a","Type":"ContainerStarted","Data":"3854208f3950aaeb955817ec4920848971f2bc2fd30b83e2000d81f79735da4f"} Mar 18 20:59:35 crc kubenswrapper[4950]: I0318 20:59:35.172691 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"9afecfad-52c5-439c-a3bc-a56640e9a79a","Type":"ContainerStarted","Data":"373caf46bc5827d2431f8f34a721dab5950d03d9fb4c103817ea2d21331ac111"} Mar 18 20:59:35 crc kubenswrapper[4950]: I0318 20:59:35.216008 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=3.379553686 podStartE2EDuration="4.21598199s" podCreationTimestamp="2026-03-18 20:59:31 +0000 UTC" firstStartedPulling="2026-03-18 20:59:32.926755235 +0000 UTC m=+3186.167597103" lastFinishedPulling="2026-03-18 20:59:33.763183539 +0000 UTC m=+3187.004025407" observedRunningTime="2026-03-18 20:59:35.203080962 +0000 UTC m=+3188.443922830" watchObservedRunningTime="2026-03-18 20:59:35.21598199 +0000 UTC m=+3188.456823858" Mar 18 20:59:35 crc kubenswrapper[4950]: I0318 20:59:35.495642 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f04bf5f-c70b-48bb-8eba-78f18fd9b60c" path="/var/lib/kubelet/pods/1f04bf5f-c70b-48bb-8eba-78f18fd9b60c/volumes" Mar 18 20:59:35 crc kubenswrapper[4950]: I0318 20:59:35.496104 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42d532dd-095d-47c2-8a85-806c38a6ed36" path="/var/lib/kubelet/pods/42d532dd-095d-47c2-8a85-806c38a6ed36/volumes" Mar 18 20:59:35 crc kubenswrapper[4950]: I0318 20:59:35.579066 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 18 20:59:35 crc kubenswrapper[4950]: I0318 20:59:35.789280 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-4qv4l" Mar 18 20:59:35 crc kubenswrapper[4950]: I0318 20:59:35.819467 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 18 20:59:35 crc kubenswrapper[4950]: I0318 20:59:35.899805 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvkp5\" (UniqueName: \"kubernetes.io/projected/e05258ae-28d6-4c4c-8bc0-c20e6f112198-kube-api-access-pvkp5\") pod \"e05258ae-28d6-4c4c-8bc0-c20e6f112198\" (UID: \"e05258ae-28d6-4c4c-8bc0-c20e6f112198\") " Mar 18 20:59:35 crc kubenswrapper[4950]: I0318 20:59:35.900170 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e05258ae-28d6-4c4c-8bc0-c20e6f112198-operator-scripts\") pod \"e05258ae-28d6-4c4c-8bc0-c20e6f112198\" (UID: \"e05258ae-28d6-4c4c-8bc0-c20e6f112198\") " Mar 18 20:59:35 crc kubenswrapper[4950]: I0318 20:59:35.903839 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e05258ae-28d6-4c4c-8bc0-c20e6f112198-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e05258ae-28d6-4c4c-8bc0-c20e6f112198" (UID: "e05258ae-28d6-4c4c-8bc0-c20e6f112198"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:59:35 crc kubenswrapper[4950]: I0318 20:59:35.914668 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e05258ae-28d6-4c4c-8bc0-c20e6f112198-kube-api-access-pvkp5" (OuterVolumeSpecName: "kube-api-access-pvkp5") pod "e05258ae-28d6-4c4c-8bc0-c20e6f112198" (UID: "e05258ae-28d6-4c4c-8bc0-c20e6f112198"). InnerVolumeSpecName "kube-api-access-pvkp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:59:35 crc kubenswrapper[4950]: I0318 20:59:35.972407 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-659bf66fb7-qw9s8"] Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.007535 4950 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e05258ae-28d6-4c4c-8bc0-c20e6f112198-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.007573 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvkp5\" (UniqueName: \"kubernetes.io/projected/e05258ae-28d6-4c4c-8bc0-c20e6f112198-kube-api-access-pvkp5\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.046460 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5bfb97c7d-h56mk"] Mar 18 20:59:36 crc kubenswrapper[4950]: E0318 20:59:36.046800 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e05258ae-28d6-4c4c-8bc0-c20e6f112198" containerName="mariadb-database-create" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.046815 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="e05258ae-28d6-4c4c-8bc0-c20e6f112198" containerName="mariadb-database-create" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.046981 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="e05258ae-28d6-4c4c-8bc0-c20e6f112198" containerName="mariadb-database-create" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.062467 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.062580 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.067778 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.122272 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5bfb97c7d-h56mk"] Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.191517 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.216696 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26e882de-ea5e-4a22-889d-85c5d4347b1e-combined-ca-bundle\") pod \"horizon-5bfb97c7d-h56mk\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.216776 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/26e882de-ea5e-4a22-889d-85c5d4347b1e-config-data\") pod \"horizon-5bfb97c7d-h56mk\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.216817 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/26e882de-ea5e-4a22-889d-85c5d4347b1e-horizon-tls-certs\") pod \"horizon-5bfb97c7d-h56mk\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.216838 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/26e882de-ea5e-4a22-889d-85c5d4347b1e-horizon-secret-key\") pod \"horizon-5bfb97c7d-h56mk\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.216885 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26e882de-ea5e-4a22-889d-85c5d4347b1e-scripts\") pod \"horizon-5bfb97c7d-h56mk\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.216908 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26e882de-ea5e-4a22-889d-85c5d4347b1e-logs\") pod \"horizon-5bfb97c7d-h56mk\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.216952 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2rfj\" (UniqueName: \"kubernetes.io/projected/26e882de-ea5e-4a22-889d-85c5d4347b1e-kube-api-access-p2rfj\") pod \"horizon-5bfb97c7d-h56mk\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.236482 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-848988ddb5-9t8cx"] Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.249520 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7449579958-pdtkd"] Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.253062 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.271272 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7449579958-pdtkd"] Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.286531 4950 generic.go:334] "Generic (PLEG): container finished" podID="4bf02c5a-528e-43e3-8f9d-faa5619091e8" containerID="93b98e9080b2a203c6474765a7e3afd36f0f3a668d12adcbcd4e48111df5e0b4" exitCode=0 Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.286604 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-ca5b-account-create-update-8l5r2" event={"ID":"4bf02c5a-528e-43e3-8f9d-faa5619091e8","Type":"ContainerDied","Data":"93b98e9080b2a203c6474765a7e3afd36f0f3a668d12adcbcd4e48111df5e0b4"} Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.296676 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-4qv4l" event={"ID":"e05258ae-28d6-4c4c-8bc0-c20e6f112198","Type":"ContainerDied","Data":"9eeb864b7c02bf07e7df6c7ae1f89d87798d453ef4998a366c8e2ba2fb9615ea"} Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.296719 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9eeb864b7c02bf07e7df6c7ae1f89d87798d453ef4998a366c8e2ba2fb9615ea" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.296788 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-4qv4l" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.313483 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"db3cf9ac-1719-4063-adae-b7bf72e3086e","Type":"ContainerStarted","Data":"5ea68e06194b1c65872e2b5537b31a499102db60a74cd5b05b3ad8c226598f18"} Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.322869 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c","Type":"ContainerStarted","Data":"9f2a703883023e66500fed154fe10dda8c7c8e5d1765fec16d5347a8acd500b7"} Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.325910 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/26e882de-ea5e-4a22-889d-85c5d4347b1e-config-data\") pod \"horizon-5bfb97c7d-h56mk\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.325967 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d648add7-37b3-44fa-a3e2-4ca57d274ca0-config-data\") pod \"horizon-7449579958-pdtkd\" (UID: \"d648add7-37b3-44fa-a3e2-4ca57d274ca0\") " pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.326000 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/26e882de-ea5e-4a22-889d-85c5d4347b1e-horizon-tls-certs\") pod \"horizon-5bfb97c7d-h56mk\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.326022 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/26e882de-ea5e-4a22-889d-85c5d4347b1e-horizon-secret-key\") pod \"horizon-5bfb97c7d-h56mk\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.326053 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d648add7-37b3-44fa-a3e2-4ca57d274ca0-combined-ca-bundle\") pod \"horizon-7449579958-pdtkd\" (UID: \"d648add7-37b3-44fa-a3e2-4ca57d274ca0\") " pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.326071 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d648add7-37b3-44fa-a3e2-4ca57d274ca0-logs\") pod \"horizon-7449579958-pdtkd\" (UID: \"d648add7-37b3-44fa-a3e2-4ca57d274ca0\") " pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.326093 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7fqn\" (UniqueName: \"kubernetes.io/projected/d648add7-37b3-44fa-a3e2-4ca57d274ca0-kube-api-access-r7fqn\") pod \"horizon-7449579958-pdtkd\" (UID: \"d648add7-37b3-44fa-a3e2-4ca57d274ca0\") " pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.326124 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d648add7-37b3-44fa-a3e2-4ca57d274ca0-horizon-tls-certs\") pod \"horizon-7449579958-pdtkd\" (UID: \"d648add7-37b3-44fa-a3e2-4ca57d274ca0\") " pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.326144 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26e882de-ea5e-4a22-889d-85c5d4347b1e-scripts\") pod \"horizon-5bfb97c7d-h56mk\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.326161 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26e882de-ea5e-4a22-889d-85c5d4347b1e-logs\") pod \"horizon-5bfb97c7d-h56mk\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.326177 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d648add7-37b3-44fa-a3e2-4ca57d274ca0-horizon-secret-key\") pod \"horizon-7449579958-pdtkd\" (UID: \"d648add7-37b3-44fa-a3e2-4ca57d274ca0\") " pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.326199 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d648add7-37b3-44fa-a3e2-4ca57d274ca0-scripts\") pod \"horizon-7449579958-pdtkd\" (UID: \"d648add7-37b3-44fa-a3e2-4ca57d274ca0\") " pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.326236 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2rfj\" (UniqueName: \"kubernetes.io/projected/26e882de-ea5e-4a22-889d-85c5d4347b1e-kube-api-access-p2rfj\") pod \"horizon-5bfb97c7d-h56mk\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.326263 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26e882de-ea5e-4a22-889d-85c5d4347b1e-combined-ca-bundle\") pod \"horizon-5bfb97c7d-h56mk\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.340809 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26e882de-ea5e-4a22-889d-85c5d4347b1e-logs\") pod \"horizon-5bfb97c7d-h56mk\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.344317 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26e882de-ea5e-4a22-889d-85c5d4347b1e-scripts\") pod \"horizon-5bfb97c7d-h56mk\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.345310 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/26e882de-ea5e-4a22-889d-85c5d4347b1e-config-data\") pod \"horizon-5bfb97c7d-h56mk\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.349295 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26e882de-ea5e-4a22-889d-85c5d4347b1e-combined-ca-bundle\") pod \"horizon-5bfb97c7d-h56mk\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.358056 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/26e882de-ea5e-4a22-889d-85c5d4347b1e-horizon-tls-certs\") pod \"horizon-5bfb97c7d-h56mk\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.374593 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2rfj\" (UniqueName: \"kubernetes.io/projected/26e882de-ea5e-4a22-889d-85c5d4347b1e-kube-api-access-p2rfj\") pod \"horizon-5bfb97c7d-h56mk\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.374839 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"e22cc0cd-ebf4-4f03-91f3-1db5f2eac881","Type":"ContainerStarted","Data":"2c17445141b3cfa9254baee1a23ff781bfafd9e338f6ebae01696869ff91e500"} Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.385890 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/26e882de-ea5e-4a22-889d-85c5d4347b1e-horizon-secret-key\") pod \"horizon-5bfb97c7d-h56mk\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.402268 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.427936 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d648add7-37b3-44fa-a3e2-4ca57d274ca0-config-data\") pod \"horizon-7449579958-pdtkd\" (UID: \"d648add7-37b3-44fa-a3e2-4ca57d274ca0\") " pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.428036 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d648add7-37b3-44fa-a3e2-4ca57d274ca0-combined-ca-bundle\") pod \"horizon-7449579958-pdtkd\" (UID: \"d648add7-37b3-44fa-a3e2-4ca57d274ca0\") " pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.428054 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d648add7-37b3-44fa-a3e2-4ca57d274ca0-logs\") pod \"horizon-7449579958-pdtkd\" (UID: \"d648add7-37b3-44fa-a3e2-4ca57d274ca0\") " pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.428074 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7fqn\" (UniqueName: \"kubernetes.io/projected/d648add7-37b3-44fa-a3e2-4ca57d274ca0-kube-api-access-r7fqn\") pod \"horizon-7449579958-pdtkd\" (UID: \"d648add7-37b3-44fa-a3e2-4ca57d274ca0\") " pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.428128 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d648add7-37b3-44fa-a3e2-4ca57d274ca0-horizon-tls-certs\") pod \"horizon-7449579958-pdtkd\" (UID: \"d648add7-37b3-44fa-a3e2-4ca57d274ca0\") " pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.428161 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d648add7-37b3-44fa-a3e2-4ca57d274ca0-horizon-secret-key\") pod \"horizon-7449579958-pdtkd\" (UID: \"d648add7-37b3-44fa-a3e2-4ca57d274ca0\") " pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.428201 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d648add7-37b3-44fa-a3e2-4ca57d274ca0-scripts\") pod \"horizon-7449579958-pdtkd\" (UID: \"d648add7-37b3-44fa-a3e2-4ca57d274ca0\") " pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.435917 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d648add7-37b3-44fa-a3e2-4ca57d274ca0-config-data\") pod \"horizon-7449579958-pdtkd\" (UID: \"d648add7-37b3-44fa-a3e2-4ca57d274ca0\") " pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.437097 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d648add7-37b3-44fa-a3e2-4ca57d274ca0-logs\") pod \"horizon-7449579958-pdtkd\" (UID: \"d648add7-37b3-44fa-a3e2-4ca57d274ca0\") " pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.440518 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d648add7-37b3-44fa-a3e2-4ca57d274ca0-scripts\") pod \"horizon-7449579958-pdtkd\" (UID: \"d648add7-37b3-44fa-a3e2-4ca57d274ca0\") " pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.440880 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=4.318109173 podStartE2EDuration="5.440867636s" podCreationTimestamp="2026-03-18 20:59:31 +0000 UTC" firstStartedPulling="2026-03-18 20:59:33.178374843 +0000 UTC m=+3186.419216711" lastFinishedPulling="2026-03-18 20:59:34.301133306 +0000 UTC m=+3187.541975174" observedRunningTime="2026-03-18 20:59:36.421673012 +0000 UTC m=+3189.662514880" watchObservedRunningTime="2026-03-18 20:59:36.440867636 +0000 UTC m=+3189.681709504" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.445172 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d648add7-37b3-44fa-a3e2-4ca57d274ca0-combined-ca-bundle\") pod \"horizon-7449579958-pdtkd\" (UID: \"d648add7-37b3-44fa-a3e2-4ca57d274ca0\") " pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.449159 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d648add7-37b3-44fa-a3e2-4ca57d274ca0-horizon-secret-key\") pod \"horizon-7449579958-pdtkd\" (UID: \"d648add7-37b3-44fa-a3e2-4ca57d274ca0\") " pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.468667 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d648add7-37b3-44fa-a3e2-4ca57d274ca0-horizon-tls-certs\") pod \"horizon-7449579958-pdtkd\" (UID: \"d648add7-37b3-44fa-a3e2-4ca57d274ca0\") " pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.473642 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7fqn\" (UniqueName: \"kubernetes.io/projected/d648add7-37b3-44fa-a3e2-4ca57d274ca0-kube-api-access-r7fqn\") pod \"horizon-7449579958-pdtkd\" (UID: \"d648add7-37b3-44fa-a3e2-4ca57d274ca0\") " pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.593832 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.866093 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Mar 18 20:59:36 crc kubenswrapper[4950]: I0318 20:59:36.886745 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:37 crc kubenswrapper[4950]: I0318 20:59:37.057084 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5bfb97c7d-h56mk"] Mar 18 20:59:37 crc kubenswrapper[4950]: W0318 20:59:37.099491 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26e882de_ea5e_4a22_889d_85c5d4347b1e.slice/crio-9ecdb31d73ee163b6f37e5c983320971ef50a3c1ef517e946e9560a1c748782c WatchSource:0}: Error finding container 9ecdb31d73ee163b6f37e5c983320971ef50a3c1ef517e946e9560a1c748782c: Status 404 returned error can't find the container with id 9ecdb31d73ee163b6f37e5c983320971ef50a3c1ef517e946e9560a1c748782c Mar 18 20:59:37 crc kubenswrapper[4950]: I0318 20:59:37.392212 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7449579958-pdtkd"] Mar 18 20:59:37 crc kubenswrapper[4950]: I0318 20:59:37.410255 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"db3cf9ac-1719-4063-adae-b7bf72e3086e","Type":"ContainerStarted","Data":"90a9304b3a5865bdd4309ff5762b12b95017e024bb9737ef4071bc8e4dfd75b5"} Mar 18 20:59:37 crc kubenswrapper[4950]: I0318 20:59:37.412383 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bfb97c7d-h56mk" event={"ID":"26e882de-ea5e-4a22-889d-85c5d4347b1e","Type":"ContainerStarted","Data":"9ecdb31d73ee163b6f37e5c983320971ef50a3c1ef517e946e9560a1c748782c"} Mar 18 20:59:37 crc kubenswrapper[4950]: I0318 20:59:37.415005 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c","Type":"ContainerStarted","Data":"45ad2074f2f51e381bced335723a4cb0b3e5930cf8c8e5986cf9f570a5802dba"} Mar 18 20:59:37 crc kubenswrapper[4950]: W0318 20:59:37.423171 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd648add7_37b3_44fa_a3e2_4ca57d274ca0.slice/crio-d8bd41dd0fb448dac00697493874c550c2e8e60bfc3855134baaf9ab10dddb56 WatchSource:0}: Error finding container d8bd41dd0fb448dac00697493874c550c2e8e60bfc3855134baaf9ab10dddb56: Status 404 returned error can't find the container with id d8bd41dd0fb448dac00697493874c550c2e8e60bfc3855134baaf9ab10dddb56 Mar 18 20:59:37 crc kubenswrapper[4950]: I0318 20:59:37.781988 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-ca5b-account-create-update-8l5r2" Mar 18 20:59:37 crc kubenswrapper[4950]: I0318 20:59:37.914272 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppvkv\" (UniqueName: \"kubernetes.io/projected/4bf02c5a-528e-43e3-8f9d-faa5619091e8-kube-api-access-ppvkv\") pod \"4bf02c5a-528e-43e3-8f9d-faa5619091e8\" (UID: \"4bf02c5a-528e-43e3-8f9d-faa5619091e8\") " Mar 18 20:59:37 crc kubenswrapper[4950]: I0318 20:59:37.915703 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bf02c5a-528e-43e3-8f9d-faa5619091e8-operator-scripts\") pod \"4bf02c5a-528e-43e3-8f9d-faa5619091e8\" (UID: \"4bf02c5a-528e-43e3-8f9d-faa5619091e8\") " Mar 18 20:59:37 crc kubenswrapper[4950]: I0318 20:59:37.916299 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bf02c5a-528e-43e3-8f9d-faa5619091e8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4bf02c5a-528e-43e3-8f9d-faa5619091e8" (UID: "4bf02c5a-528e-43e3-8f9d-faa5619091e8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 20:59:37 crc kubenswrapper[4950]: I0318 20:59:37.936282 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bf02c5a-528e-43e3-8f9d-faa5619091e8-kube-api-access-ppvkv" (OuterVolumeSpecName: "kube-api-access-ppvkv") pod "4bf02c5a-528e-43e3-8f9d-faa5619091e8" (UID: "4bf02c5a-528e-43e3-8f9d-faa5619091e8"). InnerVolumeSpecName "kube-api-access-ppvkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:59:38 crc kubenswrapper[4950]: I0318 20:59:38.018749 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppvkv\" (UniqueName: \"kubernetes.io/projected/4bf02c5a-528e-43e3-8f9d-faa5619091e8-kube-api-access-ppvkv\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:38 crc kubenswrapper[4950]: I0318 20:59:38.018789 4950 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4bf02c5a-528e-43e3-8f9d-faa5619091e8-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:38 crc kubenswrapper[4950]: I0318 20:59:38.442399 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7449579958-pdtkd" event={"ID":"d648add7-37b3-44fa-a3e2-4ca57d274ca0","Type":"ContainerStarted","Data":"d8bd41dd0fb448dac00697493874c550c2e8e60bfc3855134baaf9ab10dddb56"} Mar 18 20:59:38 crc kubenswrapper[4950]: I0318 20:59:38.447050 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-ca5b-account-create-update-8l5r2" Mar 18 20:59:38 crc kubenswrapper[4950]: I0318 20:59:38.447048 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-ca5b-account-create-update-8l5r2" event={"ID":"4bf02c5a-528e-43e3-8f9d-faa5619091e8","Type":"ContainerDied","Data":"c40d0a03711f375d680ef1c42f2a1968feb4fd34ed7c23e0cb903f0c6d3b31a6"} Mar 18 20:59:38 crc kubenswrapper[4950]: I0318 20:59:38.447244 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c40d0a03711f375d680ef1c42f2a1968feb4fd34ed7c23e0cb903f0c6d3b31a6" Mar 18 20:59:38 crc kubenswrapper[4950]: I0318 20:59:38.457448 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"db3cf9ac-1719-4063-adae-b7bf72e3086e","Type":"ContainerStarted","Data":"0106c60c4df34db29fa73a405c56841abaf2806b56b6495af29c200350a42839"} Mar 18 20:59:38 crc kubenswrapper[4950]: I0318 20:59:38.457870 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="db3cf9ac-1719-4063-adae-b7bf72e3086e" containerName="glance-httpd" containerID="cri-o://0106c60c4df34db29fa73a405c56841abaf2806b56b6495af29c200350a42839" gracePeriod=30 Mar 18 20:59:38 crc kubenswrapper[4950]: I0318 20:59:38.457749 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="db3cf9ac-1719-4063-adae-b7bf72e3086e" containerName="glance-log" containerID="cri-o://90a9304b3a5865bdd4309ff5762b12b95017e024bb9737ef4071bc8e4dfd75b5" gracePeriod=30 Mar 18 20:59:38 crc kubenswrapper[4950]: I0318 20:59:38.486624 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.486601827 podStartE2EDuration="4.486601827s" podCreationTimestamp="2026-03-18 20:59:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:59:38.482451508 +0000 UTC m=+3191.723293376" watchObservedRunningTime="2026-03-18 20:59:38.486601827 +0000 UTC m=+3191.727443695" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.005156 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fkdtp"] Mar 18 20:59:39 crc kubenswrapper[4950]: E0318 20:59:39.005847 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bf02c5a-528e-43e3-8f9d-faa5619091e8" containerName="mariadb-account-create-update" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.005861 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bf02c5a-528e-43e3-8f9d-faa5619091e8" containerName="mariadb-account-create-update" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.006034 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bf02c5a-528e-43e3-8f9d-faa5619091e8" containerName="mariadb-account-create-update" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.007343 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fkdtp" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.034978 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fkdtp"] Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.046913 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73b8c337-25a3-4f8a-88c9-1e24e96589b4-utilities\") pod \"redhat-operators-fkdtp\" (UID: \"73b8c337-25a3-4f8a-88c9-1e24e96589b4\") " pod="openshift-marketplace/redhat-operators-fkdtp" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.046949 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-565xn\" (UniqueName: \"kubernetes.io/projected/73b8c337-25a3-4f8a-88c9-1e24e96589b4-kube-api-access-565xn\") pod \"redhat-operators-fkdtp\" (UID: \"73b8c337-25a3-4f8a-88c9-1e24e96589b4\") " pod="openshift-marketplace/redhat-operators-fkdtp" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.046995 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73b8c337-25a3-4f8a-88c9-1e24e96589b4-catalog-content\") pod \"redhat-operators-fkdtp\" (UID: \"73b8c337-25a3-4f8a-88c9-1e24e96589b4\") " pod="openshift-marketplace/redhat-operators-fkdtp" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.147736 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73b8c337-25a3-4f8a-88c9-1e24e96589b4-catalog-content\") pod \"redhat-operators-fkdtp\" (UID: \"73b8c337-25a3-4f8a-88c9-1e24e96589b4\") " pod="openshift-marketplace/redhat-operators-fkdtp" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.147928 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73b8c337-25a3-4f8a-88c9-1e24e96589b4-utilities\") pod \"redhat-operators-fkdtp\" (UID: \"73b8c337-25a3-4f8a-88c9-1e24e96589b4\") " pod="openshift-marketplace/redhat-operators-fkdtp" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.147982 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-565xn\" (UniqueName: \"kubernetes.io/projected/73b8c337-25a3-4f8a-88c9-1e24e96589b4-kube-api-access-565xn\") pod \"redhat-operators-fkdtp\" (UID: \"73b8c337-25a3-4f8a-88c9-1e24e96589b4\") " pod="openshift-marketplace/redhat-operators-fkdtp" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.148432 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73b8c337-25a3-4f8a-88c9-1e24e96589b4-catalog-content\") pod \"redhat-operators-fkdtp\" (UID: \"73b8c337-25a3-4f8a-88c9-1e24e96589b4\") " pod="openshift-marketplace/redhat-operators-fkdtp" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.148869 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73b8c337-25a3-4f8a-88c9-1e24e96589b4-utilities\") pod \"redhat-operators-fkdtp\" (UID: \"73b8c337-25a3-4f8a-88c9-1e24e96589b4\") " pod="openshift-marketplace/redhat-operators-fkdtp" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.202640 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-565xn\" (UniqueName: \"kubernetes.io/projected/73b8c337-25a3-4f8a-88c9-1e24e96589b4-kube-api-access-565xn\") pod \"redhat-operators-fkdtp\" (UID: \"73b8c337-25a3-4f8a-88c9-1e24e96589b4\") " pod="openshift-marketplace/redhat-operators-fkdtp" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.347718 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fkdtp" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.494261 4950 generic.go:334] "Generic (PLEG): container finished" podID="db3cf9ac-1719-4063-adae-b7bf72e3086e" containerID="0106c60c4df34db29fa73a405c56841abaf2806b56b6495af29c200350a42839" exitCode=143 Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.494291 4950 generic.go:334] "Generic (PLEG): container finished" podID="db3cf9ac-1719-4063-adae-b7bf72e3086e" containerID="90a9304b3a5865bdd4309ff5762b12b95017e024bb9737ef4071bc8e4dfd75b5" exitCode=143 Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.496887 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"db3cf9ac-1719-4063-adae-b7bf72e3086e","Type":"ContainerDied","Data":"0106c60c4df34db29fa73a405c56841abaf2806b56b6495af29c200350a42839"} Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.496918 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"db3cf9ac-1719-4063-adae-b7bf72e3086e","Type":"ContainerDied","Data":"90a9304b3a5865bdd4309ff5762b12b95017e024bb9737ef4071bc8e4dfd75b5"} Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.496930 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"db3cf9ac-1719-4063-adae-b7bf72e3086e","Type":"ContainerDied","Data":"5ea68e06194b1c65872e2b5537b31a499102db60a74cd5b05b3ad8c226598f18"} Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.496939 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ea68e06194b1c65872e2b5537b31a499102db60a74cd5b05b3ad8c226598f18" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.501738 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c","Type":"ContainerStarted","Data":"38c63abd870af6f90e474ec150a41714c4569046f6e1b35e99ffafc93726d866"} Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.501899 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c" containerName="glance-log" containerID="cri-o://45ad2074f2f51e381bced335723a4cb0b3e5930cf8c8e5986cf9f570a5802dba" gracePeriod=30 Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.502365 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c" containerName="glance-httpd" containerID="cri-o://38c63abd870af6f90e474ec150a41714c4569046f6e1b35e99ffafc93726d866" gracePeriod=30 Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.510810 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.532115 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.532093333 podStartE2EDuration="5.532093333s" podCreationTimestamp="2026-03-18 20:59:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:59:39.526856815 +0000 UTC m=+3192.767698683" watchObservedRunningTime="2026-03-18 20:59:39.532093333 +0000 UTC m=+3192.772935201" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.554068 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-combined-ca-bundle\") pod \"db3cf9ac-1719-4063-adae-b7bf72e3086e\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.554176 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db3cf9ac-1719-4063-adae-b7bf72e3086e-httpd-run\") pod \"db3cf9ac-1719-4063-adae-b7bf72e3086e\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.554204 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db3cf9ac-1719-4063-adae-b7bf72e3086e-logs\") pod \"db3cf9ac-1719-4063-adae-b7bf72e3086e\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.554260 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-internal-tls-certs\") pod \"db3cf9ac-1719-4063-adae-b7bf72e3086e\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.554531 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-config-data\") pod \"db3cf9ac-1719-4063-adae-b7bf72e3086e\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.554679 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/db3cf9ac-1719-4063-adae-b7bf72e3086e-ceph\") pod \"db3cf9ac-1719-4063-adae-b7bf72e3086e\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.554731 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qxkm\" (UniqueName: \"kubernetes.io/projected/db3cf9ac-1719-4063-adae-b7bf72e3086e-kube-api-access-6qxkm\") pod \"db3cf9ac-1719-4063-adae-b7bf72e3086e\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.554751 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-scripts\") pod \"db3cf9ac-1719-4063-adae-b7bf72e3086e\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.554820 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"db3cf9ac-1719-4063-adae-b7bf72e3086e\" (UID: \"db3cf9ac-1719-4063-adae-b7bf72e3086e\") " Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.555085 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db3cf9ac-1719-4063-adae-b7bf72e3086e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "db3cf9ac-1719-4063-adae-b7bf72e3086e" (UID: "db3cf9ac-1719-4063-adae-b7bf72e3086e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.555712 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db3cf9ac-1719-4063-adae-b7bf72e3086e-logs" (OuterVolumeSpecName: "logs") pod "db3cf9ac-1719-4063-adae-b7bf72e3086e" (UID: "db3cf9ac-1719-4063-adae-b7bf72e3086e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.556310 4950 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db3cf9ac-1719-4063-adae-b7bf72e3086e-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.556330 4950 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db3cf9ac-1719-4063-adae-b7bf72e3086e-logs\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.641584 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-scripts" (OuterVolumeSpecName: "scripts") pod "db3cf9ac-1719-4063-adae-b7bf72e3086e" (UID: "db3cf9ac-1719-4063-adae-b7bf72e3086e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.642646 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db3cf9ac-1719-4063-adae-b7bf72e3086e-ceph" (OuterVolumeSpecName: "ceph") pod "db3cf9ac-1719-4063-adae-b7bf72e3086e" (UID: "db3cf9ac-1719-4063-adae-b7bf72e3086e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.642720 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db3cf9ac-1719-4063-adae-b7bf72e3086e-kube-api-access-6qxkm" (OuterVolumeSpecName: "kube-api-access-6qxkm") pod "db3cf9ac-1719-4063-adae-b7bf72e3086e" (UID: "db3cf9ac-1719-4063-adae-b7bf72e3086e"). InnerVolumeSpecName "kube-api-access-6qxkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.642838 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "db3cf9ac-1719-4063-adae-b7bf72e3086e" (UID: "db3cf9ac-1719-4063-adae-b7bf72e3086e"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.674035 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qxkm\" (UniqueName: \"kubernetes.io/projected/db3cf9ac-1719-4063-adae-b7bf72e3086e-kube-api-access-6qxkm\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.674102 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.674137 4950 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.674159 4950 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/db3cf9ac-1719-4063-adae-b7bf72e3086e-ceph\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.705500 4950 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.732805 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db3cf9ac-1719-4063-adae-b7bf72e3086e" (UID: "db3cf9ac-1719-4063-adae-b7bf72e3086e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.740670 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "db3cf9ac-1719-4063-adae-b7bf72e3086e" (UID: "db3cf9ac-1719-4063-adae-b7bf72e3086e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.775510 4950 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.775536 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.775546 4950 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.798621 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-config-data" (OuterVolumeSpecName: "config-data") pod "db3cf9ac-1719-4063-adae-b7bf72e3086e" (UID: "db3cf9ac-1719-4063-adae-b7bf72e3086e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:39 crc kubenswrapper[4950]: I0318 20:59:39.878330 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db3cf9ac-1719-4063-adae-b7bf72e3086e-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.150907 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fkdtp"] Mar 18 20:59:40 crc kubenswrapper[4950]: W0318 20:59:40.238095 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73b8c337_25a3_4f8a_88c9_1e24e96589b4.slice/crio-8d7e9b93360b003447d1bab1f39fb654bb5a33c0ce261952638f1ea90800f101 WatchSource:0}: Error finding container 8d7e9b93360b003447d1bab1f39fb654bb5a33c0ce261952638f1ea90800f101: Status 404 returned error can't find the container with id 8d7e9b93360b003447d1bab1f39fb654bb5a33c0ce261952638f1ea90800f101 Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.541018 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fkdtp" event={"ID":"73b8c337-25a3-4f8a-88c9-1e24e96589b4","Type":"ContainerStarted","Data":"8d7e9b93360b003447d1bab1f39fb654bb5a33c0ce261952638f1ea90800f101"} Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.554802 4950 generic.go:334] "Generic (PLEG): container finished" podID="4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c" containerID="38c63abd870af6f90e474ec150a41714c4569046f6e1b35e99ffafc93726d866" exitCode=0 Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.554848 4950 generic.go:334] "Generic (PLEG): container finished" podID="4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c" containerID="45ad2074f2f51e381bced335723a4cb0b3e5930cf8c8e5986cf9f570a5802dba" exitCode=143 Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.554849 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c","Type":"ContainerDied","Data":"38c63abd870af6f90e474ec150a41714c4569046f6e1b35e99ffafc93726d866"} Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.554887 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c","Type":"ContainerDied","Data":"45ad2074f2f51e381bced335723a4cb0b3e5930cf8c8e5986cf9f570a5802dba"} Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.554953 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.661901 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.664133 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.703285 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 18 20:59:40 crc kubenswrapper[4950]: E0318 20:59:40.703693 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db3cf9ac-1719-4063-adae-b7bf72e3086e" containerName="glance-log" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.703711 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="db3cf9ac-1719-4063-adae-b7bf72e3086e" containerName="glance-log" Mar 18 20:59:40 crc kubenswrapper[4950]: E0318 20:59:40.703744 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db3cf9ac-1719-4063-adae-b7bf72e3086e" containerName="glance-httpd" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.703751 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="db3cf9ac-1719-4063-adae-b7bf72e3086e" containerName="glance-httpd" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.703904 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="db3cf9ac-1719-4063-adae-b7bf72e3086e" containerName="glance-log" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.703927 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="db3cf9ac-1719-4063-adae-b7bf72e3086e" containerName="glance-httpd" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.705014 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.711001 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.711140 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.726608 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.797660 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.863223 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-logs\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.863259 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.863294 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-ceph\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.863468 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.863577 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjzvf\" (UniqueName: \"kubernetes.io/projected/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-kube-api-access-fjzvf\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.863650 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.863709 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.864743 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.864943 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.966096 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-scripts\") pod \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.972979 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-httpd-run\") pod \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.973004 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.973063 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-ceph\") pod \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.973099 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-logs\") pod \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.973121 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-public-tls-certs\") pod \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.973143 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-config-data\") pod \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.973163 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-combined-ca-bundle\") pod \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.973259 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95smr\" (UniqueName: \"kubernetes.io/projected/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-kube-api-access-95smr\") pod \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\" (UID: \"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c\") " Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.973499 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.973589 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.973620 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-logs\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.973641 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.973676 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-ceph\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.973741 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-logs" (OuterVolumeSpecName: "logs") pod "4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c" (UID: "4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.973880 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c" (UID: "4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.973890 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.973973 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjzvf\" (UniqueName: \"kubernetes.io/projected/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-kube-api-access-fjzvf\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.974032 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.974081 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.974514 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.975313 4950 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.975327 4950 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-logs\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.975591 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-logs\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.977106 4950 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.984794 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:40 crc kubenswrapper[4950]: I0318 20:59:40.993645 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-kube-api-access-95smr" (OuterVolumeSpecName: "kube-api-access-95smr") pod "4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c" (UID: "4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c"). InnerVolumeSpecName "kube-api-access-95smr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.003676 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c" (UID: "4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.005961 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-scripts" (OuterVolumeSpecName: "scripts") pod "4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c" (UID: "4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.007780 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.009735 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-ceph\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.018810 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.021647 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.026983 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjzvf\" (UniqueName: \"kubernetes.io/projected/3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd-kube-api-access-fjzvf\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.032637 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-ceph" (OuterVolumeSpecName: "ceph") pod "4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c" (UID: "4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.077383 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.077435 4950 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.077445 4950 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-ceph\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.077453 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95smr\" (UniqueName: \"kubernetes.io/projected/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-kube-api-access-95smr\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.082455 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c" (UID: "4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.131619 4950 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.131630 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd\") " pod="openstack/glance-default-internal-api-0" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.157705 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-config-data" (OuterVolumeSpecName: "config-data") pod "4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c" (UID: "4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.167353 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c" (UID: "4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.179966 4950 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.180000 4950 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.180010 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.180019 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.413633 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.493912 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db3cf9ac-1719-4063-adae-b7bf72e3086e" path="/var/lib/kubelet/pods/db3cf9ac-1719-4063-adae-b7bf72e3086e/volumes" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.577842 4950 generic.go:334] "Generic (PLEG): container finished" podID="73b8c337-25a3-4f8a-88c9-1e24e96589b4" containerID="356c48d5bbcb0fdb45b1fae26db49dc704d48a05cb46f0487049a2aa6c0be20a" exitCode=0 Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.577916 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fkdtp" event={"ID":"73b8c337-25a3-4f8a-88c9-1e24e96589b4","Type":"ContainerDied","Data":"356c48d5bbcb0fdb45b1fae26db49dc704d48a05cb46f0487049a2aa6c0be20a"} Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.604592 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c","Type":"ContainerDied","Data":"9f2a703883023e66500fed154fe10dda8c7c8e5d1765fec16d5347a8acd500b7"} Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.604851 4950 scope.go:117] "RemoveContainer" containerID="38c63abd870af6f90e474ec150a41714c4569046f6e1b35e99ffafc93726d866" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.604973 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.677156 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.708641 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.723495 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 18 20:59:41 crc kubenswrapper[4950]: E0318 20:59:41.723996 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c" containerName="glance-log" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.724011 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c" containerName="glance-log" Mar 18 20:59:41 crc kubenswrapper[4950]: E0318 20:59:41.724032 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c" containerName="glance-httpd" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.724040 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c" containerName="glance-httpd" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.724247 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c" containerName="glance-log" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.724269 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c" containerName="glance-httpd" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.725491 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.736538 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.736771 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.745139 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.752545 4950 scope.go:117] "RemoveContainer" containerID="45ad2074f2f51e381bced335723a4cb0b3e5930cf8c8e5986cf9f570a5802dba" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.905380 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8ec3994-fd11-4e42-80c7-01857df19a74-config-data\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.905467 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c8ec3994-fd11-4e42-80c7-01857df19a74-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.905516 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8ec3994-fd11-4e42-80c7-01857df19a74-scripts\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.905549 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c8ec3994-fd11-4e42-80c7-01857df19a74-ceph\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.905572 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8ec3994-fd11-4e42-80c7-01857df19a74-logs\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.905590 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnbhc\" (UniqueName: \"kubernetes.io/projected/c8ec3994-fd11-4e42-80c7-01857df19a74-kube-api-access-jnbhc\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.905614 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8ec3994-fd11-4e42-80c7-01857df19a74-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.905654 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:41 crc kubenswrapper[4950]: I0318 20:59:41.905866 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ec3994-fd11-4e42-80c7-01857df19a74-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.008267 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.008345 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ec3994-fd11-4e42-80c7-01857df19a74-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.008504 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8ec3994-fd11-4e42-80c7-01857df19a74-config-data\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.008538 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c8ec3994-fd11-4e42-80c7-01857df19a74-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.008601 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8ec3994-fd11-4e42-80c7-01857df19a74-scripts\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.008641 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c8ec3994-fd11-4e42-80c7-01857df19a74-ceph\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.008667 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8ec3994-fd11-4e42-80c7-01857df19a74-logs\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.008692 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnbhc\" (UniqueName: \"kubernetes.io/projected/c8ec3994-fd11-4e42-80c7-01857df19a74-kube-api-access-jnbhc\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.008720 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8ec3994-fd11-4e42-80c7-01857df19a74-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.008952 4950 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.011166 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8ec3994-fd11-4e42-80c7-01857df19a74-logs\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.026838 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c8ec3994-fd11-4e42-80c7-01857df19a74-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.027344 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c8ec3994-fd11-4e42-80c7-01857df19a74-ceph\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.031176 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8ec3994-fd11-4e42-80c7-01857df19a74-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.031756 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8ec3994-fd11-4e42-80c7-01857df19a74-config-data\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.031935 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ec3994-fd11-4e42-80c7-01857df19a74-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.032284 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8ec3994-fd11-4e42-80c7-01857df19a74-scripts\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.041941 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnbhc\" (UniqueName: \"kubernetes.io/projected/c8ec3994-fd11-4e42-80c7-01857df19a74-kube-api-access-jnbhc\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.069842 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"c8ec3994-fd11-4e42-80c7-01857df19a74\") " pod="openstack/glance-default-external-api-0" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.193053 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.288139 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 18 20:59:42 crc kubenswrapper[4950]: W0318 20:59:42.304026 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ca7bb42_8e00_44cf_827d_b6dda7b3e1dd.slice/crio-1ef9162c885856b629743c2f0a75bdbc10c14404bde1c9ef0dc94335f35c7945 WatchSource:0}: Error finding container 1ef9162c885856b629743c2f0a75bdbc10c14404bde1c9ef0dc94335f35c7945: Status 404 returned error can't find the container with id 1ef9162c885856b629743c2f0a75bdbc10c14404bde1c9ef0dc94335f35c7945 Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.358572 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.405246 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.685461 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd","Type":"ContainerStarted","Data":"1ef9162c885856b629743c2f0a75bdbc10c14404bde1c9ef0dc94335f35c7945"} Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.761782 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-5szr9"] Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.764278 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-5szr9" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.766494 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.766919 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-sgzgc" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.780893 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-5szr9"] Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.951460 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-job-config-data\") pod \"manila-db-sync-5szr9\" (UID: \"a3a7d0be-1cb1-47dc-845b-25c91adb00dd\") " pod="openstack/manila-db-sync-5szr9" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.951530 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-combined-ca-bundle\") pod \"manila-db-sync-5szr9\" (UID: \"a3a7d0be-1cb1-47dc-845b-25c91adb00dd\") " pod="openstack/manila-db-sync-5szr9" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.951596 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-config-data\") pod \"manila-db-sync-5szr9\" (UID: \"a3a7d0be-1cb1-47dc-845b-25c91adb00dd\") " pod="openstack/manila-db-sync-5szr9" Mar 18 20:59:42 crc kubenswrapper[4950]: I0318 20:59:42.951686 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttnhm\" (UniqueName: \"kubernetes.io/projected/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-kube-api-access-ttnhm\") pod \"manila-db-sync-5szr9\" (UID: \"a3a7d0be-1cb1-47dc-845b-25c91adb00dd\") " pod="openstack/manila-db-sync-5szr9" Mar 18 20:59:43 crc kubenswrapper[4950]: I0318 20:59:43.061746 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-config-data\") pod \"manila-db-sync-5szr9\" (UID: \"a3a7d0be-1cb1-47dc-845b-25c91adb00dd\") " pod="openstack/manila-db-sync-5szr9" Mar 18 20:59:43 crc kubenswrapper[4950]: I0318 20:59:43.061867 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttnhm\" (UniqueName: \"kubernetes.io/projected/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-kube-api-access-ttnhm\") pod \"manila-db-sync-5szr9\" (UID: \"a3a7d0be-1cb1-47dc-845b-25c91adb00dd\") " pod="openstack/manila-db-sync-5szr9" Mar 18 20:59:43 crc kubenswrapper[4950]: I0318 20:59:43.061902 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-job-config-data\") pod \"manila-db-sync-5szr9\" (UID: \"a3a7d0be-1cb1-47dc-845b-25c91adb00dd\") " pod="openstack/manila-db-sync-5szr9" Mar 18 20:59:43 crc kubenswrapper[4950]: I0318 20:59:43.061934 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-combined-ca-bundle\") pod \"manila-db-sync-5szr9\" (UID: \"a3a7d0be-1cb1-47dc-845b-25c91adb00dd\") " pod="openstack/manila-db-sync-5szr9" Mar 18 20:59:43 crc kubenswrapper[4950]: I0318 20:59:43.077627 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-combined-ca-bundle\") pod \"manila-db-sync-5szr9\" (UID: \"a3a7d0be-1cb1-47dc-845b-25c91adb00dd\") " pod="openstack/manila-db-sync-5szr9" Mar 18 20:59:43 crc kubenswrapper[4950]: I0318 20:59:43.083148 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-config-data\") pod \"manila-db-sync-5szr9\" (UID: \"a3a7d0be-1cb1-47dc-845b-25c91adb00dd\") " pod="openstack/manila-db-sync-5szr9" Mar 18 20:59:43 crc kubenswrapper[4950]: I0318 20:59:43.083473 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-job-config-data\") pod \"manila-db-sync-5szr9\" (UID: \"a3a7d0be-1cb1-47dc-845b-25c91adb00dd\") " pod="openstack/manila-db-sync-5szr9" Mar 18 20:59:43 crc kubenswrapper[4950]: I0318 20:59:43.104970 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttnhm\" (UniqueName: \"kubernetes.io/projected/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-kube-api-access-ttnhm\") pod \"manila-db-sync-5szr9\" (UID: \"a3a7d0be-1cb1-47dc-845b-25c91adb00dd\") " pod="openstack/manila-db-sync-5szr9" Mar 18 20:59:43 crc kubenswrapper[4950]: I0318 20:59:43.121285 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-5szr9" Mar 18 20:59:43 crc kubenswrapper[4950]: I0318 20:59:43.273578 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 18 20:59:43 crc kubenswrapper[4950]: W0318 20:59:43.295749 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8ec3994_fd11_4e42_80c7_01857df19a74.slice/crio-fc453b63e89dbe6fb56fce8e0755c5fc4e37bbac14ac8822cbfc73f8ea15949e WatchSource:0}: Error finding container fc453b63e89dbe6fb56fce8e0755c5fc4e37bbac14ac8822cbfc73f8ea15949e: Status 404 returned error can't find the container with id fc453b63e89dbe6fb56fce8e0755c5fc4e37bbac14ac8822cbfc73f8ea15949e Mar 18 20:59:43 crc kubenswrapper[4950]: I0318 20:59:43.527971 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c" path="/var/lib/kubelet/pods/4e45f9f4-2015-4fef-a26b-cc10cbaa8c0c/volumes" Mar 18 20:59:43 crc kubenswrapper[4950]: I0318 20:59:43.740297 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fkdtp" event={"ID":"73b8c337-25a3-4f8a-88c9-1e24e96589b4","Type":"ContainerStarted","Data":"cdc902dcf161ce80f7fa862793b85274322032d048cecea57a76141aa626f347"} Mar 18 20:59:43 crc kubenswrapper[4950]: I0318 20:59:43.742487 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c8ec3994-fd11-4e42-80c7-01857df19a74","Type":"ContainerStarted","Data":"fc453b63e89dbe6fb56fce8e0755c5fc4e37bbac14ac8822cbfc73f8ea15949e"} Mar 18 20:59:43 crc kubenswrapper[4950]: I0318 20:59:43.746101 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd","Type":"ContainerStarted","Data":"ae0780b9be72d59ccc46465cbb93f627f37fffab0436618af674a031601c920e"} Mar 18 20:59:43 crc kubenswrapper[4950]: I0318 20:59:43.793693 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-5szr9"] Mar 18 20:59:43 crc kubenswrapper[4950]: W0318 20:59:43.795504 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3a7d0be_1cb1_47dc_845b_25c91adb00dd.slice/crio-0b3b4eaaa96bd1971c19e03fbbfb9ac13cdb5bd8634bc1fd58f0e83f7a384a8d WatchSource:0}: Error finding container 0b3b4eaaa96bd1971c19e03fbbfb9ac13cdb5bd8634bc1fd58f0e83f7a384a8d: Status 404 returned error can't find the container with id 0b3b4eaaa96bd1971c19e03fbbfb9ac13cdb5bd8634bc1fd58f0e83f7a384a8d Mar 18 20:59:44 crc kubenswrapper[4950]: I0318 20:59:44.766716 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c8ec3994-fd11-4e42-80c7-01857df19a74","Type":"ContainerStarted","Data":"abeef7bc3661bfefc779b81b2c9c6dbc93fb8afcf1a1b6d780f0de40d165f038"} Mar 18 20:59:44 crc kubenswrapper[4950]: I0318 20:59:44.770086 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd","Type":"ContainerStarted","Data":"c234e5e77e671dba222b13a58454eadc0e00c6a7164a54a23d60bab12da19a99"} Mar 18 20:59:44 crc kubenswrapper[4950]: I0318 20:59:44.775470 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-5szr9" event={"ID":"a3a7d0be-1cb1-47dc-845b-25c91adb00dd","Type":"ContainerStarted","Data":"0b3b4eaaa96bd1971c19e03fbbfb9ac13cdb5bd8634bc1fd58f0e83f7a384a8d"} Mar 18 20:59:44 crc kubenswrapper[4950]: I0318 20:59:44.795620 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.795599773 podStartE2EDuration="4.795599773s" podCreationTimestamp="2026-03-18 20:59:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:59:44.795279005 +0000 UTC m=+3198.036120873" watchObservedRunningTime="2026-03-18 20:59:44.795599773 +0000 UTC m=+3198.036441641" Mar 18 20:59:45 crc kubenswrapper[4950]: I0318 20:59:45.796193 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c8ec3994-fd11-4e42-80c7-01857df19a74","Type":"ContainerStarted","Data":"1b297a0229a9836740e70eade1ec49a5e90b7a2b6b5d12ada813af828e44b812"} Mar 18 20:59:45 crc kubenswrapper[4950]: I0318 20:59:45.833807 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.833783416 podStartE2EDuration="4.833783416s" podCreationTimestamp="2026-03-18 20:59:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 20:59:45.822065799 +0000 UTC m=+3199.062907667" watchObservedRunningTime="2026-03-18 20:59:45.833783416 +0000 UTC m=+3199.074625284" Mar 18 20:59:50 crc kubenswrapper[4950]: I0318 20:59:50.845572 4950 generic.go:334] "Generic (PLEG): container finished" podID="73b8c337-25a3-4f8a-88c9-1e24e96589b4" containerID="cdc902dcf161ce80f7fa862793b85274322032d048cecea57a76141aa626f347" exitCode=0 Mar 18 20:59:50 crc kubenswrapper[4950]: I0318 20:59:50.846513 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fkdtp" event={"ID":"73b8c337-25a3-4f8a-88c9-1e24e96589b4","Type":"ContainerDied","Data":"cdc902dcf161ce80f7fa862793b85274322032d048cecea57a76141aa626f347"} Mar 18 20:59:51 crc kubenswrapper[4950]: I0318 20:59:51.414497 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 18 20:59:51 crc kubenswrapper[4950]: I0318 20:59:51.415643 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 18 20:59:51 crc kubenswrapper[4950]: I0318 20:59:51.457260 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 18 20:59:51 crc kubenswrapper[4950]: I0318 20:59:51.494834 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 18 20:59:51 crc kubenswrapper[4950]: I0318 20:59:51.855373 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 18 20:59:51 crc kubenswrapper[4950]: I0318 20:59:51.855452 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 18 20:59:52 crc kubenswrapper[4950]: I0318 20:59:52.359323 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 18 20:59:52 crc kubenswrapper[4950]: I0318 20:59:52.359383 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 18 20:59:52 crc kubenswrapper[4950]: I0318 20:59:52.413387 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 18 20:59:52 crc kubenswrapper[4950]: I0318 20:59:52.425149 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 18 20:59:52 crc kubenswrapper[4950]: I0318 20:59:52.861895 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 18 20:59:52 crc kubenswrapper[4950]: I0318 20:59:52.861934 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 18 20:59:53 crc kubenswrapper[4950]: I0318 20:59:53.869979 4950 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 18 20:59:53 crc kubenswrapper[4950]: I0318 20:59:53.870365 4950 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 18 20:59:55 crc kubenswrapper[4950]: I0318 20:59:55.891743 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-848988ddb5-9t8cx" event={"ID":"d592b2fb-1e76-410f-8b0d-a48bc85b8b58","Type":"ContainerStarted","Data":"a8f9756eb93bf64f7b97cc373b0fd7a76096c3a6b8113ba1e0a0c9081d54251e"} Mar 18 20:59:55 crc kubenswrapper[4950]: I0318 20:59:55.892343 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-848988ddb5-9t8cx" event={"ID":"d592b2fb-1e76-410f-8b0d-a48bc85b8b58","Type":"ContainerStarted","Data":"249aabae5cccde84300386f75b0257ea6574925978759cf790ad555443ff5616"} Mar 18 20:59:55 crc kubenswrapper[4950]: I0318 20:59:55.891824 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-848988ddb5-9t8cx" podUID="d592b2fb-1e76-410f-8b0d-a48bc85b8b58" containerName="horizon-log" containerID="cri-o://249aabae5cccde84300386f75b0257ea6574925978759cf790ad555443ff5616" gracePeriod=30 Mar 18 20:59:55 crc kubenswrapper[4950]: I0318 20:59:55.891886 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-848988ddb5-9t8cx" podUID="d592b2fb-1e76-410f-8b0d-a48bc85b8b58" containerName="horizon" containerID="cri-o://a8f9756eb93bf64f7b97cc373b0fd7a76096c3a6b8113ba1e0a0c9081d54251e" gracePeriod=30 Mar 18 20:59:55 crc kubenswrapper[4950]: I0318 20:59:55.898159 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7449579958-pdtkd" event={"ID":"d648add7-37b3-44fa-a3e2-4ca57d274ca0","Type":"ContainerStarted","Data":"3422f5a69b2e76d65ddb51ebb3cdcea5b200341b7d09816469e7a46d2a2d8104"} Mar 18 20:59:55 crc kubenswrapper[4950]: I0318 20:59:55.898207 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7449579958-pdtkd" event={"ID":"d648add7-37b3-44fa-a3e2-4ca57d274ca0","Type":"ContainerStarted","Data":"bde4b6dd97889b398063d00256cae8269562c5113c089022b5e918a2870f160f"} Mar 18 20:59:55 crc kubenswrapper[4950]: I0318 20:59:55.903759 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bfb97c7d-h56mk" event={"ID":"26e882de-ea5e-4a22-889d-85c5d4347b1e","Type":"ContainerStarted","Data":"aa50a9c7f3ed240952a5dacf482e2676dcb49617b9f6bddc43a68f1aa9938ed8"} Mar 18 20:59:55 crc kubenswrapper[4950]: I0318 20:59:55.903970 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bfb97c7d-h56mk" event={"ID":"26e882de-ea5e-4a22-889d-85c5d4347b1e","Type":"ContainerStarted","Data":"36339553eb5a8810d3a99d963f412a51a7f943a23f9e3e57db6fa61a732993dc"} Mar 18 20:59:55 crc kubenswrapper[4950]: I0318 20:59:55.908046 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fkdtp" event={"ID":"73b8c337-25a3-4f8a-88c9-1e24e96589b4","Type":"ContainerStarted","Data":"ae6b85bb22621397ef4605bb22edb4ba8730fea01af4f5b8477234037824fb44"} Mar 18 20:59:55 crc kubenswrapper[4950]: I0318 20:59:55.910196 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-5szr9" event={"ID":"a3a7d0be-1cb1-47dc-845b-25c91adb00dd","Type":"ContainerStarted","Data":"a1b7b36b64c69a4bc5dbf5bc4c35c0f6a0d2d705da4804ce955847c102f01be3"} Mar 18 20:59:55 crc kubenswrapper[4950]: I0318 20:59:55.914621 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-659bf66fb7-qw9s8" event={"ID":"746eaa13-6098-490f-9ada-f8338fe419be","Type":"ContainerStarted","Data":"7fa9cd942f4e0076f6010f77d48d9fc39d501bdcc0be17307c0be365c614ca53"} Mar 18 20:59:55 crc kubenswrapper[4950]: I0318 20:59:55.914814 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-659bf66fb7-qw9s8" event={"ID":"746eaa13-6098-490f-9ada-f8338fe419be","Type":"ContainerStarted","Data":"e0003fa0675dfc576e0bb2f940084a7e9e5ddb8f132eb87eb5ac0c9eef09e779"} Mar 18 20:59:55 crc kubenswrapper[4950]: I0318 20:59:55.914756 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-659bf66fb7-qw9s8" podUID="746eaa13-6098-490f-9ada-f8338fe419be" containerName="horizon" containerID="cri-o://7fa9cd942f4e0076f6010f77d48d9fc39d501bdcc0be17307c0be365c614ca53" gracePeriod=30 Mar 18 20:59:55 crc kubenswrapper[4950]: I0318 20:59:55.914714 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-659bf66fb7-qw9s8" podUID="746eaa13-6098-490f-9ada-f8338fe419be" containerName="horizon-log" containerID="cri-o://e0003fa0675dfc576e0bb2f940084a7e9e5ddb8f132eb87eb5ac0c9eef09e779" gracePeriod=30 Mar 18 20:59:55 crc kubenswrapper[4950]: I0318 20:59:55.923107 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-848988ddb5-9t8cx" podStartSLOduration=3.160693967 podStartE2EDuration="23.923085624s" podCreationTimestamp="2026-03-18 20:59:32 +0000 UTC" firstStartedPulling="2026-03-18 20:59:33.843050176 +0000 UTC m=+3187.083892044" lastFinishedPulling="2026-03-18 20:59:54.605441843 +0000 UTC m=+3207.846283701" observedRunningTime="2026-03-18 20:59:55.915058783 +0000 UTC m=+3209.155900681" watchObservedRunningTime="2026-03-18 20:59:55.923085624 +0000 UTC m=+3209.163927492" Mar 18 20:59:55 crc kubenswrapper[4950]: I0318 20:59:55.948227 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fkdtp" podStartSLOduration=4.887335701 podStartE2EDuration="17.948207793s" podCreationTimestamp="2026-03-18 20:59:38 +0000 UTC" firstStartedPulling="2026-03-18 20:59:41.586226584 +0000 UTC m=+3194.827068452" lastFinishedPulling="2026-03-18 20:59:54.647098676 +0000 UTC m=+3207.887940544" observedRunningTime="2026-03-18 20:59:55.944746162 +0000 UTC m=+3209.185588030" watchObservedRunningTime="2026-03-18 20:59:55.948207793 +0000 UTC m=+3209.189049661" Mar 18 20:59:55 crc kubenswrapper[4950]: I0318 20:59:55.967967 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7449579958-pdtkd" podStartSLOduration=2.792747926 podStartE2EDuration="19.967950912s" podCreationTimestamp="2026-03-18 20:59:36 +0000 UTC" firstStartedPulling="2026-03-18 20:59:37.428016738 +0000 UTC m=+3190.668858606" lastFinishedPulling="2026-03-18 20:59:54.603219724 +0000 UTC m=+3207.844061592" observedRunningTime="2026-03-18 20:59:55.964803769 +0000 UTC m=+3209.205645637" watchObservedRunningTime="2026-03-18 20:59:55.967950912 +0000 UTC m=+3209.208792780" Mar 18 20:59:55 crc kubenswrapper[4950]: I0318 20:59:55.991905 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5bfb97c7d-h56mk" podStartSLOduration=3.499960348 podStartE2EDuration="20.99188533s" podCreationTimestamp="2026-03-18 20:59:35 +0000 UTC" firstStartedPulling="2026-03-18 20:59:37.112026791 +0000 UTC m=+3190.352868659" lastFinishedPulling="2026-03-18 20:59:54.603951773 +0000 UTC m=+3207.844793641" observedRunningTime="2026-03-18 20:59:55.98883751 +0000 UTC m=+3209.229679378" watchObservedRunningTime="2026-03-18 20:59:55.99188533 +0000 UTC m=+3209.232727198" Mar 18 20:59:56 crc kubenswrapper[4950]: I0318 20:59:56.039994 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-659bf66fb7-qw9s8" podStartSLOduration=3.3149027970000002 podStartE2EDuration="24.039978303s" podCreationTimestamp="2026-03-18 20:59:32 +0000 UTC" firstStartedPulling="2026-03-18 20:59:33.880389627 +0000 UTC m=+3187.121231495" lastFinishedPulling="2026-03-18 20:59:54.605465133 +0000 UTC m=+3207.846307001" observedRunningTime="2026-03-18 20:59:56.025661287 +0000 UTC m=+3209.266503155" watchObservedRunningTime="2026-03-18 20:59:56.039978303 +0000 UTC m=+3209.280820161" Mar 18 20:59:56 crc kubenswrapper[4950]: I0318 20:59:56.054087 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-5szr9" podStartSLOduration=3.153346708 podStartE2EDuration="14.054067723s" podCreationTimestamp="2026-03-18 20:59:42 +0000 UTC" firstStartedPulling="2026-03-18 20:59:43.798817948 +0000 UTC m=+3197.039659816" lastFinishedPulling="2026-03-18 20:59:54.699538963 +0000 UTC m=+3207.940380831" observedRunningTime="2026-03-18 20:59:56.052736488 +0000 UTC m=+3209.293578356" watchObservedRunningTime="2026-03-18 20:59:56.054067723 +0000 UTC m=+3209.294909591" Mar 18 20:59:56 crc kubenswrapper[4950]: I0318 20:59:56.403914 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:56 crc kubenswrapper[4950]: I0318 20:59:56.403956 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 20:59:56 crc kubenswrapper[4950]: I0318 20:59:56.443056 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 18 20:59:56 crc kubenswrapper[4950]: I0318 20:59:56.443189 4950 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 18 20:59:56 crc kubenswrapper[4950]: I0318 20:59:56.582749 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 18 20:59:56 crc kubenswrapper[4950]: I0318 20:59:56.594368 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:56 crc kubenswrapper[4950]: I0318 20:59:56.594440 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7449579958-pdtkd" Mar 18 20:59:56 crc kubenswrapper[4950]: I0318 20:59:56.901575 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 18 20:59:56 crc kubenswrapper[4950]: I0318 20:59:56.901980 4950 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 18 20:59:56 crc kubenswrapper[4950]: I0318 20:59:56.905199 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 18 20:59:59 crc kubenswrapper[4950]: I0318 20:59:59.351902 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fkdtp" Mar 18 20:59:59 crc kubenswrapper[4950]: I0318 20:59:59.352254 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fkdtp" Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.160086 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564460-krxkx"] Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.161600 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564460-krxkx" Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.164381 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.164476 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.165080 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.172266 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564460-nq4hd"] Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.173611 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564460-nq4hd" Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.179637 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.181533 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.185757 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564460-krxkx"] Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.240539 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564460-nq4hd"] Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.269143 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4qhn\" (UniqueName: \"kubernetes.io/projected/703737d3-cdc5-4db1-8b1f-2714f5261d4e-kube-api-access-c4qhn\") pod \"auto-csr-approver-29564460-krxkx\" (UID: \"703737d3-cdc5-4db1-8b1f-2714f5261d4e\") " pod="openshift-infra/auto-csr-approver-29564460-krxkx" Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.269492 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cba64ffb-d332-4b22-a17b-a0e8de767301-config-volume\") pod \"collect-profiles-29564460-nq4hd\" (UID: \"cba64ffb-d332-4b22-a17b-a0e8de767301\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564460-nq4hd" Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.269533 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cba64ffb-d332-4b22-a17b-a0e8de767301-secret-volume\") pod \"collect-profiles-29564460-nq4hd\" (UID: \"cba64ffb-d332-4b22-a17b-a0e8de767301\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564460-nq4hd" Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.269611 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtxfb\" (UniqueName: \"kubernetes.io/projected/cba64ffb-d332-4b22-a17b-a0e8de767301-kube-api-access-vtxfb\") pod \"collect-profiles-29564460-nq4hd\" (UID: \"cba64ffb-d332-4b22-a17b-a0e8de767301\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564460-nq4hd" Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.371130 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cba64ffb-d332-4b22-a17b-a0e8de767301-config-volume\") pod \"collect-profiles-29564460-nq4hd\" (UID: \"cba64ffb-d332-4b22-a17b-a0e8de767301\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564460-nq4hd" Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.371187 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cba64ffb-d332-4b22-a17b-a0e8de767301-secret-volume\") pod \"collect-profiles-29564460-nq4hd\" (UID: \"cba64ffb-d332-4b22-a17b-a0e8de767301\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564460-nq4hd" Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.371280 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtxfb\" (UniqueName: \"kubernetes.io/projected/cba64ffb-d332-4b22-a17b-a0e8de767301-kube-api-access-vtxfb\") pod \"collect-profiles-29564460-nq4hd\" (UID: \"cba64ffb-d332-4b22-a17b-a0e8de767301\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564460-nq4hd" Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.371359 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4qhn\" (UniqueName: \"kubernetes.io/projected/703737d3-cdc5-4db1-8b1f-2714f5261d4e-kube-api-access-c4qhn\") pod \"auto-csr-approver-29564460-krxkx\" (UID: \"703737d3-cdc5-4db1-8b1f-2714f5261d4e\") " pod="openshift-infra/auto-csr-approver-29564460-krxkx" Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.372054 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cba64ffb-d332-4b22-a17b-a0e8de767301-config-volume\") pod \"collect-profiles-29564460-nq4hd\" (UID: \"cba64ffb-d332-4b22-a17b-a0e8de767301\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564460-nq4hd" Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.387475 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cba64ffb-d332-4b22-a17b-a0e8de767301-secret-volume\") pod \"collect-profiles-29564460-nq4hd\" (UID: \"cba64ffb-d332-4b22-a17b-a0e8de767301\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564460-nq4hd" Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.388008 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4qhn\" (UniqueName: \"kubernetes.io/projected/703737d3-cdc5-4db1-8b1f-2714f5261d4e-kube-api-access-c4qhn\") pod \"auto-csr-approver-29564460-krxkx\" (UID: \"703737d3-cdc5-4db1-8b1f-2714f5261d4e\") " pod="openshift-infra/auto-csr-approver-29564460-krxkx" Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.396496 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtxfb\" (UniqueName: \"kubernetes.io/projected/cba64ffb-d332-4b22-a17b-a0e8de767301-kube-api-access-vtxfb\") pod \"collect-profiles-29564460-nq4hd\" (UID: \"cba64ffb-d332-4b22-a17b-a0e8de767301\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564460-nq4hd" Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.421385 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fkdtp" podUID="73b8c337-25a3-4f8a-88c9-1e24e96589b4" containerName="registry-server" probeResult="failure" output=< Mar 18 21:00:00 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 21:00:00 crc kubenswrapper[4950]: > Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.485349 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564460-krxkx" Mar 18 21:00:00 crc kubenswrapper[4950]: I0318 21:00:00.492720 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564460-nq4hd" Mar 18 21:00:01 crc kubenswrapper[4950]: I0318 21:00:01.568404 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564460-krxkx"] Mar 18 21:00:01 crc kubenswrapper[4950]: I0318 21:00:01.588975 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564460-nq4hd"] Mar 18 21:00:01 crc kubenswrapper[4950]: W0318 21:00:01.597693 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcba64ffb_d332_4b22_a17b_a0e8de767301.slice/crio-643939d403350f7bd2abcb4a4b91a0a1ce4b3e1314db60f27ac721c23d92ba07 WatchSource:0}: Error finding container 643939d403350f7bd2abcb4a4b91a0a1ce4b3e1314db60f27ac721c23d92ba07: Status 404 returned error can't find the container with id 643939d403350f7bd2abcb4a4b91a0a1ce4b3e1314db60f27ac721c23d92ba07 Mar 18 21:00:01 crc kubenswrapper[4950]: I0318 21:00:01.963960 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564460-krxkx" event={"ID":"703737d3-cdc5-4db1-8b1f-2714f5261d4e","Type":"ContainerStarted","Data":"525ec1df9585d34e79eac47d6f2f7953bf52b1b8483cfb04fcb5fde3a7010e7d"} Mar 18 21:00:01 crc kubenswrapper[4950]: I0318 21:00:01.965499 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564460-nq4hd" event={"ID":"cba64ffb-d332-4b22-a17b-a0e8de767301","Type":"ContainerStarted","Data":"741f719728fc4cb9d0652f75d31a87868d5a00f2327aa3d3edbeab03e64e1df7"} Mar 18 21:00:01 crc kubenswrapper[4950]: I0318 21:00:01.965540 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564460-nq4hd" event={"ID":"cba64ffb-d332-4b22-a17b-a0e8de767301","Type":"ContainerStarted","Data":"643939d403350f7bd2abcb4a4b91a0a1ce4b3e1314db60f27ac721c23d92ba07"} Mar 18 21:00:01 crc kubenswrapper[4950]: I0318 21:00:01.989223 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29564460-nq4hd" podStartSLOduration=1.9891978620000001 podStartE2EDuration="1.989197862s" podCreationTimestamp="2026-03-18 21:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 21:00:01.980023901 +0000 UTC m=+3215.220865769" watchObservedRunningTime="2026-03-18 21:00:01.989197862 +0000 UTC m=+3215.230039740" Mar 18 21:00:02 crc kubenswrapper[4950]: I0318 21:00:02.979553 4950 generic.go:334] "Generic (PLEG): container finished" podID="cba64ffb-d332-4b22-a17b-a0e8de767301" containerID="741f719728fc4cb9d0652f75d31a87868d5a00f2327aa3d3edbeab03e64e1df7" exitCode=0 Mar 18 21:00:02 crc kubenswrapper[4950]: I0318 21:00:02.979646 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564460-nq4hd" event={"ID":"cba64ffb-d332-4b22-a17b-a0e8de767301","Type":"ContainerDied","Data":"741f719728fc4cb9d0652f75d31a87868d5a00f2327aa3d3edbeab03e64e1df7"} Mar 18 21:00:03 crc kubenswrapper[4950]: I0318 21:00:03.000658 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-659bf66fb7-qw9s8" Mar 18 21:00:03 crc kubenswrapper[4950]: I0318 21:00:03.132534 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-848988ddb5-9t8cx" Mar 18 21:00:03 crc kubenswrapper[4950]: I0318 21:00:03.836293 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 21:00:03 crc kubenswrapper[4950]: I0318 21:00:03.836349 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 21:00:03 crc kubenswrapper[4950]: I0318 21:00:03.836397 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 21:00:03 crc kubenswrapper[4950]: I0318 21:00:03.837228 4950 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1"} pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 21:00:03 crc kubenswrapper[4950]: I0318 21:00:03.837330 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" containerID="cri-o://afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" gracePeriod=600 Mar 18 21:00:04 crc kubenswrapper[4950]: I0318 21:00:04.079035 4950 generic.go:334] "Generic (PLEG): container finished" podID="4048b439-3266-46e7-9de0-22377efacc46" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" exitCode=0 Mar 18 21:00:04 crc kubenswrapper[4950]: I0318 21:00:04.079204 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerDied","Data":"afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1"} Mar 18 21:00:04 crc kubenswrapper[4950]: I0318 21:00:04.079280 4950 scope.go:117] "RemoveContainer" containerID="659b90655ee76a1c728bf80c079e29e36775b2f88753e6ed85575c3829a23583" Mar 18 21:00:04 crc kubenswrapper[4950]: I0318 21:00:04.474086 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564460-nq4hd" Mar 18 21:00:04 crc kubenswrapper[4950]: I0318 21:00:04.559159 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtxfb\" (UniqueName: \"kubernetes.io/projected/cba64ffb-d332-4b22-a17b-a0e8de767301-kube-api-access-vtxfb\") pod \"cba64ffb-d332-4b22-a17b-a0e8de767301\" (UID: \"cba64ffb-d332-4b22-a17b-a0e8de767301\") " Mar 18 21:00:04 crc kubenswrapper[4950]: I0318 21:00:04.559329 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cba64ffb-d332-4b22-a17b-a0e8de767301-config-volume\") pod \"cba64ffb-d332-4b22-a17b-a0e8de767301\" (UID: \"cba64ffb-d332-4b22-a17b-a0e8de767301\") " Mar 18 21:00:04 crc kubenswrapper[4950]: I0318 21:00:04.559373 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cba64ffb-d332-4b22-a17b-a0e8de767301-secret-volume\") pod \"cba64ffb-d332-4b22-a17b-a0e8de767301\" (UID: \"cba64ffb-d332-4b22-a17b-a0e8de767301\") " Mar 18 21:00:04 crc kubenswrapper[4950]: I0318 21:00:04.561211 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cba64ffb-d332-4b22-a17b-a0e8de767301-config-volume" (OuterVolumeSpecName: "config-volume") pod "cba64ffb-d332-4b22-a17b-a0e8de767301" (UID: "cba64ffb-d332-4b22-a17b-a0e8de767301"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 21:00:04 crc kubenswrapper[4950]: I0318 21:00:04.568973 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cba64ffb-d332-4b22-a17b-a0e8de767301-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "cba64ffb-d332-4b22-a17b-a0e8de767301" (UID: "cba64ffb-d332-4b22-a17b-a0e8de767301"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:04 crc kubenswrapper[4950]: I0318 21:00:04.571590 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cba64ffb-d332-4b22-a17b-a0e8de767301-kube-api-access-vtxfb" (OuterVolumeSpecName: "kube-api-access-vtxfb") pod "cba64ffb-d332-4b22-a17b-a0e8de767301" (UID: "cba64ffb-d332-4b22-a17b-a0e8de767301"). InnerVolumeSpecName "kube-api-access-vtxfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:00:04 crc kubenswrapper[4950]: E0318 21:00:04.581557 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:00:04 crc kubenswrapper[4950]: I0318 21:00:04.661647 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtxfb\" (UniqueName: \"kubernetes.io/projected/cba64ffb-d332-4b22-a17b-a0e8de767301-kube-api-access-vtxfb\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:04 crc kubenswrapper[4950]: I0318 21:00:04.661683 4950 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cba64ffb-d332-4b22-a17b-a0e8de767301-config-volume\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:04 crc kubenswrapper[4950]: I0318 21:00:04.661693 4950 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cba64ffb-d332-4b22-a17b-a0e8de767301-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:04 crc kubenswrapper[4950]: I0318 21:00:04.688506 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564415-rpc7c"] Mar 18 21:00:04 crc kubenswrapper[4950]: I0318 21:00:04.699274 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564415-rpc7c"] Mar 18 21:00:05 crc kubenswrapper[4950]: I0318 21:00:05.092648 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:00:05 crc kubenswrapper[4950]: E0318 21:00:05.093102 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:00:05 crc kubenswrapper[4950]: I0318 21:00:05.096490 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564460-nq4hd" event={"ID":"cba64ffb-d332-4b22-a17b-a0e8de767301","Type":"ContainerDied","Data":"643939d403350f7bd2abcb4a4b91a0a1ce4b3e1314db60f27ac721c23d92ba07"} Mar 18 21:00:05 crc kubenswrapper[4950]: I0318 21:00:05.096543 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="643939d403350f7bd2abcb4a4b91a0a1ce4b3e1314db60f27ac721c23d92ba07" Mar 18 21:00:05 crc kubenswrapper[4950]: I0318 21:00:05.096600 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564460-nq4hd" Mar 18 21:00:05 crc kubenswrapper[4950]: I0318 21:00:05.516801 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49fb4fb7-6a60-4eeb-b1bf-349a1948b08f" path="/var/lib/kubelet/pods/49fb4fb7-6a60-4eeb-b1bf-349a1948b08f/volumes" Mar 18 21:00:06 crc kubenswrapper[4950]: I0318 21:00:06.405181 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5bfb97c7d-h56mk" podUID="26e882de-ea5e-4a22-889d-85c5d4347b1e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.9:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.9:8443: connect: connection refused" Mar 18 21:00:06 crc kubenswrapper[4950]: I0318 21:00:06.596611 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7449579958-pdtkd" podUID="d648add7-37b3-44fa-a3e2-4ca57d274ca0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.10:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.10:8443: connect: connection refused" Mar 18 21:00:09 crc kubenswrapper[4950]: I0318 21:00:09.152299 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564460-krxkx" event={"ID":"703737d3-cdc5-4db1-8b1f-2714f5261d4e","Type":"ContainerStarted","Data":"b18dc7ae008b2e958bcd61cbb5faf4c4cf93ce26ea3399ab2cf053160cf355c1"} Mar 18 21:00:10 crc kubenswrapper[4950]: I0318 21:00:10.174631 4950 generic.go:334] "Generic (PLEG): container finished" podID="703737d3-cdc5-4db1-8b1f-2714f5261d4e" containerID="b18dc7ae008b2e958bcd61cbb5faf4c4cf93ce26ea3399ab2cf053160cf355c1" exitCode=0 Mar 18 21:00:10 crc kubenswrapper[4950]: I0318 21:00:10.174692 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564460-krxkx" event={"ID":"703737d3-cdc5-4db1-8b1f-2714f5261d4e","Type":"ContainerDied","Data":"b18dc7ae008b2e958bcd61cbb5faf4c4cf93ce26ea3399ab2cf053160cf355c1"} Mar 18 21:00:10 crc kubenswrapper[4950]: I0318 21:00:10.416964 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fkdtp" podUID="73b8c337-25a3-4f8a-88c9-1e24e96589b4" containerName="registry-server" probeResult="failure" output=< Mar 18 21:00:10 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 21:00:10 crc kubenswrapper[4950]: > Mar 18 21:00:11 crc kubenswrapper[4950]: I0318 21:00:11.620673 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564460-krxkx" Mar 18 21:00:11 crc kubenswrapper[4950]: I0318 21:00:11.772566 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4qhn\" (UniqueName: \"kubernetes.io/projected/703737d3-cdc5-4db1-8b1f-2714f5261d4e-kube-api-access-c4qhn\") pod \"703737d3-cdc5-4db1-8b1f-2714f5261d4e\" (UID: \"703737d3-cdc5-4db1-8b1f-2714f5261d4e\") " Mar 18 21:00:11 crc kubenswrapper[4950]: I0318 21:00:11.803471 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/703737d3-cdc5-4db1-8b1f-2714f5261d4e-kube-api-access-c4qhn" (OuterVolumeSpecName: "kube-api-access-c4qhn") pod "703737d3-cdc5-4db1-8b1f-2714f5261d4e" (UID: "703737d3-cdc5-4db1-8b1f-2714f5261d4e"). InnerVolumeSpecName "kube-api-access-c4qhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:00:11 crc kubenswrapper[4950]: I0318 21:00:11.876113 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4qhn\" (UniqueName: \"kubernetes.io/projected/703737d3-cdc5-4db1-8b1f-2714f5261d4e-kube-api-access-c4qhn\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:12 crc kubenswrapper[4950]: I0318 21:00:12.193078 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564460-krxkx" event={"ID":"703737d3-cdc5-4db1-8b1f-2714f5261d4e","Type":"ContainerDied","Data":"525ec1df9585d34e79eac47d6f2f7953bf52b1b8483cfb04fcb5fde3a7010e7d"} Mar 18 21:00:12 crc kubenswrapper[4950]: I0318 21:00:12.193125 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="525ec1df9585d34e79eac47d6f2f7953bf52b1b8483cfb04fcb5fde3a7010e7d" Mar 18 21:00:12 crc kubenswrapper[4950]: I0318 21:00:12.193134 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564460-krxkx" Mar 18 21:00:12 crc kubenswrapper[4950]: I0318 21:00:12.236204 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564454-jll6h"] Mar 18 21:00:12 crc kubenswrapper[4950]: I0318 21:00:12.243577 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564454-jll6h"] Mar 18 21:00:13 crc kubenswrapper[4950]: I0318 21:00:13.488936 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b83aea1b-8a6b-4d04-8f82-2293bc696179" path="/var/lib/kubelet/pods/b83aea1b-8a6b-4d04-8f82-2293bc696179/volumes" Mar 18 21:00:14 crc kubenswrapper[4950]: I0318 21:00:14.211397 4950 generic.go:334] "Generic (PLEG): container finished" podID="a3a7d0be-1cb1-47dc-845b-25c91adb00dd" containerID="a1b7b36b64c69a4bc5dbf5bc4c35c0f6a0d2d705da4804ce955847c102f01be3" exitCode=0 Mar 18 21:00:14 crc kubenswrapper[4950]: I0318 21:00:14.211498 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-5szr9" event={"ID":"a3a7d0be-1cb1-47dc-845b-25c91adb00dd","Type":"ContainerDied","Data":"a1b7b36b64c69a4bc5dbf5bc4c35c0f6a0d2d705da4804ce955847c102f01be3"} Mar 18 21:00:15 crc kubenswrapper[4950]: I0318 21:00:15.745529 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-5szr9" Mar 18 21:00:15 crc kubenswrapper[4950]: I0318 21:00:15.851130 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-combined-ca-bundle\") pod \"a3a7d0be-1cb1-47dc-845b-25c91adb00dd\" (UID: \"a3a7d0be-1cb1-47dc-845b-25c91adb00dd\") " Mar 18 21:00:15 crc kubenswrapper[4950]: I0318 21:00:15.851219 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-job-config-data\") pod \"a3a7d0be-1cb1-47dc-845b-25c91adb00dd\" (UID: \"a3a7d0be-1cb1-47dc-845b-25c91adb00dd\") " Mar 18 21:00:15 crc kubenswrapper[4950]: I0318 21:00:15.851337 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-config-data\") pod \"a3a7d0be-1cb1-47dc-845b-25c91adb00dd\" (UID: \"a3a7d0be-1cb1-47dc-845b-25c91adb00dd\") " Mar 18 21:00:15 crc kubenswrapper[4950]: I0318 21:00:15.851468 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttnhm\" (UniqueName: \"kubernetes.io/projected/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-kube-api-access-ttnhm\") pod \"a3a7d0be-1cb1-47dc-845b-25c91adb00dd\" (UID: \"a3a7d0be-1cb1-47dc-845b-25c91adb00dd\") " Mar 18 21:00:15 crc kubenswrapper[4950]: I0318 21:00:15.859442 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-kube-api-access-ttnhm" (OuterVolumeSpecName: "kube-api-access-ttnhm") pod "a3a7d0be-1cb1-47dc-845b-25c91adb00dd" (UID: "a3a7d0be-1cb1-47dc-845b-25c91adb00dd"). InnerVolumeSpecName "kube-api-access-ttnhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:00:15 crc kubenswrapper[4950]: I0318 21:00:15.863521 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-config-data" (OuterVolumeSpecName: "config-data") pod "a3a7d0be-1cb1-47dc-845b-25c91adb00dd" (UID: "a3a7d0be-1cb1-47dc-845b-25c91adb00dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:15 crc kubenswrapper[4950]: I0318 21:00:15.863561 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "a3a7d0be-1cb1-47dc-845b-25c91adb00dd" (UID: "a3a7d0be-1cb1-47dc-845b-25c91adb00dd"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:15 crc kubenswrapper[4950]: I0318 21:00:15.892028 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3a7d0be-1cb1-47dc-845b-25c91adb00dd" (UID: "a3a7d0be-1cb1-47dc-845b-25c91adb00dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:15 crc kubenswrapper[4950]: I0318 21:00:15.953473 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:15 crc kubenswrapper[4950]: I0318 21:00:15.953516 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttnhm\" (UniqueName: \"kubernetes.io/projected/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-kube-api-access-ttnhm\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:15 crc kubenswrapper[4950]: I0318 21:00:15.953527 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:15 crc kubenswrapper[4950]: I0318 21:00:15.953537 4950 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/a3a7d0be-1cb1-47dc-845b-25c91adb00dd-job-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.228823 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-5szr9" event={"ID":"a3a7d0be-1cb1-47dc-845b-25c91adb00dd","Type":"ContainerDied","Data":"0b3b4eaaa96bd1971c19e03fbbfb9ac13cdb5bd8634bc1fd58f0e83f7a384a8d"} Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.228874 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b3b4eaaa96bd1971c19e03fbbfb9ac13cdb5bd8634bc1fd58f0e83f7a384a8d" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.228921 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-5szr9" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.420271 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5bfb97c7d-h56mk" podUID="26e882de-ea5e-4a22-889d-85c5d4347b1e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.9:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.9:8443: connect: connection refused" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.583808 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Mar 18 21:00:16 crc kubenswrapper[4950]: E0318 21:00:16.584206 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3a7d0be-1cb1-47dc-845b-25c91adb00dd" containerName="manila-db-sync" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.584229 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3a7d0be-1cb1-47dc-845b-25c91adb00dd" containerName="manila-db-sync" Mar 18 21:00:16 crc kubenswrapper[4950]: E0318 21:00:16.584241 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="703737d3-cdc5-4db1-8b1f-2714f5261d4e" containerName="oc" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.584247 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="703737d3-cdc5-4db1-8b1f-2714f5261d4e" containerName="oc" Mar 18 21:00:16 crc kubenswrapper[4950]: E0318 21:00:16.584264 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cba64ffb-d332-4b22-a17b-a0e8de767301" containerName="collect-profiles" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.584270 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="cba64ffb-d332-4b22-a17b-a0e8de767301" containerName="collect-profiles" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.584445 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3a7d0be-1cb1-47dc-845b-25c91adb00dd" containerName="manila-db-sync" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.584459 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="703737d3-cdc5-4db1-8b1f-2714f5261d4e" containerName="oc" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.584476 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="cba64ffb-d332-4b22-a17b-a0e8de767301" containerName="collect-profiles" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.585457 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.594713 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.594997 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.595294 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.595486 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-sgzgc" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.596536 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7449579958-pdtkd" podUID="d648add7-37b3-44fa-a3e2-4ca57d274ca0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.10:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.10:8443: connect: connection refused" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.622490 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.624110 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.627852 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.655567 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.679908 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.773807 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/47aef657-3e55-4afa-9b2f-5650cb789e8a-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.773876 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/47aef657-3e55-4afa-9b2f-5650cb789e8a-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.773912 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.773957 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/47aef657-3e55-4afa-9b2f-5650cb789e8a-ceph\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.773985 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-config-data\") pod \"manila-scheduler-0\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.774026 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx4tn\" (UniqueName: \"kubernetes.io/projected/fcb05ff2-359e-4728-a148-b8f8678e4ce8-kube-api-access-jx4tn\") pod \"manila-scheduler-0\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.774053 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.774086 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fcb05ff2-359e-4728-a148-b8f8678e4ce8-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.774123 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq94g\" (UniqueName: \"kubernetes.io/projected/47aef657-3e55-4afa-9b2f-5650cb789e8a-kube-api-access-pq94g\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.774145 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-config-data\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.774167 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.774193 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.774234 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-scripts\") pod \"manila-scheduler-0\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.774271 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-scripts\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.807568 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f55f6cb75-46f5l"] Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.809116 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.842901 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.844803 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.854610 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.874830 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f55f6cb75-46f5l"] Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.876178 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-scripts\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.876247 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/47aef657-3e55-4afa-9b2f-5650cb789e8a-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.876278 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/47aef657-3e55-4afa-9b2f-5650cb789e8a-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.876303 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.876333 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/47aef657-3e55-4afa-9b2f-5650cb789e8a-ceph\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.876351 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-config-data\") pod \"manila-scheduler-0\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.876381 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx4tn\" (UniqueName: \"kubernetes.io/projected/fcb05ff2-359e-4728-a148-b8f8678e4ce8-kube-api-access-jx4tn\") pod \"manila-scheduler-0\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.876401 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.876437 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fcb05ff2-359e-4728-a148-b8f8678e4ce8-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.876465 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq94g\" (UniqueName: \"kubernetes.io/projected/47aef657-3e55-4afa-9b2f-5650cb789e8a-kube-api-access-pq94g\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.876479 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-config-data\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.876496 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.876515 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.876548 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-scripts\") pod \"manila-scheduler-0\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.878049 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fcb05ff2-359e-4728-a148-b8f8678e4ce8-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.879930 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/47aef657-3e55-4afa-9b2f-5650cb789e8a-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.879998 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/47aef657-3e55-4afa-9b2f-5650cb789e8a-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.889547 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-config-data\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.890983 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-config-data\") pod \"manila-scheduler-0\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.892929 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/47aef657-3e55-4afa-9b2f-5650cb789e8a-ceph\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.903919 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-scripts\") pod \"manila-scheduler-0\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.904234 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.904585 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-scripts\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.909823 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.912895 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.916387 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.920834 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.944102 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq94g\" (UniqueName: \"kubernetes.io/projected/47aef657-3e55-4afa-9b2f-5650cb789e8a-kube-api-access-pq94g\") pod \"manila-share-share1-0\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.964085 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx4tn\" (UniqueName: \"kubernetes.io/projected/fcb05ff2-359e-4728-a148-b8f8678e4ce8-kube-api-access-jx4tn\") pod \"manila-scheduler-0\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.978419 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-config-data\") pod \"manila-api-0\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " pod="openstack/manila-api-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.978463 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab176318-8186-41e2-8277-1d7b4d4f7618-logs\") pod \"manila-api-0\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " pod="openstack/manila-api-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.978496 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f9997bd5-b01a-490e-b4c3-b258b277d73c-ovsdbserver-sb\") pod \"dnsmasq-dns-7f55f6cb75-46f5l\" (UID: \"f9997bd5-b01a-490e-b4c3-b258b277d73c\") " pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.978558 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9997bd5-b01a-490e-b4c3-b258b277d73c-config\") pod \"dnsmasq-dns-7f55f6cb75-46f5l\" (UID: \"f9997bd5-b01a-490e-b4c3-b258b277d73c\") " pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.978576 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk4nq\" (UniqueName: \"kubernetes.io/projected/f9997bd5-b01a-490e-b4c3-b258b277d73c-kube-api-access-zk4nq\") pod \"dnsmasq-dns-7f55f6cb75-46f5l\" (UID: \"f9997bd5-b01a-490e-b4c3-b258b277d73c\") " pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.978608 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ab176318-8186-41e2-8277-1d7b4d4f7618-etc-machine-id\") pod \"manila-api-0\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " pod="openstack/manila-api-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.978642 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w685v\" (UniqueName: \"kubernetes.io/projected/ab176318-8186-41e2-8277-1d7b4d4f7618-kube-api-access-w685v\") pod \"manila-api-0\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " pod="openstack/manila-api-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.978665 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9997bd5-b01a-490e-b4c3-b258b277d73c-dns-svc\") pod \"dnsmasq-dns-7f55f6cb75-46f5l\" (UID: \"f9997bd5-b01a-490e-b4c3-b258b277d73c\") " pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.978688 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " pod="openstack/manila-api-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.978712 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f9997bd5-b01a-490e-b4c3-b258b277d73c-openstack-edpm-ipam\") pod \"dnsmasq-dns-7f55f6cb75-46f5l\" (UID: \"f9997bd5-b01a-490e-b4c3-b258b277d73c\") " pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.978729 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f9997bd5-b01a-490e-b4c3-b258b277d73c-ovsdbserver-nb\") pod \"dnsmasq-dns-7f55f6cb75-46f5l\" (UID: \"f9997bd5-b01a-490e-b4c3-b258b277d73c\") " pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.978749 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-scripts\") pod \"manila-api-0\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " pod="openstack/manila-api-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.978774 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-config-data-custom\") pod \"manila-api-0\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " pod="openstack/manila-api-0" Mar 18 21:00:16 crc kubenswrapper[4950]: I0318 21:00:16.991600 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.081548 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w685v\" (UniqueName: \"kubernetes.io/projected/ab176318-8186-41e2-8277-1d7b4d4f7618-kube-api-access-w685v\") pod \"manila-api-0\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " pod="openstack/manila-api-0" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.082335 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9997bd5-b01a-490e-b4c3-b258b277d73c-dns-svc\") pod \"dnsmasq-dns-7f55f6cb75-46f5l\" (UID: \"f9997bd5-b01a-490e-b4c3-b258b277d73c\") " pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.082378 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f9997bd5-b01a-490e-b4c3-b258b277d73c-dns-svc\") pod \"dnsmasq-dns-7f55f6cb75-46f5l\" (UID: \"f9997bd5-b01a-490e-b4c3-b258b277d73c\") " pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.082424 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " pod="openstack/manila-api-0" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.082461 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f9997bd5-b01a-490e-b4c3-b258b277d73c-openstack-edpm-ipam\") pod \"dnsmasq-dns-7f55f6cb75-46f5l\" (UID: \"f9997bd5-b01a-490e-b4c3-b258b277d73c\") " pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.082484 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f9997bd5-b01a-490e-b4c3-b258b277d73c-ovsdbserver-nb\") pod \"dnsmasq-dns-7f55f6cb75-46f5l\" (UID: \"f9997bd5-b01a-490e-b4c3-b258b277d73c\") " pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.082519 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-scripts\") pod \"manila-api-0\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " pod="openstack/manila-api-0" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.082559 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-config-data-custom\") pod \"manila-api-0\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " pod="openstack/manila-api-0" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.082664 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-config-data\") pod \"manila-api-0\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " pod="openstack/manila-api-0" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.082693 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab176318-8186-41e2-8277-1d7b4d4f7618-logs\") pod \"manila-api-0\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " pod="openstack/manila-api-0" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.082727 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f9997bd5-b01a-490e-b4c3-b258b277d73c-ovsdbserver-sb\") pod \"dnsmasq-dns-7f55f6cb75-46f5l\" (UID: \"f9997bd5-b01a-490e-b4c3-b258b277d73c\") " pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.082809 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9997bd5-b01a-490e-b4c3-b258b277d73c-config\") pod \"dnsmasq-dns-7f55f6cb75-46f5l\" (UID: \"f9997bd5-b01a-490e-b4c3-b258b277d73c\") " pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.082832 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk4nq\" (UniqueName: \"kubernetes.io/projected/f9997bd5-b01a-490e-b4c3-b258b277d73c-kube-api-access-zk4nq\") pod \"dnsmasq-dns-7f55f6cb75-46f5l\" (UID: \"f9997bd5-b01a-490e-b4c3-b258b277d73c\") " pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.082874 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ab176318-8186-41e2-8277-1d7b4d4f7618-etc-machine-id\") pod \"manila-api-0\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " pod="openstack/manila-api-0" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.082999 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ab176318-8186-41e2-8277-1d7b4d4f7618-etc-machine-id\") pod \"manila-api-0\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " pod="openstack/manila-api-0" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.083293 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f9997bd5-b01a-490e-b4c3-b258b277d73c-ovsdbserver-nb\") pod \"dnsmasq-dns-7f55f6cb75-46f5l\" (UID: \"f9997bd5-b01a-490e-b4c3-b258b277d73c\") " pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.085231 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f9997bd5-b01a-490e-b4c3-b258b277d73c-ovsdbserver-sb\") pod \"dnsmasq-dns-7f55f6cb75-46f5l\" (UID: \"f9997bd5-b01a-490e-b4c3-b258b277d73c\") " pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.086272 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab176318-8186-41e2-8277-1d7b4d4f7618-logs\") pod \"manila-api-0\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " pod="openstack/manila-api-0" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.086840 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9997bd5-b01a-490e-b4c3-b258b277d73c-config\") pod \"dnsmasq-dns-7f55f6cb75-46f5l\" (UID: \"f9997bd5-b01a-490e-b4c3-b258b277d73c\") " pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.092897 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f9997bd5-b01a-490e-b4c3-b258b277d73c-openstack-edpm-ipam\") pod \"dnsmasq-dns-7f55f6cb75-46f5l\" (UID: \"f9997bd5-b01a-490e-b4c3-b258b277d73c\") " pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.099004 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-config-data-custom\") pod \"manila-api-0\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " pod="openstack/manila-api-0" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.100644 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " pod="openstack/manila-api-0" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.101811 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-scripts\") pod \"manila-api-0\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " pod="openstack/manila-api-0" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.108130 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-config-data\") pod \"manila-api-0\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " pod="openstack/manila-api-0" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.125821 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk4nq\" (UniqueName: \"kubernetes.io/projected/f9997bd5-b01a-490e-b4c3-b258b277d73c-kube-api-access-zk4nq\") pod \"dnsmasq-dns-7f55f6cb75-46f5l\" (UID: \"f9997bd5-b01a-490e-b4c3-b258b277d73c\") " pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.129963 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.131244 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w685v\" (UniqueName: \"kubernetes.io/projected/ab176318-8186-41e2-8277-1d7b4d4f7618-kube-api-access-w685v\") pod \"manila-api-0\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " pod="openstack/manila-api-0" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.166915 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.258494 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.773043 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Mar 18 21:00:17 crc kubenswrapper[4950]: I0318 21:00:17.965486 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f55f6cb75-46f5l"] Mar 18 21:00:18 crc kubenswrapper[4950]: I0318 21:00:18.122268 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Mar 18 21:00:18 crc kubenswrapper[4950]: I0318 21:00:18.232952 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Mar 18 21:00:18 crc kubenswrapper[4950]: W0318 21:00:18.239871 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab176318_8186_41e2_8277_1d7b4d4f7618.slice/crio-5d40b272004a94b01f6bf929379d39a839d989329f4cc838e0cbf7bda19e4c69 WatchSource:0}: Error finding container 5d40b272004a94b01f6bf929379d39a839d989329f4cc838e0cbf7bda19e4c69: Status 404 returned error can't find the container with id 5d40b272004a94b01f6bf929379d39a839d989329f4cc838e0cbf7bda19e4c69 Mar 18 21:00:18 crc kubenswrapper[4950]: I0318 21:00:18.324944 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"fcb05ff2-359e-4728-a148-b8f8678e4ce8","Type":"ContainerStarted","Data":"f9985702ff146febd91169282cf4d3f72ad678ca7361b090ba881d22831b4645"} Mar 18 21:00:18 crc kubenswrapper[4950]: I0318 21:00:18.344678 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" event={"ID":"f9997bd5-b01a-490e-b4c3-b258b277d73c","Type":"ContainerStarted","Data":"1ff6d06e56f40a26310062218ea41b9623ae58c57de6ed780784a7d4e1278cc4"} Mar 18 21:00:18 crc kubenswrapper[4950]: I0318 21:00:18.355075 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"47aef657-3e55-4afa-9b2f-5650cb789e8a","Type":"ContainerStarted","Data":"2ec34286594a538e040b8e0121ad51759a803001ac959bd5cf4bcfd2db5edf9a"} Mar 18 21:00:18 crc kubenswrapper[4950]: I0318 21:00:18.376503 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"ab176318-8186-41e2-8277-1d7b4d4f7618","Type":"ContainerStarted","Data":"5d40b272004a94b01f6bf929379d39a839d989329f4cc838e0cbf7bda19e4c69"} Mar 18 21:00:18 crc kubenswrapper[4950]: I0318 21:00:18.480391 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:00:18 crc kubenswrapper[4950]: E0318 21:00:18.480670 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:00:19 crc kubenswrapper[4950]: I0318 21:00:19.402582 4950 generic.go:334] "Generic (PLEG): container finished" podID="f9997bd5-b01a-490e-b4c3-b258b277d73c" containerID="5e6603a132510777e48822dbf24378c7d9a73cc01a470a91c5192de6e447b680" exitCode=0 Mar 18 21:00:19 crc kubenswrapper[4950]: I0318 21:00:19.402892 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" event={"ID":"f9997bd5-b01a-490e-b4c3-b258b277d73c","Type":"ContainerDied","Data":"5e6603a132510777e48822dbf24378c7d9a73cc01a470a91c5192de6e447b680"} Mar 18 21:00:19 crc kubenswrapper[4950]: I0318 21:00:19.422548 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"ab176318-8186-41e2-8277-1d7b4d4f7618","Type":"ContainerStarted","Data":"abe31e08ae8889a6500d3527ed7dfdc24654d4f161ce78309046281901aaefef"} Mar 18 21:00:20 crc kubenswrapper[4950]: I0318 21:00:20.024085 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Mar 18 21:00:20 crc kubenswrapper[4950]: I0318 21:00:20.465653 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fkdtp" podUID="73b8c337-25a3-4f8a-88c9-1e24e96589b4" containerName="registry-server" probeResult="failure" output=< Mar 18 21:00:20 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 21:00:20 crc kubenswrapper[4950]: > Mar 18 21:00:20 crc kubenswrapper[4950]: I0318 21:00:20.466962 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" event={"ID":"f9997bd5-b01a-490e-b4c3-b258b277d73c","Type":"ContainerStarted","Data":"e399b454b13cf53fd75a270535b9876f2c68d9f49ddf938801e619b9f69bdf5b"} Mar 18 21:00:20 crc kubenswrapper[4950]: I0318 21:00:20.467015 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" Mar 18 21:00:20 crc kubenswrapper[4950]: I0318 21:00:20.490724 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"ab176318-8186-41e2-8277-1d7b4d4f7618","Type":"ContainerStarted","Data":"3165fe6d9e468847f34b57eac9432e3c988f8d350e96d51224903e78665a4694"} Mar 18 21:00:20 crc kubenswrapper[4950]: I0318 21:00:20.491561 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Mar 18 21:00:20 crc kubenswrapper[4950]: I0318 21:00:20.510969 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"fcb05ff2-359e-4728-a148-b8f8678e4ce8","Type":"ContainerStarted","Data":"8ab94c7a21fb072ce10d18b5450749842b6d8a52d33e406c4dde29e82d766b8e"} Mar 18 21:00:20 crc kubenswrapper[4950]: I0318 21:00:20.517563 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" podStartSLOduration=4.5175410110000005 podStartE2EDuration="4.517541011s" podCreationTimestamp="2026-03-18 21:00:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 21:00:20.507526938 +0000 UTC m=+3233.748368806" watchObservedRunningTime="2026-03-18 21:00:20.517541011 +0000 UTC m=+3233.758382869" Mar 18 21:00:20 crc kubenswrapper[4950]: I0318 21:00:20.534599 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=4.534578568 podStartE2EDuration="4.534578568s" podCreationTimestamp="2026-03-18 21:00:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 21:00:20.534165217 +0000 UTC m=+3233.775007085" watchObservedRunningTime="2026-03-18 21:00:20.534578568 +0000 UTC m=+3233.775420436" Mar 18 21:00:21 crc kubenswrapper[4950]: I0318 21:00:21.531238 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"fcb05ff2-359e-4728-a148-b8f8678e4ce8","Type":"ContainerStarted","Data":"36c82e8c490fb8f94c4df1767d29546a61825206aa668c3a2a92b4d3b5da6be8"} Mar 18 21:00:21 crc kubenswrapper[4950]: I0318 21:00:21.531887 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="ab176318-8186-41e2-8277-1d7b4d4f7618" containerName="manila-api" containerID="cri-o://3165fe6d9e468847f34b57eac9432e3c988f8d350e96d51224903e78665a4694" gracePeriod=30 Mar 18 21:00:21 crc kubenswrapper[4950]: I0318 21:00:21.531618 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="ab176318-8186-41e2-8277-1d7b4d4f7618" containerName="manila-api-log" containerID="cri-o://abe31e08ae8889a6500d3527ed7dfdc24654d4f161ce78309046281901aaefef" gracePeriod=30 Mar 18 21:00:21 crc kubenswrapper[4950]: I0318 21:00:21.593662 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=4.208058244 podStartE2EDuration="5.59364692s" podCreationTimestamp="2026-03-18 21:00:16 +0000 UTC" firstStartedPulling="2026-03-18 21:00:18.139521894 +0000 UTC m=+3231.380363762" lastFinishedPulling="2026-03-18 21:00:19.52511057 +0000 UTC m=+3232.765952438" observedRunningTime="2026-03-18 21:00:21.588878854 +0000 UTC m=+3234.829720722" watchObservedRunningTime="2026-03-18 21:00:21.59364692 +0000 UTC m=+3234.834488788" Mar 18 21:00:22 crc kubenswrapper[4950]: I0318 21:00:22.592586 4950 generic.go:334] "Generic (PLEG): container finished" podID="ab176318-8186-41e2-8277-1d7b4d4f7618" containerID="3165fe6d9e468847f34b57eac9432e3c988f8d350e96d51224903e78665a4694" exitCode=0 Mar 18 21:00:22 crc kubenswrapper[4950]: I0318 21:00:22.593242 4950 generic.go:334] "Generic (PLEG): container finished" podID="ab176318-8186-41e2-8277-1d7b4d4f7618" containerID="abe31e08ae8889a6500d3527ed7dfdc24654d4f161ce78309046281901aaefef" exitCode=143 Mar 18 21:00:22 crc kubenswrapper[4950]: I0318 21:00:22.592777 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"ab176318-8186-41e2-8277-1d7b4d4f7618","Type":"ContainerDied","Data":"3165fe6d9e468847f34b57eac9432e3c988f8d350e96d51224903e78665a4694"} Mar 18 21:00:22 crc kubenswrapper[4950]: I0318 21:00:22.594117 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"ab176318-8186-41e2-8277-1d7b4d4f7618","Type":"ContainerDied","Data":"abe31e08ae8889a6500d3527ed7dfdc24654d4f161ce78309046281901aaefef"} Mar 18 21:00:22 crc kubenswrapper[4950]: I0318 21:00:22.837660 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Mar 18 21:00:22 crc kubenswrapper[4950]: I0318 21:00:22.942123 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ab176318-8186-41e2-8277-1d7b4d4f7618-etc-machine-id\") pod \"ab176318-8186-41e2-8277-1d7b4d4f7618\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " Mar 18 21:00:22 crc kubenswrapper[4950]: I0318 21:00:22.942310 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-scripts\") pod \"ab176318-8186-41e2-8277-1d7b4d4f7618\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " Mar 18 21:00:22 crc kubenswrapper[4950]: I0318 21:00:22.942386 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-combined-ca-bundle\") pod \"ab176318-8186-41e2-8277-1d7b4d4f7618\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " Mar 18 21:00:22 crc kubenswrapper[4950]: I0318 21:00:22.942472 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w685v\" (UniqueName: \"kubernetes.io/projected/ab176318-8186-41e2-8277-1d7b4d4f7618-kube-api-access-w685v\") pod \"ab176318-8186-41e2-8277-1d7b4d4f7618\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " Mar 18 21:00:22 crc kubenswrapper[4950]: I0318 21:00:22.942508 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-config-data\") pod \"ab176318-8186-41e2-8277-1d7b4d4f7618\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " Mar 18 21:00:22 crc kubenswrapper[4950]: I0318 21:00:22.942503 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab176318-8186-41e2-8277-1d7b4d4f7618-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ab176318-8186-41e2-8277-1d7b4d4f7618" (UID: "ab176318-8186-41e2-8277-1d7b4d4f7618"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 21:00:22 crc kubenswrapper[4950]: I0318 21:00:22.942527 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-config-data-custom\") pod \"ab176318-8186-41e2-8277-1d7b4d4f7618\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " Mar 18 21:00:22 crc kubenswrapper[4950]: I0318 21:00:22.942589 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab176318-8186-41e2-8277-1d7b4d4f7618-logs\") pod \"ab176318-8186-41e2-8277-1d7b4d4f7618\" (UID: \"ab176318-8186-41e2-8277-1d7b4d4f7618\") " Mar 18 21:00:22 crc kubenswrapper[4950]: I0318 21:00:22.943512 4950 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ab176318-8186-41e2-8277-1d7b4d4f7618-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:22 crc kubenswrapper[4950]: I0318 21:00:22.943778 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab176318-8186-41e2-8277-1d7b4d4f7618-logs" (OuterVolumeSpecName: "logs") pod "ab176318-8186-41e2-8277-1d7b4d4f7618" (UID: "ab176318-8186-41e2-8277-1d7b4d4f7618"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:00:22 crc kubenswrapper[4950]: I0318 21:00:22.952623 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-scripts" (OuterVolumeSpecName: "scripts") pod "ab176318-8186-41e2-8277-1d7b4d4f7618" (UID: "ab176318-8186-41e2-8277-1d7b4d4f7618"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:22 crc kubenswrapper[4950]: I0318 21:00:22.974681 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ab176318-8186-41e2-8277-1d7b4d4f7618" (UID: "ab176318-8186-41e2-8277-1d7b4d4f7618"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:22 crc kubenswrapper[4950]: I0318 21:00:22.974780 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab176318-8186-41e2-8277-1d7b4d4f7618-kube-api-access-w685v" (OuterVolumeSpecName: "kube-api-access-w685v") pod "ab176318-8186-41e2-8277-1d7b4d4f7618" (UID: "ab176318-8186-41e2-8277-1d7b4d4f7618"). InnerVolumeSpecName "kube-api-access-w685v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.010636 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab176318-8186-41e2-8277-1d7b4d4f7618" (UID: "ab176318-8186-41e2-8277-1d7b4d4f7618"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.046889 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.046919 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.046927 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w685v\" (UniqueName: \"kubernetes.io/projected/ab176318-8186-41e2-8277-1d7b4d4f7618-kube-api-access-w685v\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.046938 4950 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.046947 4950 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab176318-8186-41e2-8277-1d7b4d4f7618-logs\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.088607 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-config-data" (OuterVolumeSpecName: "config-data") pod "ab176318-8186-41e2-8277-1d7b4d4f7618" (UID: "ab176318-8186-41e2-8277-1d7b4d4f7618"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.149807 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab176318-8186-41e2-8277-1d7b4d4f7618-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.618805 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"ab176318-8186-41e2-8277-1d7b4d4f7618","Type":"ContainerDied","Data":"5d40b272004a94b01f6bf929379d39a839d989329f4cc838e0cbf7bda19e4c69"} Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.618871 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.618889 4950 scope.go:117] "RemoveContainer" containerID="3165fe6d9e468847f34b57eac9432e3c988f8d350e96d51224903e78665a4694" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.652494 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.661056 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.676419 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Mar 18 21:00:23 crc kubenswrapper[4950]: E0318 21:00:23.676799 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab176318-8186-41e2-8277-1d7b4d4f7618" containerName="manila-api-log" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.676814 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab176318-8186-41e2-8277-1d7b4d4f7618" containerName="manila-api-log" Mar 18 21:00:23 crc kubenswrapper[4950]: E0318 21:00:23.676831 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab176318-8186-41e2-8277-1d7b4d4f7618" containerName="manila-api" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.676837 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab176318-8186-41e2-8277-1d7b4d4f7618" containerName="manila-api" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.676942 4950 scope.go:117] "RemoveContainer" containerID="abe31e08ae8889a6500d3527ed7dfdc24654d4f161ce78309046281901aaefef" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.677008 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab176318-8186-41e2-8277-1d7b4d4f7618" containerName="manila-api" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.677036 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab176318-8186-41e2-8277-1d7b4d4f7618" containerName="manila-api-log" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.678093 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.681499 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.681894 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.682134 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.695438 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.873688 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f52ad06-7e1b-4736-9441-9b1749f6d325-config-data-custom\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.873742 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9f52ad06-7e1b-4736-9441-9b1749f6d325-etc-machine-id\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.873773 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f52ad06-7e1b-4736-9441-9b1749f6d325-internal-tls-certs\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.873804 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f52ad06-7e1b-4736-9441-9b1749f6d325-public-tls-certs\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.873834 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f52ad06-7e1b-4736-9441-9b1749f6d325-logs\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.873854 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f52ad06-7e1b-4736-9441-9b1749f6d325-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.873874 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f52ad06-7e1b-4736-9441-9b1749f6d325-scripts\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.873923 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f52ad06-7e1b-4736-9441-9b1749f6d325-config-data\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.873946 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpjn4\" (UniqueName: \"kubernetes.io/projected/9f52ad06-7e1b-4736-9441-9b1749f6d325-kube-api-access-mpjn4\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.976202 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f52ad06-7e1b-4736-9441-9b1749f6d325-config-data-custom\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.976258 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9f52ad06-7e1b-4736-9441-9b1749f6d325-etc-machine-id\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.976286 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f52ad06-7e1b-4736-9441-9b1749f6d325-internal-tls-certs\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.976319 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f52ad06-7e1b-4736-9441-9b1749f6d325-public-tls-certs\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.976346 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f52ad06-7e1b-4736-9441-9b1749f6d325-logs\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.976360 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f52ad06-7e1b-4736-9441-9b1749f6d325-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.976379 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f52ad06-7e1b-4736-9441-9b1749f6d325-scripts\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.976459 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f52ad06-7e1b-4736-9441-9b1749f6d325-config-data\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.976492 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpjn4\" (UniqueName: \"kubernetes.io/projected/9f52ad06-7e1b-4736-9441-9b1749f6d325-kube-api-access-mpjn4\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.977992 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f52ad06-7e1b-4736-9441-9b1749f6d325-logs\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.978292 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9f52ad06-7e1b-4736-9441-9b1749f6d325-etc-machine-id\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.985381 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f52ad06-7e1b-4736-9441-9b1749f6d325-public-tls-certs\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.985664 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f52ad06-7e1b-4736-9441-9b1749f6d325-internal-tls-certs\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.985705 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f52ad06-7e1b-4736-9441-9b1749f6d325-scripts\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.985722 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f52ad06-7e1b-4736-9441-9b1749f6d325-config-data\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.993177 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f52ad06-7e1b-4736-9441-9b1749f6d325-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:23 crc kubenswrapper[4950]: I0318 21:00:23.993614 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9f52ad06-7e1b-4736-9441-9b1749f6d325-config-data-custom\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:24 crc kubenswrapper[4950]: I0318 21:00:24.012691 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpjn4\" (UniqueName: \"kubernetes.io/projected/9f52ad06-7e1b-4736-9441-9b1749f6d325-kube-api-access-mpjn4\") pod \"manila-api-0\" (UID: \"9f52ad06-7e1b-4736-9441-9b1749f6d325\") " pod="openstack/manila-api-0" Mar 18 21:00:24 crc kubenswrapper[4950]: I0318 21:00:24.033673 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Mar 18 21:00:24 crc kubenswrapper[4950]: I0318 21:00:24.892098 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Mar 18 21:00:25 crc kubenswrapper[4950]: I0318 21:00:25.505801 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab176318-8186-41e2-8277-1d7b4d4f7618" path="/var/lib/kubelet/pods/ab176318-8186-41e2-8277-1d7b4d4f7618/volumes" Mar 18 21:00:25 crc kubenswrapper[4950]: I0318 21:00:25.670097 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"9f52ad06-7e1b-4736-9441-9b1749f6d325","Type":"ContainerStarted","Data":"a19f363388bc66b47286fd4ab46d5ca890d6073dd0d89222fb669e1474558bcd"} Mar 18 21:00:25 crc kubenswrapper[4950]: I0318 21:00:25.670520 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"9f52ad06-7e1b-4736-9441-9b1749f6d325","Type":"ContainerStarted","Data":"29d7ca6e824dfffacfd5613b611f67fa2d3a3ccfc50ba6db764f4da59f7e8b4a"} Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.564010 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-848988ddb5-9t8cx" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.631190 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-659bf66fb7-qw9s8" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.646547 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/746eaa13-6098-490f-9ada-f8338fe419be-config-data\") pod \"746eaa13-6098-490f-9ada-f8338fe419be\" (UID: \"746eaa13-6098-490f-9ada-f8338fe419be\") " Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.646651 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-horizon-secret-key\") pod \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\" (UID: \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\") " Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.646689 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-config-data\") pod \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\" (UID: \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\") " Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.646774 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7tx5\" (UniqueName: \"kubernetes.io/projected/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-kube-api-access-r7tx5\") pod \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\" (UID: \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\") " Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.646835 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgzf5\" (UniqueName: \"kubernetes.io/projected/746eaa13-6098-490f-9ada-f8338fe419be-kube-api-access-sgzf5\") pod \"746eaa13-6098-490f-9ada-f8338fe419be\" (UID: \"746eaa13-6098-490f-9ada-f8338fe419be\") " Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.646906 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-logs\") pod \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\" (UID: \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\") " Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.646941 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/746eaa13-6098-490f-9ada-f8338fe419be-horizon-secret-key\") pod \"746eaa13-6098-490f-9ada-f8338fe419be\" (UID: \"746eaa13-6098-490f-9ada-f8338fe419be\") " Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.646955 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/746eaa13-6098-490f-9ada-f8338fe419be-logs\") pod \"746eaa13-6098-490f-9ada-f8338fe419be\" (UID: \"746eaa13-6098-490f-9ada-f8338fe419be\") " Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.646981 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-scripts\") pod \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\" (UID: \"d592b2fb-1e76-410f-8b0d-a48bc85b8b58\") " Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.647025 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/746eaa13-6098-490f-9ada-f8338fe419be-scripts\") pod \"746eaa13-6098-490f-9ada-f8338fe419be\" (UID: \"746eaa13-6098-490f-9ada-f8338fe419be\") " Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.661918 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d592b2fb-1e76-410f-8b0d-a48bc85b8b58" (UID: "d592b2fb-1e76-410f-8b0d-a48bc85b8b58"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.671663 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-logs" (OuterVolumeSpecName: "logs") pod "d592b2fb-1e76-410f-8b0d-a48bc85b8b58" (UID: "d592b2fb-1e76-410f-8b0d-a48bc85b8b58"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.671973 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/746eaa13-6098-490f-9ada-f8338fe419be-logs" (OuterVolumeSpecName: "logs") pod "746eaa13-6098-490f-9ada-f8338fe419be" (UID: "746eaa13-6098-490f-9ada-f8338fe419be"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.686286 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-kube-api-access-r7tx5" (OuterVolumeSpecName: "kube-api-access-r7tx5") pod "d592b2fb-1e76-410f-8b0d-a48bc85b8b58" (UID: "d592b2fb-1e76-410f-8b0d-a48bc85b8b58"). InnerVolumeSpecName "kube-api-access-r7tx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.687015 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/746eaa13-6098-490f-9ada-f8338fe419be-kube-api-access-sgzf5" (OuterVolumeSpecName: "kube-api-access-sgzf5") pod "746eaa13-6098-490f-9ada-f8338fe419be" (UID: "746eaa13-6098-490f-9ada-f8338fe419be"). InnerVolumeSpecName "kube-api-access-sgzf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.688750 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/746eaa13-6098-490f-9ada-f8338fe419be-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "746eaa13-6098-490f-9ada-f8338fe419be" (UID: "746eaa13-6098-490f-9ada-f8338fe419be"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.699299 4950 generic.go:334] "Generic (PLEG): container finished" podID="746eaa13-6098-490f-9ada-f8338fe419be" containerID="7fa9cd942f4e0076f6010f77d48d9fc39d501bdcc0be17307c0be365c614ca53" exitCode=137 Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.699339 4950 generic.go:334] "Generic (PLEG): container finished" podID="746eaa13-6098-490f-9ada-f8338fe419be" containerID="e0003fa0675dfc576e0bb2f940084a7e9e5ddb8f132eb87eb5ac0c9eef09e779" exitCode=137 Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.699385 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-659bf66fb7-qw9s8" event={"ID":"746eaa13-6098-490f-9ada-f8338fe419be","Type":"ContainerDied","Data":"7fa9cd942f4e0076f6010f77d48d9fc39d501bdcc0be17307c0be365c614ca53"} Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.699566 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-659bf66fb7-qw9s8" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.699983 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-659bf66fb7-qw9s8" event={"ID":"746eaa13-6098-490f-9ada-f8338fe419be","Type":"ContainerDied","Data":"e0003fa0675dfc576e0bb2f940084a7e9e5ddb8f132eb87eb5ac0c9eef09e779"} Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.700013 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-659bf66fb7-qw9s8" event={"ID":"746eaa13-6098-490f-9ada-f8338fe419be","Type":"ContainerDied","Data":"88a48b23d683ef643ac9af0eb09e936825ca0a4a82641d233b8bd1fe37cb3979"} Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.700035 4950 scope.go:117] "RemoveContainer" containerID="7fa9cd942f4e0076f6010f77d48d9fc39d501bdcc0be17307c0be365c614ca53" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.707026 4950 generic.go:334] "Generic (PLEG): container finished" podID="d592b2fb-1e76-410f-8b0d-a48bc85b8b58" containerID="a8f9756eb93bf64f7b97cc373b0fd7a76096c3a6b8113ba1e0a0c9081d54251e" exitCode=137 Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.707059 4950 generic.go:334] "Generic (PLEG): container finished" podID="d592b2fb-1e76-410f-8b0d-a48bc85b8b58" containerID="249aabae5cccde84300386f75b0257ea6574925978759cf790ad555443ff5616" exitCode=137 Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.707117 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-848988ddb5-9t8cx" event={"ID":"d592b2fb-1e76-410f-8b0d-a48bc85b8b58","Type":"ContainerDied","Data":"a8f9756eb93bf64f7b97cc373b0fd7a76096c3a6b8113ba1e0a0c9081d54251e"} Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.707144 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-848988ddb5-9t8cx" event={"ID":"d592b2fb-1e76-410f-8b0d-a48bc85b8b58","Type":"ContainerDied","Data":"249aabae5cccde84300386f75b0257ea6574925978759cf790ad555443ff5616"} Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.707155 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-848988ddb5-9t8cx" event={"ID":"d592b2fb-1e76-410f-8b0d-a48bc85b8b58","Type":"ContainerDied","Data":"48b0938e7e2dcfc63891667ff8ebb4b9e1a5ddd5aa6417abb99fde5118c0b9e2"} Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.707296 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-848988ddb5-9t8cx" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.712656 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"9f52ad06-7e1b-4736-9441-9b1749f6d325","Type":"ContainerStarted","Data":"6b6c405d03947c6a437f34cbaac008c67c4cf8e4c3e3901678cb667423a00a89"} Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.713181 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.717866 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/746eaa13-6098-490f-9ada-f8338fe419be-config-data" (OuterVolumeSpecName: "config-data") pod "746eaa13-6098-490f-9ada-f8338fe419be" (UID: "746eaa13-6098-490f-9ada-f8338fe419be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.721254 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-scripts" (OuterVolumeSpecName: "scripts") pod "d592b2fb-1e76-410f-8b0d-a48bc85b8b58" (UID: "d592b2fb-1e76-410f-8b0d-a48bc85b8b58"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.734510 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-config-data" (OuterVolumeSpecName: "config-data") pod "d592b2fb-1e76-410f-8b0d-a48bc85b8b58" (UID: "d592b2fb-1e76-410f-8b0d-a48bc85b8b58"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.749988 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.750046 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7tx5\" (UniqueName: \"kubernetes.io/projected/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-kube-api-access-r7tx5\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.750056 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgzf5\" (UniqueName: \"kubernetes.io/projected/746eaa13-6098-490f-9ada-f8338fe419be-kube-api-access-sgzf5\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.750067 4950 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-logs\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.750075 4950 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/746eaa13-6098-490f-9ada-f8338fe419be-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.750083 4950 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/746eaa13-6098-490f-9ada-f8338fe419be-logs\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.750091 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.750099 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/746eaa13-6098-490f-9ada-f8338fe419be-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.750108 4950 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d592b2fb-1e76-410f-8b0d-a48bc85b8b58-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.753183 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.75317224 podStartE2EDuration="3.75317224s" podCreationTimestamp="2026-03-18 21:00:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 21:00:26.741078943 +0000 UTC m=+3239.981920811" watchObservedRunningTime="2026-03-18 21:00:26.75317224 +0000 UTC m=+3239.994014108" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.787356 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/746eaa13-6098-490f-9ada-f8338fe419be-scripts" (OuterVolumeSpecName: "scripts") pod "746eaa13-6098-490f-9ada-f8338fe419be" (UID: "746eaa13-6098-490f-9ada-f8338fe419be"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.851445 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/746eaa13-6098-490f-9ada-f8338fe419be-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.924810 4950 scope.go:117] "RemoveContainer" containerID="e0003fa0675dfc576e0bb2f940084a7e9e5ddb8f132eb87eb5ac0c9eef09e779" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.954654 4950 scope.go:117] "RemoveContainer" containerID="7fa9cd942f4e0076f6010f77d48d9fc39d501bdcc0be17307c0be365c614ca53" Mar 18 21:00:26 crc kubenswrapper[4950]: E0318 21:00:26.957404 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fa9cd942f4e0076f6010f77d48d9fc39d501bdcc0be17307c0be365c614ca53\": container with ID starting with 7fa9cd942f4e0076f6010f77d48d9fc39d501bdcc0be17307c0be365c614ca53 not found: ID does not exist" containerID="7fa9cd942f4e0076f6010f77d48d9fc39d501bdcc0be17307c0be365c614ca53" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.957482 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fa9cd942f4e0076f6010f77d48d9fc39d501bdcc0be17307c0be365c614ca53"} err="failed to get container status \"7fa9cd942f4e0076f6010f77d48d9fc39d501bdcc0be17307c0be365c614ca53\": rpc error: code = NotFound desc = could not find container \"7fa9cd942f4e0076f6010f77d48d9fc39d501bdcc0be17307c0be365c614ca53\": container with ID starting with 7fa9cd942f4e0076f6010f77d48d9fc39d501bdcc0be17307c0be365c614ca53 not found: ID does not exist" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.957507 4950 scope.go:117] "RemoveContainer" containerID="e0003fa0675dfc576e0bb2f940084a7e9e5ddb8f132eb87eb5ac0c9eef09e779" Mar 18 21:00:26 crc kubenswrapper[4950]: E0318 21:00:26.958502 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0003fa0675dfc576e0bb2f940084a7e9e5ddb8f132eb87eb5ac0c9eef09e779\": container with ID starting with e0003fa0675dfc576e0bb2f940084a7e9e5ddb8f132eb87eb5ac0c9eef09e779 not found: ID does not exist" containerID="e0003fa0675dfc576e0bb2f940084a7e9e5ddb8f132eb87eb5ac0c9eef09e779" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.958529 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0003fa0675dfc576e0bb2f940084a7e9e5ddb8f132eb87eb5ac0c9eef09e779"} err="failed to get container status \"e0003fa0675dfc576e0bb2f940084a7e9e5ddb8f132eb87eb5ac0c9eef09e779\": rpc error: code = NotFound desc = could not find container \"e0003fa0675dfc576e0bb2f940084a7e9e5ddb8f132eb87eb5ac0c9eef09e779\": container with ID starting with e0003fa0675dfc576e0bb2f940084a7e9e5ddb8f132eb87eb5ac0c9eef09e779 not found: ID does not exist" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.958545 4950 scope.go:117] "RemoveContainer" containerID="7fa9cd942f4e0076f6010f77d48d9fc39d501bdcc0be17307c0be365c614ca53" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.959198 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fa9cd942f4e0076f6010f77d48d9fc39d501bdcc0be17307c0be365c614ca53"} err="failed to get container status \"7fa9cd942f4e0076f6010f77d48d9fc39d501bdcc0be17307c0be365c614ca53\": rpc error: code = NotFound desc = could not find container \"7fa9cd942f4e0076f6010f77d48d9fc39d501bdcc0be17307c0be365c614ca53\": container with ID starting with 7fa9cd942f4e0076f6010f77d48d9fc39d501bdcc0be17307c0be365c614ca53 not found: ID does not exist" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.959225 4950 scope.go:117] "RemoveContainer" containerID="e0003fa0675dfc576e0bb2f940084a7e9e5ddb8f132eb87eb5ac0c9eef09e779" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.960054 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0003fa0675dfc576e0bb2f940084a7e9e5ddb8f132eb87eb5ac0c9eef09e779"} err="failed to get container status \"e0003fa0675dfc576e0bb2f940084a7e9e5ddb8f132eb87eb5ac0c9eef09e779\": rpc error: code = NotFound desc = could not find container \"e0003fa0675dfc576e0bb2f940084a7e9e5ddb8f132eb87eb5ac0c9eef09e779\": container with ID starting with e0003fa0675dfc576e0bb2f940084a7e9e5ddb8f132eb87eb5ac0c9eef09e779 not found: ID does not exist" Mar 18 21:00:26 crc kubenswrapper[4950]: I0318 21:00:26.960076 4950 scope.go:117] "RemoveContainer" containerID="a8f9756eb93bf64f7b97cc373b0fd7a76096c3a6b8113ba1e0a0c9081d54251e" Mar 18 21:00:27 crc kubenswrapper[4950]: I0318 21:00:27.057889 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-659bf66fb7-qw9s8"] Mar 18 21:00:27 crc kubenswrapper[4950]: I0318 21:00:27.070872 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-659bf66fb7-qw9s8"] Mar 18 21:00:27 crc kubenswrapper[4950]: I0318 21:00:27.079051 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-848988ddb5-9t8cx"] Mar 18 21:00:27 crc kubenswrapper[4950]: I0318 21:00:27.089896 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-848988ddb5-9t8cx"] Mar 18 21:00:27 crc kubenswrapper[4950]: I0318 21:00:27.131604 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7f55f6cb75-46f5l" Mar 18 21:00:27 crc kubenswrapper[4950]: I0318 21:00:27.224436 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84f6c65c87-mmngf"] Mar 18 21:00:27 crc kubenswrapper[4950]: I0318 21:00:27.224638 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" podUID="74de5806-1889-472c-b423-dfbdc2a63020" containerName="dnsmasq-dns" containerID="cri-o://6ebacc2249dba5b2f473a4da950e2dd3ee04679510a00ffe911f4046844db248" gracePeriod=10 Mar 18 21:00:27 crc kubenswrapper[4950]: I0318 21:00:27.261160 4950 scope.go:117] "RemoveContainer" containerID="249aabae5cccde84300386f75b0257ea6574925978759cf790ad555443ff5616" Mar 18 21:00:27 crc kubenswrapper[4950]: I0318 21:00:27.261297 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Mar 18 21:00:27 crc kubenswrapper[4950]: I0318 21:00:27.305071 4950 scope.go:117] "RemoveContainer" containerID="a8f9756eb93bf64f7b97cc373b0fd7a76096c3a6b8113ba1e0a0c9081d54251e" Mar 18 21:00:27 crc kubenswrapper[4950]: E0318 21:00:27.306600 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8f9756eb93bf64f7b97cc373b0fd7a76096c3a6b8113ba1e0a0c9081d54251e\": container with ID starting with a8f9756eb93bf64f7b97cc373b0fd7a76096c3a6b8113ba1e0a0c9081d54251e not found: ID does not exist" containerID="a8f9756eb93bf64f7b97cc373b0fd7a76096c3a6b8113ba1e0a0c9081d54251e" Mar 18 21:00:27 crc kubenswrapper[4950]: I0318 21:00:27.306651 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8f9756eb93bf64f7b97cc373b0fd7a76096c3a6b8113ba1e0a0c9081d54251e"} err="failed to get container status \"a8f9756eb93bf64f7b97cc373b0fd7a76096c3a6b8113ba1e0a0c9081d54251e\": rpc error: code = NotFound desc = could not find container \"a8f9756eb93bf64f7b97cc373b0fd7a76096c3a6b8113ba1e0a0c9081d54251e\": container with ID starting with a8f9756eb93bf64f7b97cc373b0fd7a76096c3a6b8113ba1e0a0c9081d54251e not found: ID does not exist" Mar 18 21:00:27 crc kubenswrapper[4950]: I0318 21:00:27.306673 4950 scope.go:117] "RemoveContainer" containerID="249aabae5cccde84300386f75b0257ea6574925978759cf790ad555443ff5616" Mar 18 21:00:27 crc kubenswrapper[4950]: E0318 21:00:27.308590 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"249aabae5cccde84300386f75b0257ea6574925978759cf790ad555443ff5616\": container with ID starting with 249aabae5cccde84300386f75b0257ea6574925978759cf790ad555443ff5616 not found: ID does not exist" containerID="249aabae5cccde84300386f75b0257ea6574925978759cf790ad555443ff5616" Mar 18 21:00:27 crc kubenswrapper[4950]: I0318 21:00:27.308635 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"249aabae5cccde84300386f75b0257ea6574925978759cf790ad555443ff5616"} err="failed to get container status \"249aabae5cccde84300386f75b0257ea6574925978759cf790ad555443ff5616\": rpc error: code = NotFound desc = could not find container \"249aabae5cccde84300386f75b0257ea6574925978759cf790ad555443ff5616\": container with ID starting with 249aabae5cccde84300386f75b0257ea6574925978759cf790ad555443ff5616 not found: ID does not exist" Mar 18 21:00:27 crc kubenswrapper[4950]: I0318 21:00:27.308649 4950 scope.go:117] "RemoveContainer" containerID="a8f9756eb93bf64f7b97cc373b0fd7a76096c3a6b8113ba1e0a0c9081d54251e" Mar 18 21:00:27 crc kubenswrapper[4950]: I0318 21:00:27.309053 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8f9756eb93bf64f7b97cc373b0fd7a76096c3a6b8113ba1e0a0c9081d54251e"} err="failed to get container status \"a8f9756eb93bf64f7b97cc373b0fd7a76096c3a6b8113ba1e0a0c9081d54251e\": rpc error: code = NotFound desc = could not find container \"a8f9756eb93bf64f7b97cc373b0fd7a76096c3a6b8113ba1e0a0c9081d54251e\": container with ID starting with a8f9756eb93bf64f7b97cc373b0fd7a76096c3a6b8113ba1e0a0c9081d54251e not found: ID does not exist" Mar 18 21:00:27 crc kubenswrapper[4950]: I0318 21:00:27.309074 4950 scope.go:117] "RemoveContainer" containerID="249aabae5cccde84300386f75b0257ea6574925978759cf790ad555443ff5616" Mar 18 21:00:27 crc kubenswrapper[4950]: I0318 21:00:27.309962 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"249aabae5cccde84300386f75b0257ea6574925978759cf790ad555443ff5616"} err="failed to get container status \"249aabae5cccde84300386f75b0257ea6574925978759cf790ad555443ff5616\": rpc error: code = NotFound desc = could not find container \"249aabae5cccde84300386f75b0257ea6574925978759cf790ad555443ff5616\": container with ID starting with 249aabae5cccde84300386f75b0257ea6574925978759cf790ad555443ff5616 not found: ID does not exist" Mar 18 21:00:27 crc kubenswrapper[4950]: I0318 21:00:27.527866 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="746eaa13-6098-490f-9ada-f8338fe419be" path="/var/lib/kubelet/pods/746eaa13-6098-490f-9ada-f8338fe419be/volumes" Mar 18 21:00:27 crc kubenswrapper[4950]: I0318 21:00:27.528613 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d592b2fb-1e76-410f-8b0d-a48bc85b8b58" path="/var/lib/kubelet/pods/d592b2fb-1e76-410f-8b0d-a48bc85b8b58/volumes" Mar 18 21:00:27 crc kubenswrapper[4950]: I0318 21:00:27.754201 4950 generic.go:334] "Generic (PLEG): container finished" podID="74de5806-1889-472c-b423-dfbdc2a63020" containerID="6ebacc2249dba5b2f473a4da950e2dd3ee04679510a00ffe911f4046844db248" exitCode=0 Mar 18 21:00:27 crc kubenswrapper[4950]: I0318 21:00:27.754286 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" event={"ID":"74de5806-1889-472c-b423-dfbdc2a63020","Type":"ContainerDied","Data":"6ebacc2249dba5b2f473a4da950e2dd3ee04679510a00ffe911f4046844db248"} Mar 18 21:00:27 crc kubenswrapper[4950]: I0318 21:00:27.925357 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 21:00:28 crc kubenswrapper[4950]: I0318 21:00:28.119989 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzbxj\" (UniqueName: \"kubernetes.io/projected/74de5806-1889-472c-b423-dfbdc2a63020-kube-api-access-hzbxj\") pod \"74de5806-1889-472c-b423-dfbdc2a63020\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " Mar 18 21:00:28 crc kubenswrapper[4950]: I0318 21:00:28.120103 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-openstack-edpm-ipam\") pod \"74de5806-1889-472c-b423-dfbdc2a63020\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " Mar 18 21:00:28 crc kubenswrapper[4950]: I0318 21:00:28.120950 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-config\") pod \"74de5806-1889-472c-b423-dfbdc2a63020\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " Mar 18 21:00:28 crc kubenswrapper[4950]: I0318 21:00:28.120990 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-ovsdbserver-nb\") pod \"74de5806-1889-472c-b423-dfbdc2a63020\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " Mar 18 21:00:28 crc kubenswrapper[4950]: I0318 21:00:28.121019 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-dns-svc\") pod \"74de5806-1889-472c-b423-dfbdc2a63020\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " Mar 18 21:00:28 crc kubenswrapper[4950]: I0318 21:00:28.121190 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-ovsdbserver-sb\") pod \"74de5806-1889-472c-b423-dfbdc2a63020\" (UID: \"74de5806-1889-472c-b423-dfbdc2a63020\") " Mar 18 21:00:28 crc kubenswrapper[4950]: I0318 21:00:28.129626 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74de5806-1889-472c-b423-dfbdc2a63020-kube-api-access-hzbxj" (OuterVolumeSpecName: "kube-api-access-hzbxj") pod "74de5806-1889-472c-b423-dfbdc2a63020" (UID: "74de5806-1889-472c-b423-dfbdc2a63020"). InnerVolumeSpecName "kube-api-access-hzbxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:00:28 crc kubenswrapper[4950]: I0318 21:00:28.177696 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "74de5806-1889-472c-b423-dfbdc2a63020" (UID: "74de5806-1889-472c-b423-dfbdc2a63020"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 21:00:28 crc kubenswrapper[4950]: I0318 21:00:28.186554 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "74de5806-1889-472c-b423-dfbdc2a63020" (UID: "74de5806-1889-472c-b423-dfbdc2a63020"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 21:00:28 crc kubenswrapper[4950]: I0318 21:00:28.194145 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "74de5806-1889-472c-b423-dfbdc2a63020" (UID: "74de5806-1889-472c-b423-dfbdc2a63020"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 21:00:28 crc kubenswrapper[4950]: I0318 21:00:28.223302 4950 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:28 crc kubenswrapper[4950]: I0318 21:00:28.223333 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzbxj\" (UniqueName: \"kubernetes.io/projected/74de5806-1889-472c-b423-dfbdc2a63020-kube-api-access-hzbxj\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:28 crc kubenswrapper[4950]: I0318 21:00:28.223344 4950 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:28 crc kubenswrapper[4950]: I0318 21:00:28.223357 4950 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:28 crc kubenswrapper[4950]: I0318 21:00:28.239614 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "74de5806-1889-472c-b423-dfbdc2a63020" (UID: "74de5806-1889-472c-b423-dfbdc2a63020"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 21:00:28 crc kubenswrapper[4950]: I0318 21:00:28.290692 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-config" (OuterVolumeSpecName: "config") pod "74de5806-1889-472c-b423-dfbdc2a63020" (UID: "74de5806-1889-472c-b423-dfbdc2a63020"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 21:00:28 crc kubenswrapper[4950]: I0318 21:00:28.324595 4950 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-config\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:28 crc kubenswrapper[4950]: I0318 21:00:28.324632 4950 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74de5806-1889-472c-b423-dfbdc2a63020-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:28 crc kubenswrapper[4950]: I0318 21:00:28.770086 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" event={"ID":"74de5806-1889-472c-b423-dfbdc2a63020","Type":"ContainerDied","Data":"c217b96a0bde9026a07c7184e2b4075f8923056f16782388425c19c5532aaab3"} Mar 18 21:00:28 crc kubenswrapper[4950]: I0318 21:00:28.770137 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84f6c65c87-mmngf" Mar 18 21:00:28 crc kubenswrapper[4950]: I0318 21:00:28.770141 4950 scope.go:117] "RemoveContainer" containerID="6ebacc2249dba5b2f473a4da950e2dd3ee04679510a00ffe911f4046844db248" Mar 18 21:00:28 crc kubenswrapper[4950]: I0318 21:00:28.826044 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84f6c65c87-mmngf"] Mar 18 21:00:28 crc kubenswrapper[4950]: I0318 21:00:28.843583 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84f6c65c87-mmngf"] Mar 18 21:00:29 crc kubenswrapper[4950]: I0318 21:00:29.491901 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74de5806-1889-472c-b423-dfbdc2a63020" path="/var/lib/kubelet/pods/74de5806-1889-472c-b423-dfbdc2a63020/volumes" Mar 18 21:00:30 crc kubenswrapper[4950]: I0318 21:00:30.339860 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 21:00:30 crc kubenswrapper[4950]: I0318 21:00:30.341855 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7449579958-pdtkd" Mar 18 21:00:30 crc kubenswrapper[4950]: I0318 21:00:30.424206 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fkdtp" podUID="73b8c337-25a3-4f8a-88c9-1e24e96589b4" containerName="registry-server" probeResult="failure" output=< Mar 18 21:00:30 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 21:00:30 crc kubenswrapper[4950]: > Mar 18 21:00:31 crc kubenswrapper[4950]: I0318 21:00:31.497223 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 18 21:00:31 crc kubenswrapper[4950]: I0318 21:00:31.497507 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9703e257-1e49-46e0-bbdf-754fee621b4b" containerName="ceilometer-central-agent" containerID="cri-o://6ac425705d5c6fa63f048c24cc9ed5e70a1c540588ab51d299b9b060a652d2b2" gracePeriod=30 Mar 18 21:00:31 crc kubenswrapper[4950]: I0318 21:00:31.497611 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9703e257-1e49-46e0-bbdf-754fee621b4b" containerName="proxy-httpd" containerID="cri-o://a4960228dd8e0cdea9b485c596699b1a53e857466115f3997b8d6047aa7aecb7" gracePeriod=30 Mar 18 21:00:31 crc kubenswrapper[4950]: I0318 21:00:31.497648 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9703e257-1e49-46e0-bbdf-754fee621b4b" containerName="sg-core" containerID="cri-o://f39bf9b9db435c44a5af5d72d57bfbf948467d00f15a7ae19bd9527a0f06c5c5" gracePeriod=30 Mar 18 21:00:31 crc kubenswrapper[4950]: I0318 21:00:31.497678 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9703e257-1e49-46e0-bbdf-754fee621b4b" containerName="ceilometer-notification-agent" containerID="cri-o://a43c585b19441af8a3824713bad197517a01bc13c492af5db16fd853273c86fc" gracePeriod=30 Mar 18 21:00:31 crc kubenswrapper[4950]: I0318 21:00:31.822292 4950 generic.go:334] "Generic (PLEG): container finished" podID="9703e257-1e49-46e0-bbdf-754fee621b4b" containerID="a4960228dd8e0cdea9b485c596699b1a53e857466115f3997b8d6047aa7aecb7" exitCode=0 Mar 18 21:00:31 crc kubenswrapper[4950]: I0318 21:00:31.822599 4950 generic.go:334] "Generic (PLEG): container finished" podID="9703e257-1e49-46e0-bbdf-754fee621b4b" containerID="f39bf9b9db435c44a5af5d72d57bfbf948467d00f15a7ae19bd9527a0f06c5c5" exitCode=2 Mar 18 21:00:31 crc kubenswrapper[4950]: I0318 21:00:31.822369 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9703e257-1e49-46e0-bbdf-754fee621b4b","Type":"ContainerDied","Data":"a4960228dd8e0cdea9b485c596699b1a53e857466115f3997b8d6047aa7aecb7"} Mar 18 21:00:31 crc kubenswrapper[4950]: I0318 21:00:31.822645 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9703e257-1e49-46e0-bbdf-754fee621b4b","Type":"ContainerDied","Data":"f39bf9b9db435c44a5af5d72d57bfbf948467d00f15a7ae19bd9527a0f06c5c5"} Mar 18 21:00:32 crc kubenswrapper[4950]: I0318 21:00:32.442615 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7449579958-pdtkd" Mar 18 21:00:32 crc kubenswrapper[4950]: I0318 21:00:32.515485 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5bfb97c7d-h56mk"] Mar 18 21:00:32 crc kubenswrapper[4950]: I0318 21:00:32.515718 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5bfb97c7d-h56mk" podUID="26e882de-ea5e-4a22-889d-85c5d4347b1e" containerName="horizon-log" containerID="cri-o://36339553eb5a8810d3a99d963f412a51a7f943a23f9e3e57db6fa61a732993dc" gracePeriod=30 Mar 18 21:00:32 crc kubenswrapper[4950]: I0318 21:00:32.515804 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5bfb97c7d-h56mk" podUID="26e882de-ea5e-4a22-889d-85c5d4347b1e" containerName="horizon" containerID="cri-o://aa50a9c7f3ed240952a5dacf482e2676dcb49617b9f6bddc43a68f1aa9938ed8" gracePeriod=30 Mar 18 21:00:32 crc kubenswrapper[4950]: I0318 21:00:32.535988 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5bfb97c7d-h56mk" podUID="26e882de-ea5e-4a22-889d-85c5d4347b1e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.9:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Mar 18 21:00:32 crc kubenswrapper[4950]: I0318 21:00:32.836749 4950 generic.go:334] "Generic (PLEG): container finished" podID="9703e257-1e49-46e0-bbdf-754fee621b4b" containerID="a43c585b19441af8a3824713bad197517a01bc13c492af5db16fd853273c86fc" exitCode=0 Mar 18 21:00:32 crc kubenswrapper[4950]: I0318 21:00:32.837087 4950 generic.go:334] "Generic (PLEG): container finished" podID="9703e257-1e49-46e0-bbdf-754fee621b4b" containerID="6ac425705d5c6fa63f048c24cc9ed5e70a1c540588ab51d299b9b060a652d2b2" exitCode=0 Mar 18 21:00:32 crc kubenswrapper[4950]: I0318 21:00:32.836826 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9703e257-1e49-46e0-bbdf-754fee621b4b","Type":"ContainerDied","Data":"a43c585b19441af8a3824713bad197517a01bc13c492af5db16fd853273c86fc"} Mar 18 21:00:32 crc kubenswrapper[4950]: I0318 21:00:32.837578 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9703e257-1e49-46e0-bbdf-754fee621b4b","Type":"ContainerDied","Data":"6ac425705d5c6fa63f048c24cc9ed5e70a1c540588ab51d299b9b060a652d2b2"} Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.242719 4950 scope.go:117] "RemoveContainer" containerID="654f0d6c8a9b433b09cd88437a2df12e449a86f79c0b22d19f564e39e30c5d5c" Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.481995 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:00:33 crc kubenswrapper[4950]: E0318 21:00:33.482812 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.655292 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.756236 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-scripts\") pod \"9703e257-1e49-46e0-bbdf-754fee621b4b\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.756314 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-sg-core-conf-yaml\") pod \"9703e257-1e49-46e0-bbdf-754fee621b4b\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.756369 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9703e257-1e49-46e0-bbdf-754fee621b4b-log-httpd\") pod \"9703e257-1e49-46e0-bbdf-754fee621b4b\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.758362 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8g54r\" (UniqueName: \"kubernetes.io/projected/9703e257-1e49-46e0-bbdf-754fee621b4b-kube-api-access-8g54r\") pod \"9703e257-1e49-46e0-bbdf-754fee621b4b\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.758463 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-config-data\") pod \"9703e257-1e49-46e0-bbdf-754fee621b4b\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.758498 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-combined-ca-bundle\") pod \"9703e257-1e49-46e0-bbdf-754fee621b4b\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.758542 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-ceilometer-tls-certs\") pod \"9703e257-1e49-46e0-bbdf-754fee621b4b\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.758559 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9703e257-1e49-46e0-bbdf-754fee621b4b-run-httpd\") pod \"9703e257-1e49-46e0-bbdf-754fee621b4b\" (UID: \"9703e257-1e49-46e0-bbdf-754fee621b4b\") " Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.759148 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9703e257-1e49-46e0-bbdf-754fee621b4b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9703e257-1e49-46e0-bbdf-754fee621b4b" (UID: "9703e257-1e49-46e0-bbdf-754fee621b4b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.760559 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9703e257-1e49-46e0-bbdf-754fee621b4b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9703e257-1e49-46e0-bbdf-754fee621b4b" (UID: "9703e257-1e49-46e0-bbdf-754fee621b4b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.763845 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-scripts" (OuterVolumeSpecName: "scripts") pod "9703e257-1e49-46e0-bbdf-754fee621b4b" (UID: "9703e257-1e49-46e0-bbdf-754fee621b4b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.777055 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9703e257-1e49-46e0-bbdf-754fee621b4b-kube-api-access-8g54r" (OuterVolumeSpecName: "kube-api-access-8g54r") pod "9703e257-1e49-46e0-bbdf-754fee621b4b" (UID: "9703e257-1e49-46e0-bbdf-754fee621b4b"). InnerVolumeSpecName "kube-api-access-8g54r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.795540 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9703e257-1e49-46e0-bbdf-754fee621b4b" (UID: "9703e257-1e49-46e0-bbdf-754fee621b4b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.857561 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9703e257-1e49-46e0-bbdf-754fee621b4b","Type":"ContainerDied","Data":"c17a68aca9578a4302873d6a5b8d8c23d033296270f81cf3681f5b96d28da2e5"} Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.857630 4950 scope.go:117] "RemoveContainer" containerID="a4960228dd8e0cdea9b485c596699b1a53e857466115f3997b8d6047aa7aecb7" Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.858559 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.861687 4950 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9703e257-1e49-46e0-bbdf-754fee621b4b-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.861713 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8g54r\" (UniqueName: \"kubernetes.io/projected/9703e257-1e49-46e0-bbdf-754fee621b4b-kube-api-access-8g54r\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.861722 4950 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9703e257-1e49-46e0-bbdf-754fee621b4b-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.861731 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.861741 4950 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.861979 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "9703e257-1e49-46e0-bbdf-754fee621b4b" (UID: "9703e257-1e49-46e0-bbdf-754fee621b4b"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.884315 4950 scope.go:117] "RemoveContainer" containerID="f39bf9b9db435c44a5af5d72d57bfbf948467d00f15a7ae19bd9527a0f06c5c5" Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.911531 4950 scope.go:117] "RemoveContainer" containerID="a43c585b19441af8a3824713bad197517a01bc13c492af5db16fd853273c86fc" Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.924512 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9703e257-1e49-46e0-bbdf-754fee621b4b" (UID: "9703e257-1e49-46e0-bbdf-754fee621b4b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.939967 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-config-data" (OuterVolumeSpecName: "config-data") pod "9703e257-1e49-46e0-bbdf-754fee621b4b" (UID: "9703e257-1e49-46e0-bbdf-754fee621b4b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.941131 4950 scope.go:117] "RemoveContainer" containerID="6ac425705d5c6fa63f048c24cc9ed5e70a1c540588ab51d299b9b060a652d2b2" Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.964931 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.964969 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:33 crc kubenswrapper[4950]: I0318 21:00:33.964981 4950 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9703e257-1e49-46e0-bbdf-754fee621b4b-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.189459 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.197574 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.225302 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 18 21:00:34 crc kubenswrapper[4950]: E0318 21:00:34.225671 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9703e257-1e49-46e0-bbdf-754fee621b4b" containerName="ceilometer-notification-agent" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.225689 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="9703e257-1e49-46e0-bbdf-754fee621b4b" containerName="ceilometer-notification-agent" Mar 18 21:00:34 crc kubenswrapper[4950]: E0318 21:00:34.225706 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="746eaa13-6098-490f-9ada-f8338fe419be" containerName="horizon" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.225712 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="746eaa13-6098-490f-9ada-f8338fe419be" containerName="horizon" Mar 18 21:00:34 crc kubenswrapper[4950]: E0318 21:00:34.225725 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74de5806-1889-472c-b423-dfbdc2a63020" containerName="dnsmasq-dns" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.225731 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="74de5806-1889-472c-b423-dfbdc2a63020" containerName="dnsmasq-dns" Mar 18 21:00:34 crc kubenswrapper[4950]: E0318 21:00:34.225745 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9703e257-1e49-46e0-bbdf-754fee621b4b" containerName="ceilometer-central-agent" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.225762 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="9703e257-1e49-46e0-bbdf-754fee621b4b" containerName="ceilometer-central-agent" Mar 18 21:00:34 crc kubenswrapper[4950]: E0318 21:00:34.225776 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="746eaa13-6098-490f-9ada-f8338fe419be" containerName="horizon-log" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.225782 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="746eaa13-6098-490f-9ada-f8338fe419be" containerName="horizon-log" Mar 18 21:00:34 crc kubenswrapper[4950]: E0318 21:00:34.225793 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d592b2fb-1e76-410f-8b0d-a48bc85b8b58" containerName="horizon-log" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.225798 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="d592b2fb-1e76-410f-8b0d-a48bc85b8b58" containerName="horizon-log" Mar 18 21:00:34 crc kubenswrapper[4950]: E0318 21:00:34.225806 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9703e257-1e49-46e0-bbdf-754fee621b4b" containerName="proxy-httpd" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.225812 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="9703e257-1e49-46e0-bbdf-754fee621b4b" containerName="proxy-httpd" Mar 18 21:00:34 crc kubenswrapper[4950]: E0318 21:00:34.225818 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74de5806-1889-472c-b423-dfbdc2a63020" containerName="init" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.225826 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="74de5806-1889-472c-b423-dfbdc2a63020" containerName="init" Mar 18 21:00:34 crc kubenswrapper[4950]: E0318 21:00:34.225840 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9703e257-1e49-46e0-bbdf-754fee621b4b" containerName="sg-core" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.225845 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="9703e257-1e49-46e0-bbdf-754fee621b4b" containerName="sg-core" Mar 18 21:00:34 crc kubenswrapper[4950]: E0318 21:00:34.225859 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d592b2fb-1e76-410f-8b0d-a48bc85b8b58" containerName="horizon" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.225866 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="d592b2fb-1e76-410f-8b0d-a48bc85b8b58" containerName="horizon" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.226019 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="d592b2fb-1e76-410f-8b0d-a48bc85b8b58" containerName="horizon" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.226038 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="746eaa13-6098-490f-9ada-f8338fe419be" containerName="horizon" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.226044 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="d592b2fb-1e76-410f-8b0d-a48bc85b8b58" containerName="horizon-log" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.226057 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="746eaa13-6098-490f-9ada-f8338fe419be" containerName="horizon-log" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.226068 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="9703e257-1e49-46e0-bbdf-754fee621b4b" containerName="ceilometer-central-agent" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.226081 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="74de5806-1889-472c-b423-dfbdc2a63020" containerName="dnsmasq-dns" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.226095 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="9703e257-1e49-46e0-bbdf-754fee621b4b" containerName="ceilometer-notification-agent" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.226109 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="9703e257-1e49-46e0-bbdf-754fee621b4b" containerName="proxy-httpd" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.226122 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="9703e257-1e49-46e0-bbdf-754fee621b4b" containerName="sg-core" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.227752 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.235385 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.235550 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.235834 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.249748 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.274210 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99fbd787-9df6-4174-9c32-333f258751b7-run-httpd\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.274295 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.274400 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-config-data\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.274454 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.274480 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2mnh\" (UniqueName: \"kubernetes.io/projected/99fbd787-9df6-4174-9c32-333f258751b7-kube-api-access-c2mnh\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.274501 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.274526 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99fbd787-9df6-4174-9c32-333f258751b7-log-httpd\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.274581 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-scripts\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.376460 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-scripts\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.376535 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99fbd787-9df6-4174-9c32-333f258751b7-run-httpd\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.376622 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.376658 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-config-data\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.376688 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.376709 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2mnh\" (UniqueName: \"kubernetes.io/projected/99fbd787-9df6-4174-9c32-333f258751b7-kube-api-access-c2mnh\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.376727 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.376748 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99fbd787-9df6-4174-9c32-333f258751b7-log-httpd\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.377692 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99fbd787-9df6-4174-9c32-333f258751b7-log-httpd\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.378622 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99fbd787-9df6-4174-9c32-333f258751b7-run-httpd\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.384851 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.385848 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.387639 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-config-data\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.390831 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.394453 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-scripts\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.398705 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2mnh\" (UniqueName: \"kubernetes.io/projected/99fbd787-9df6-4174-9c32-333f258751b7-kube-api-access-c2mnh\") pod \"ceilometer-0\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.631925 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.656186 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.881423 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"47aef657-3e55-4afa-9b2f-5650cb789e8a","Type":"ContainerStarted","Data":"9859043d9f89adb1179cfacf90e38db05af68485649e91692d7499cf83008d46"} Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.881765 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"47aef657-3e55-4afa-9b2f-5650cb789e8a","Type":"ContainerStarted","Data":"f0de70ba641985ce7b88f6798197c5f1fb231055b551d2ef84ffede3f344489b"} Mar 18 21:00:34 crc kubenswrapper[4950]: I0318 21:00:34.914233 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.361489073 podStartE2EDuration="18.914215031s" podCreationTimestamp="2026-03-18 21:00:16 +0000 UTC" firstStartedPulling="2026-03-18 21:00:17.785010364 +0000 UTC m=+3231.025852232" lastFinishedPulling="2026-03-18 21:00:33.337736322 +0000 UTC m=+3246.578578190" observedRunningTime="2026-03-18 21:00:34.903846368 +0000 UTC m=+3248.144688236" watchObservedRunningTime="2026-03-18 21:00:34.914215031 +0000 UTC m=+3248.155056899" Mar 18 21:00:35 crc kubenswrapper[4950]: I0318 21:00:35.137391 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 18 21:00:35 crc kubenswrapper[4950]: I0318 21:00:35.490713 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9703e257-1e49-46e0-bbdf-754fee621b4b" path="/var/lib/kubelet/pods/9703e257-1e49-46e0-bbdf-754fee621b4b/volumes" Mar 18 21:00:35 crc kubenswrapper[4950]: I0318 21:00:35.895307 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99fbd787-9df6-4174-9c32-333f258751b7","Type":"ContainerStarted","Data":"e44dc7ec8a36d575ada5d0722f3bfe253989188da89d641e1d4aab820d10676a"} Mar 18 21:00:36 crc kubenswrapper[4950]: I0318 21:00:36.904200 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99fbd787-9df6-4174-9c32-333f258751b7","Type":"ContainerStarted","Data":"65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197"} Mar 18 21:00:36 crc kubenswrapper[4950]: I0318 21:00:36.993033 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Mar 18 21:00:37 crc kubenswrapper[4950]: I0318 21:00:37.915759 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99fbd787-9df6-4174-9c32-333f258751b7","Type":"ContainerStarted","Data":"0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee"} Mar 18 21:00:37 crc kubenswrapper[4950]: I0318 21:00:37.964970 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5bfb97c7d-h56mk" podUID="26e882de-ea5e-4a22-889d-85c5d4347b1e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.9:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:51622->10.217.1.9:8443: read: connection reset by peer" Mar 18 21:00:38 crc kubenswrapper[4950]: I0318 21:00:38.927505 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99fbd787-9df6-4174-9c32-333f258751b7","Type":"ContainerStarted","Data":"805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f"} Mar 18 21:00:38 crc kubenswrapper[4950]: I0318 21:00:38.935325 4950 generic.go:334] "Generic (PLEG): container finished" podID="26e882de-ea5e-4a22-889d-85c5d4347b1e" containerID="aa50a9c7f3ed240952a5dacf482e2676dcb49617b9f6bddc43a68f1aa9938ed8" exitCode=0 Mar 18 21:00:38 crc kubenswrapper[4950]: I0318 21:00:38.935376 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bfb97c7d-h56mk" event={"ID":"26e882de-ea5e-4a22-889d-85c5d4347b1e","Type":"ContainerDied","Data":"aa50a9c7f3ed240952a5dacf482e2676dcb49617b9f6bddc43a68f1aa9938ed8"} Mar 18 21:00:39 crc kubenswrapper[4950]: I0318 21:00:39.084583 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Mar 18 21:00:39 crc kubenswrapper[4950]: I0318 21:00:39.120466 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Mar 18 21:00:39 crc kubenswrapper[4950]: I0318 21:00:39.953132 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="fcb05ff2-359e-4728-a148-b8f8678e4ce8" containerName="manila-scheduler" containerID="cri-o://8ab94c7a21fb072ce10d18b5450749842b6d8a52d33e406c4dde29e82d766b8e" gracePeriod=30 Mar 18 21:00:39 crc kubenswrapper[4950]: I0318 21:00:39.953760 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="fcb05ff2-359e-4728-a148-b8f8678e4ce8" containerName="probe" containerID="cri-o://36c82e8c490fb8f94c4df1767d29546a61825206aa668c3a2a92b4d3b5da6be8" gracePeriod=30 Mar 18 21:00:40 crc kubenswrapper[4950]: I0318 21:00:40.402637 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fkdtp" podUID="73b8c337-25a3-4f8a-88c9-1e24e96589b4" containerName="registry-server" probeResult="failure" output=< Mar 18 21:00:40 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 21:00:40 crc kubenswrapper[4950]: > Mar 18 21:00:40 crc kubenswrapper[4950]: I0318 21:00:40.963131 4950 generic.go:334] "Generic (PLEG): container finished" podID="fcb05ff2-359e-4728-a148-b8f8678e4ce8" containerID="36c82e8c490fb8f94c4df1767d29546a61825206aa668c3a2a92b4d3b5da6be8" exitCode=0 Mar 18 21:00:40 crc kubenswrapper[4950]: I0318 21:00:40.963215 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"fcb05ff2-359e-4728-a148-b8f8678e4ce8","Type":"ContainerDied","Data":"36c82e8c490fb8f94c4df1767d29546a61825206aa668c3a2a92b4d3b5da6be8"} Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.698210 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.735515 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jx4tn\" (UniqueName: \"kubernetes.io/projected/fcb05ff2-359e-4728-a148-b8f8678e4ce8-kube-api-access-jx4tn\") pod \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.735606 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-scripts\") pod \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.735695 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-config-data\") pod \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.735739 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-config-data-custom\") pod \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.735781 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fcb05ff2-359e-4728-a148-b8f8678e4ce8-etc-machine-id\") pod \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.735941 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-combined-ca-bundle\") pod \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\" (UID: \"fcb05ff2-359e-4728-a148-b8f8678e4ce8\") " Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.736960 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fcb05ff2-359e-4728-a148-b8f8678e4ce8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "fcb05ff2-359e-4728-a148-b8f8678e4ce8" (UID: "fcb05ff2-359e-4728-a148-b8f8678e4ce8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.764679 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fcb05ff2-359e-4728-a148-b8f8678e4ce8" (UID: "fcb05ff2-359e-4728-a148-b8f8678e4ce8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.765644 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcb05ff2-359e-4728-a148-b8f8678e4ce8-kube-api-access-jx4tn" (OuterVolumeSpecName: "kube-api-access-jx4tn") pod "fcb05ff2-359e-4728-a148-b8f8678e4ce8" (UID: "fcb05ff2-359e-4728-a148-b8f8678e4ce8"). InnerVolumeSpecName "kube-api-access-jx4tn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.776912 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-scripts" (OuterVolumeSpecName: "scripts") pod "fcb05ff2-359e-4728-a148-b8f8678e4ce8" (UID: "fcb05ff2-359e-4728-a148-b8f8678e4ce8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.837938 4950 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.837967 4950 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fcb05ff2-359e-4728-a148-b8f8678e4ce8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.837976 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jx4tn\" (UniqueName: \"kubernetes.io/projected/fcb05ff2-359e-4728-a148-b8f8678e4ce8-kube-api-access-jx4tn\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.837986 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.845058 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fcb05ff2-359e-4728-a148-b8f8678e4ce8" (UID: "fcb05ff2-359e-4728-a148-b8f8678e4ce8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.910581 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-config-data" (OuterVolumeSpecName: "config-data") pod "fcb05ff2-359e-4728-a148-b8f8678e4ce8" (UID: "fcb05ff2-359e-4728-a148-b8f8678e4ce8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.940438 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.940468 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcb05ff2-359e-4728-a148-b8f8678e4ce8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.972407 4950 generic.go:334] "Generic (PLEG): container finished" podID="fcb05ff2-359e-4728-a148-b8f8678e4ce8" containerID="8ab94c7a21fb072ce10d18b5450749842b6d8a52d33e406c4dde29e82d766b8e" exitCode=0 Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.972474 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.972517 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"fcb05ff2-359e-4728-a148-b8f8678e4ce8","Type":"ContainerDied","Data":"8ab94c7a21fb072ce10d18b5450749842b6d8a52d33e406c4dde29e82d766b8e"} Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.972551 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"fcb05ff2-359e-4728-a148-b8f8678e4ce8","Type":"ContainerDied","Data":"f9985702ff146febd91169282cf4d3f72ad678ca7361b090ba881d22831b4645"} Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.972572 4950 scope.go:117] "RemoveContainer" containerID="36c82e8c490fb8f94c4df1767d29546a61825206aa668c3a2a92b4d3b5da6be8" Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.976189 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99fbd787-9df6-4174-9c32-333f258751b7","Type":"ContainerStarted","Data":"1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3"} Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.976907 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 18 21:00:41 crc kubenswrapper[4950]: I0318 21:00:41.994722 4950 scope.go:117] "RemoveContainer" containerID="8ab94c7a21fb072ce10d18b5450749842b6d8a52d33e406c4dde29e82d766b8e" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.011704 4950 scope.go:117] "RemoveContainer" containerID="36c82e8c490fb8f94c4df1767d29546a61825206aa668c3a2a92b4d3b5da6be8" Mar 18 21:00:42 crc kubenswrapper[4950]: E0318 21:00:42.012509 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36c82e8c490fb8f94c4df1767d29546a61825206aa668c3a2a92b4d3b5da6be8\": container with ID starting with 36c82e8c490fb8f94c4df1767d29546a61825206aa668c3a2a92b4d3b5da6be8 not found: ID does not exist" containerID="36c82e8c490fb8f94c4df1767d29546a61825206aa668c3a2a92b4d3b5da6be8" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.012544 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36c82e8c490fb8f94c4df1767d29546a61825206aa668c3a2a92b4d3b5da6be8"} err="failed to get container status \"36c82e8c490fb8f94c4df1767d29546a61825206aa668c3a2a92b4d3b5da6be8\": rpc error: code = NotFound desc = could not find container \"36c82e8c490fb8f94c4df1767d29546a61825206aa668c3a2a92b4d3b5da6be8\": container with ID starting with 36c82e8c490fb8f94c4df1767d29546a61825206aa668c3a2a92b4d3b5da6be8 not found: ID does not exist" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.012577 4950 scope.go:117] "RemoveContainer" containerID="8ab94c7a21fb072ce10d18b5450749842b6d8a52d33e406c4dde29e82d766b8e" Mar 18 21:00:42 crc kubenswrapper[4950]: E0318 21:00:42.014289 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ab94c7a21fb072ce10d18b5450749842b6d8a52d33e406c4dde29e82d766b8e\": container with ID starting with 8ab94c7a21fb072ce10d18b5450749842b6d8a52d33e406c4dde29e82d766b8e not found: ID does not exist" containerID="8ab94c7a21fb072ce10d18b5450749842b6d8a52d33e406c4dde29e82d766b8e" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.014317 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ab94c7a21fb072ce10d18b5450749842b6d8a52d33e406c4dde29e82d766b8e"} err="failed to get container status \"8ab94c7a21fb072ce10d18b5450749842b6d8a52d33e406c4dde29e82d766b8e\": rpc error: code = NotFound desc = could not find container \"8ab94c7a21fb072ce10d18b5450749842b6d8a52d33e406c4dde29e82d766b8e\": container with ID starting with 8ab94c7a21fb072ce10d18b5450749842b6d8a52d33e406c4dde29e82d766b8e not found: ID does not exist" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.018401 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.396872476 podStartE2EDuration="8.018381658s" podCreationTimestamp="2026-03-18 21:00:34 +0000 UTC" firstStartedPulling="2026-03-18 21:00:35.133706865 +0000 UTC m=+3248.374548733" lastFinishedPulling="2026-03-18 21:00:40.755216047 +0000 UTC m=+3253.996057915" observedRunningTime="2026-03-18 21:00:42.013047838 +0000 UTC m=+3255.253889706" watchObservedRunningTime="2026-03-18 21:00:42.018381658 +0000 UTC m=+3255.259223526" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.034869 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.041615 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.060436 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Mar 18 21:00:42 crc kubenswrapper[4950]: E0318 21:00:42.060840 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcb05ff2-359e-4728-a148-b8f8678e4ce8" containerName="manila-scheduler" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.060859 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcb05ff2-359e-4728-a148-b8f8678e4ce8" containerName="manila-scheduler" Mar 18 21:00:42 crc kubenswrapper[4950]: E0318 21:00:42.060881 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcb05ff2-359e-4728-a148-b8f8678e4ce8" containerName="probe" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.060887 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcb05ff2-359e-4728-a148-b8f8678e4ce8" containerName="probe" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.061161 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcb05ff2-359e-4728-a148-b8f8678e4ce8" containerName="manila-scheduler" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.061181 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcb05ff2-359e-4728-a148-b8f8678e4ce8" containerName="probe" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.062200 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.069121 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.076534 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.143811 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28rpm\" (UniqueName: \"kubernetes.io/projected/a9bcf400-c9cc-4556-a2e2-6b7214c68579-kube-api-access-28rpm\") pod \"manila-scheduler-0\" (UID: \"a9bcf400-c9cc-4556-a2e2-6b7214c68579\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.143871 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9bcf400-c9cc-4556-a2e2-6b7214c68579-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"a9bcf400-c9cc-4556-a2e2-6b7214c68579\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.143954 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9bcf400-c9cc-4556-a2e2-6b7214c68579-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"a9bcf400-c9cc-4556-a2e2-6b7214c68579\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.144122 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a9bcf400-c9cc-4556-a2e2-6b7214c68579-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"a9bcf400-c9cc-4556-a2e2-6b7214c68579\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.144221 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9bcf400-c9cc-4556-a2e2-6b7214c68579-scripts\") pod \"manila-scheduler-0\" (UID: \"a9bcf400-c9cc-4556-a2e2-6b7214c68579\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.144313 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9bcf400-c9cc-4556-a2e2-6b7214c68579-config-data\") pod \"manila-scheduler-0\" (UID: \"a9bcf400-c9cc-4556-a2e2-6b7214c68579\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.246689 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28rpm\" (UniqueName: \"kubernetes.io/projected/a9bcf400-c9cc-4556-a2e2-6b7214c68579-kube-api-access-28rpm\") pod \"manila-scheduler-0\" (UID: \"a9bcf400-c9cc-4556-a2e2-6b7214c68579\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.246771 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9bcf400-c9cc-4556-a2e2-6b7214c68579-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"a9bcf400-c9cc-4556-a2e2-6b7214c68579\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.246829 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9bcf400-c9cc-4556-a2e2-6b7214c68579-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"a9bcf400-c9cc-4556-a2e2-6b7214c68579\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.246868 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a9bcf400-c9cc-4556-a2e2-6b7214c68579-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"a9bcf400-c9cc-4556-a2e2-6b7214c68579\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.246905 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9bcf400-c9cc-4556-a2e2-6b7214c68579-scripts\") pod \"manila-scheduler-0\" (UID: \"a9bcf400-c9cc-4556-a2e2-6b7214c68579\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.246946 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9bcf400-c9cc-4556-a2e2-6b7214c68579-config-data\") pod \"manila-scheduler-0\" (UID: \"a9bcf400-c9cc-4556-a2e2-6b7214c68579\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.247018 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a9bcf400-c9cc-4556-a2e2-6b7214c68579-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"a9bcf400-c9cc-4556-a2e2-6b7214c68579\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.250197 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9bcf400-c9cc-4556-a2e2-6b7214c68579-scripts\") pod \"manila-scheduler-0\" (UID: \"a9bcf400-c9cc-4556-a2e2-6b7214c68579\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.250850 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9bcf400-c9cc-4556-a2e2-6b7214c68579-config-data\") pod \"manila-scheduler-0\" (UID: \"a9bcf400-c9cc-4556-a2e2-6b7214c68579\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.251023 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9bcf400-c9cc-4556-a2e2-6b7214c68579-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"a9bcf400-c9cc-4556-a2e2-6b7214c68579\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.251740 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9bcf400-c9cc-4556-a2e2-6b7214c68579-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"a9bcf400-c9cc-4556-a2e2-6b7214c68579\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.264898 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28rpm\" (UniqueName: \"kubernetes.io/projected/a9bcf400-c9cc-4556-a2e2-6b7214c68579-kube-api-access-28rpm\") pod \"manila-scheduler-0\" (UID: \"a9bcf400-c9cc-4556-a2e2-6b7214c68579\") " pod="openstack/manila-scheduler-0" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.411656 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.925401 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Mar 18 21:00:42 crc kubenswrapper[4950]: I0318 21:00:42.995903 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"a9bcf400-c9cc-4556-a2e2-6b7214c68579","Type":"ContainerStarted","Data":"0c9ebc0fdfceafb9addb9fa95181cee2230f61c6ae81ae80322676c0fb1b6c5f"} Mar 18 21:00:43 crc kubenswrapper[4950]: I0318 21:00:43.494052 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcb05ff2-359e-4728-a148-b8f8678e4ce8" path="/var/lib/kubelet/pods/fcb05ff2-359e-4728-a148-b8f8678e4ce8/volumes" Mar 18 21:00:43 crc kubenswrapper[4950]: I0318 21:00:43.851881 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.005532 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"a9bcf400-c9cc-4556-a2e2-6b7214c68579","Type":"ContainerStarted","Data":"0ed848bf9b529761969c73911f4b27b8e0603588001953a007242271a68c9456"} Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.005571 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"a9bcf400-c9cc-4556-a2e2-6b7214c68579","Type":"ContainerStarted","Data":"825a9f3ea904e9ef8a73b9b08685e234c499fd5d9d0ac741c1a3f315a36602d9"} Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.005680 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="99fbd787-9df6-4174-9c32-333f258751b7" containerName="ceilometer-central-agent" containerID="cri-o://65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197" gracePeriod=30 Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.005723 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="99fbd787-9df6-4174-9c32-333f258751b7" containerName="proxy-httpd" containerID="cri-o://1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3" gracePeriod=30 Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.005766 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="99fbd787-9df6-4174-9c32-333f258751b7" containerName="ceilometer-notification-agent" containerID="cri-o://0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee" gracePeriod=30 Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.005767 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="99fbd787-9df6-4174-9c32-333f258751b7" containerName="sg-core" containerID="cri-o://805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f" gracePeriod=30 Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.044296 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=2.044276819 podStartE2EDuration="2.044276819s" podCreationTimestamp="2026-03-18 21:00:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 21:00:44.030275981 +0000 UTC m=+3257.271117849" watchObservedRunningTime="2026-03-18 21:00:44.044276819 +0000 UTC m=+3257.285118687" Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.480692 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:00:44 crc kubenswrapper[4950]: E0318 21:00:44.481005 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.826457 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.894587 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-scripts\") pod \"99fbd787-9df6-4174-9c32-333f258751b7\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.894636 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99fbd787-9df6-4174-9c32-333f258751b7-run-httpd\") pod \"99fbd787-9df6-4174-9c32-333f258751b7\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.894676 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2mnh\" (UniqueName: \"kubernetes.io/projected/99fbd787-9df6-4174-9c32-333f258751b7-kube-api-access-c2mnh\") pod \"99fbd787-9df6-4174-9c32-333f258751b7\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.894707 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-combined-ca-bundle\") pod \"99fbd787-9df6-4174-9c32-333f258751b7\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.894870 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-ceilometer-tls-certs\") pod \"99fbd787-9df6-4174-9c32-333f258751b7\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.894893 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-sg-core-conf-yaml\") pod \"99fbd787-9df6-4174-9c32-333f258751b7\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.894909 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-config-data\") pod \"99fbd787-9df6-4174-9c32-333f258751b7\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.894943 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99fbd787-9df6-4174-9c32-333f258751b7-log-httpd\") pod \"99fbd787-9df6-4174-9c32-333f258751b7\" (UID: \"99fbd787-9df6-4174-9c32-333f258751b7\") " Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.895788 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99fbd787-9df6-4174-9c32-333f258751b7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "99fbd787-9df6-4174-9c32-333f258751b7" (UID: "99fbd787-9df6-4174-9c32-333f258751b7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.896875 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99fbd787-9df6-4174-9c32-333f258751b7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "99fbd787-9df6-4174-9c32-333f258751b7" (UID: "99fbd787-9df6-4174-9c32-333f258751b7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.906553 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99fbd787-9df6-4174-9c32-333f258751b7-kube-api-access-c2mnh" (OuterVolumeSpecName: "kube-api-access-c2mnh") pod "99fbd787-9df6-4174-9c32-333f258751b7" (UID: "99fbd787-9df6-4174-9c32-333f258751b7"). InnerVolumeSpecName "kube-api-access-c2mnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.919835 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-scripts" (OuterVolumeSpecName: "scripts") pod "99fbd787-9df6-4174-9c32-333f258751b7" (UID: "99fbd787-9df6-4174-9c32-333f258751b7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.945395 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "99fbd787-9df6-4174-9c32-333f258751b7" (UID: "99fbd787-9df6-4174-9c32-333f258751b7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.965900 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "99fbd787-9df6-4174-9c32-333f258751b7" (UID: "99fbd787-9df6-4174-9c32-333f258751b7"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.997945 4950 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.998164 4950 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.998174 4950 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99fbd787-9df6-4174-9c32-333f258751b7-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.998182 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.998191 4950 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99fbd787-9df6-4174-9c32-333f258751b7-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:44 crc kubenswrapper[4950]: I0318 21:00:44.998201 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2mnh\" (UniqueName: \"kubernetes.io/projected/99fbd787-9df6-4174-9c32-333f258751b7-kube-api-access-c2mnh\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.006366 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99fbd787-9df6-4174-9c32-333f258751b7" (UID: "99fbd787-9df6-4174-9c32-333f258751b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.023742 4950 generic.go:334] "Generic (PLEG): container finished" podID="99fbd787-9df6-4174-9c32-333f258751b7" containerID="1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3" exitCode=0 Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.023808 4950 generic.go:334] "Generic (PLEG): container finished" podID="99fbd787-9df6-4174-9c32-333f258751b7" containerID="805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f" exitCode=2 Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.023820 4950 generic.go:334] "Generic (PLEG): container finished" podID="99fbd787-9df6-4174-9c32-333f258751b7" containerID="0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee" exitCode=0 Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.023854 4950 generic.go:334] "Generic (PLEG): container finished" podID="99fbd787-9df6-4174-9c32-333f258751b7" containerID="65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197" exitCode=0 Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.023874 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99fbd787-9df6-4174-9c32-333f258751b7","Type":"ContainerDied","Data":"1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3"} Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.023912 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99fbd787-9df6-4174-9c32-333f258751b7","Type":"ContainerDied","Data":"805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f"} Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.023921 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99fbd787-9df6-4174-9c32-333f258751b7","Type":"ContainerDied","Data":"0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee"} Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.023931 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99fbd787-9df6-4174-9c32-333f258751b7","Type":"ContainerDied","Data":"65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197"} Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.023943 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99fbd787-9df6-4174-9c32-333f258751b7","Type":"ContainerDied","Data":"e44dc7ec8a36d575ada5d0722f3bfe253989188da89d641e1d4aab820d10676a"} Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.023963 4950 scope.go:117] "RemoveContainer" containerID="1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.023792 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.045616 4950 scope.go:117] "RemoveContainer" containerID="805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.050776 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-config-data" (OuterVolumeSpecName: "config-data") pod "99fbd787-9df6-4174-9c32-333f258751b7" (UID: "99fbd787-9df6-4174-9c32-333f258751b7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.066606 4950 scope.go:117] "RemoveContainer" containerID="0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.085010 4950 scope.go:117] "RemoveContainer" containerID="65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.099739 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.099773 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99fbd787-9df6-4174-9c32-333f258751b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.105035 4950 scope.go:117] "RemoveContainer" containerID="1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3" Mar 18 21:00:45 crc kubenswrapper[4950]: E0318 21:00:45.106751 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3\": container with ID starting with 1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3 not found: ID does not exist" containerID="1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.106790 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3"} err="failed to get container status \"1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3\": rpc error: code = NotFound desc = could not find container \"1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3\": container with ID starting with 1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3 not found: ID does not exist" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.106815 4950 scope.go:117] "RemoveContainer" containerID="805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f" Mar 18 21:00:45 crc kubenswrapper[4950]: E0318 21:00:45.107482 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f\": container with ID starting with 805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f not found: ID does not exist" containerID="805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.107512 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f"} err="failed to get container status \"805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f\": rpc error: code = NotFound desc = could not find container \"805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f\": container with ID starting with 805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f not found: ID does not exist" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.107534 4950 scope.go:117] "RemoveContainer" containerID="0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee" Mar 18 21:00:45 crc kubenswrapper[4950]: E0318 21:00:45.107731 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee\": container with ID starting with 0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee not found: ID does not exist" containerID="0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.107762 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee"} err="failed to get container status \"0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee\": rpc error: code = NotFound desc = could not find container \"0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee\": container with ID starting with 0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee not found: ID does not exist" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.107781 4950 scope.go:117] "RemoveContainer" containerID="65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197" Mar 18 21:00:45 crc kubenswrapper[4950]: E0318 21:00:45.107984 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197\": container with ID starting with 65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197 not found: ID does not exist" containerID="65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.108005 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197"} err="failed to get container status \"65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197\": rpc error: code = NotFound desc = could not find container \"65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197\": container with ID starting with 65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197 not found: ID does not exist" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.108019 4950 scope.go:117] "RemoveContainer" containerID="1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.108188 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3"} err="failed to get container status \"1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3\": rpc error: code = NotFound desc = could not find container \"1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3\": container with ID starting with 1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3 not found: ID does not exist" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.108205 4950 scope.go:117] "RemoveContainer" containerID="805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.108372 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f"} err="failed to get container status \"805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f\": rpc error: code = NotFound desc = could not find container \"805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f\": container with ID starting with 805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f not found: ID does not exist" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.108396 4950 scope.go:117] "RemoveContainer" containerID="0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.108648 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee"} err="failed to get container status \"0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee\": rpc error: code = NotFound desc = could not find container \"0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee\": container with ID starting with 0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee not found: ID does not exist" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.108667 4950 scope.go:117] "RemoveContainer" containerID="65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.108823 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197"} err="failed to get container status \"65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197\": rpc error: code = NotFound desc = could not find container \"65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197\": container with ID starting with 65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197 not found: ID does not exist" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.108846 4950 scope.go:117] "RemoveContainer" containerID="1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.108996 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3"} err="failed to get container status \"1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3\": rpc error: code = NotFound desc = could not find container \"1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3\": container with ID starting with 1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3 not found: ID does not exist" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.109013 4950 scope.go:117] "RemoveContainer" containerID="805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.109196 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f"} err="failed to get container status \"805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f\": rpc error: code = NotFound desc = could not find container \"805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f\": container with ID starting with 805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f not found: ID does not exist" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.109213 4950 scope.go:117] "RemoveContainer" containerID="0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.109397 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee"} err="failed to get container status \"0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee\": rpc error: code = NotFound desc = could not find container \"0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee\": container with ID starting with 0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee not found: ID does not exist" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.109447 4950 scope.go:117] "RemoveContainer" containerID="65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.109634 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197"} err="failed to get container status \"65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197\": rpc error: code = NotFound desc = could not find container \"65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197\": container with ID starting with 65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197 not found: ID does not exist" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.109653 4950 scope.go:117] "RemoveContainer" containerID="1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.109802 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3"} err="failed to get container status \"1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3\": rpc error: code = NotFound desc = could not find container \"1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3\": container with ID starting with 1520d5a881bab813a11a4da8f0f895bc120e8a59090837aec3efc5e8bf39dcd3 not found: ID does not exist" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.109819 4950 scope.go:117] "RemoveContainer" containerID="805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.109977 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f"} err="failed to get container status \"805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f\": rpc error: code = NotFound desc = could not find container \"805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f\": container with ID starting with 805f126ba9f436cadacfb2a1e1c3064d5c7741e6750ca12f3d74e6cc71ef298f not found: ID does not exist" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.110001 4950 scope.go:117] "RemoveContainer" containerID="0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.110239 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee"} err="failed to get container status \"0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee\": rpc error: code = NotFound desc = could not find container \"0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee\": container with ID starting with 0f1cbf2612e77432904eb947e4f3db2b23f8038670e8a18317052571529085ee not found: ID does not exist" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.110261 4950 scope.go:117] "RemoveContainer" containerID="65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.110477 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197"} err="failed to get container status \"65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197\": rpc error: code = NotFound desc = could not find container \"65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197\": container with ID starting with 65c767b8053d4387e9adbf480ab7f794610d914a87618e85776dd1cb1108e197 not found: ID does not exist" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.374446 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.390288 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.403621 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 18 21:00:45 crc kubenswrapper[4950]: E0318 21:00:45.404050 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99fbd787-9df6-4174-9c32-333f258751b7" containerName="proxy-httpd" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.404065 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="99fbd787-9df6-4174-9c32-333f258751b7" containerName="proxy-httpd" Mar 18 21:00:45 crc kubenswrapper[4950]: E0318 21:00:45.404081 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99fbd787-9df6-4174-9c32-333f258751b7" containerName="ceilometer-central-agent" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.404087 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="99fbd787-9df6-4174-9c32-333f258751b7" containerName="ceilometer-central-agent" Mar 18 21:00:45 crc kubenswrapper[4950]: E0318 21:00:45.404099 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99fbd787-9df6-4174-9c32-333f258751b7" containerName="ceilometer-notification-agent" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.404106 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="99fbd787-9df6-4174-9c32-333f258751b7" containerName="ceilometer-notification-agent" Mar 18 21:00:45 crc kubenswrapper[4950]: E0318 21:00:45.404131 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99fbd787-9df6-4174-9c32-333f258751b7" containerName="sg-core" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.404137 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="99fbd787-9df6-4174-9c32-333f258751b7" containerName="sg-core" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.404296 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="99fbd787-9df6-4174-9c32-333f258751b7" containerName="ceilometer-central-agent" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.404312 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="99fbd787-9df6-4174-9c32-333f258751b7" containerName="proxy-httpd" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.404324 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="99fbd787-9df6-4174-9c32-333f258751b7" containerName="ceilometer-notification-agent" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.404343 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="99fbd787-9df6-4174-9c32-333f258751b7" containerName="sg-core" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.405929 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.414855 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.415842 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.415928 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.416774 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.491439 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99fbd787-9df6-4174-9c32-333f258751b7" path="/var/lib/kubelet/pods/99fbd787-9df6-4174-9c32-333f258751b7/volumes" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.503272 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.506583 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b5ef7308-b687-45b3-bcf1-1ef804092712-run-httpd\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.506620 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b5ef7308-b687-45b3-bcf1-1ef804092712-log-httpd\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.506685 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7sjl\" (UniqueName: \"kubernetes.io/projected/b5ef7308-b687-45b3-bcf1-1ef804092712-kube-api-access-n7sjl\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.506722 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5ef7308-b687-45b3-bcf1-1ef804092712-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.506768 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5ef7308-b687-45b3-bcf1-1ef804092712-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.506829 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5ef7308-b687-45b3-bcf1-1ef804092712-scripts\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.506877 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5ef7308-b687-45b3-bcf1-1ef804092712-config-data\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.506916 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b5ef7308-b687-45b3-bcf1-1ef804092712-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.609143 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b5ef7308-b687-45b3-bcf1-1ef804092712-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.609199 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b5ef7308-b687-45b3-bcf1-1ef804092712-run-httpd\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.609217 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b5ef7308-b687-45b3-bcf1-1ef804092712-log-httpd\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.609290 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7sjl\" (UniqueName: \"kubernetes.io/projected/b5ef7308-b687-45b3-bcf1-1ef804092712-kube-api-access-n7sjl\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.609333 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5ef7308-b687-45b3-bcf1-1ef804092712-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.609388 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5ef7308-b687-45b3-bcf1-1ef804092712-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.609573 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5ef7308-b687-45b3-bcf1-1ef804092712-scripts\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.609622 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5ef7308-b687-45b3-bcf1-1ef804092712-config-data\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.612117 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b5ef7308-b687-45b3-bcf1-1ef804092712-run-httpd\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.612354 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b5ef7308-b687-45b3-bcf1-1ef804092712-log-httpd\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.618564 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5ef7308-b687-45b3-bcf1-1ef804092712-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.626430 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5ef7308-b687-45b3-bcf1-1ef804092712-scripts\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.628727 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5ef7308-b687-45b3-bcf1-1ef804092712-config-data\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.631288 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5ef7308-b687-45b3-bcf1-1ef804092712-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.635218 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b5ef7308-b687-45b3-bcf1-1ef804092712-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.635654 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7sjl\" (UniqueName: \"kubernetes.io/projected/b5ef7308-b687-45b3-bcf1-1ef804092712-kube-api-access-n7sjl\") pod \"ceilometer-0\" (UID: \"b5ef7308-b687-45b3-bcf1-1ef804092712\") " pod="openstack/ceilometer-0" Mar 18 21:00:45 crc kubenswrapper[4950]: I0318 21:00:45.788572 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 18 21:00:46 crc kubenswrapper[4950]: I0318 21:00:46.294009 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 18 21:00:46 crc kubenswrapper[4950]: I0318 21:00:46.403674 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5bfb97c7d-h56mk" podUID="26e882de-ea5e-4a22-889d-85c5d4347b1e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.9:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.9:8443: connect: connection refused" Mar 18 21:00:47 crc kubenswrapper[4950]: I0318 21:00:47.064256 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b5ef7308-b687-45b3-bcf1-1ef804092712","Type":"ContainerStarted","Data":"994ad7a992f88c1b8b67d24025c5a5998316ba994019259e74638214bb99e430"} Mar 18 21:00:48 crc kubenswrapper[4950]: I0318 21:00:48.079082 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b5ef7308-b687-45b3-bcf1-1ef804092712","Type":"ContainerStarted","Data":"0aa63fb6b3101be8257be8043313aada7c57153d6779acf84e680d75d68b9fb4"} Mar 18 21:00:48 crc kubenswrapper[4950]: I0318 21:00:48.079668 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b5ef7308-b687-45b3-bcf1-1ef804092712","Type":"ContainerStarted","Data":"8e693e3cf6d021f74ceaabb394df03b85d37db1b16cb7c697c26c9746a20811b"} Mar 18 21:00:48 crc kubenswrapper[4950]: I0318 21:00:48.638957 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Mar 18 21:00:48 crc kubenswrapper[4950]: I0318 21:00:48.724344 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Mar 18 21:00:49 crc kubenswrapper[4950]: I0318 21:00:49.086023 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="47aef657-3e55-4afa-9b2f-5650cb789e8a" containerName="manila-share" containerID="cri-o://f0de70ba641985ce7b88f6798197c5f1fb231055b551d2ef84ffede3f344489b" gracePeriod=30 Mar 18 21:00:49 crc kubenswrapper[4950]: I0318 21:00:49.086536 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="47aef657-3e55-4afa-9b2f-5650cb789e8a" containerName="probe" containerID="cri-o://9859043d9f89adb1179cfacf90e38db05af68485649e91692d7499cf83008d46" gracePeriod=30 Mar 18 21:00:49 crc kubenswrapper[4950]: I0318 21:00:49.409709 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fkdtp" Mar 18 21:00:49 crc kubenswrapper[4950]: I0318 21:00:49.459755 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fkdtp" Mar 18 21:00:49 crc kubenswrapper[4950]: I0318 21:00:49.644262 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fkdtp"] Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.044897 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.097452 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b5ef7308-b687-45b3-bcf1-1ef804092712","Type":"ContainerStarted","Data":"327b9bd50c87c2c64bfb56298cc96ebcb702eafb4f75c4aa0d3c1edd161b4458"} Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.100173 4950 generic.go:334] "Generic (PLEG): container finished" podID="47aef657-3e55-4afa-9b2f-5650cb789e8a" containerID="9859043d9f89adb1179cfacf90e38db05af68485649e91692d7499cf83008d46" exitCode=0 Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.100199 4950 generic.go:334] "Generic (PLEG): container finished" podID="47aef657-3e55-4afa-9b2f-5650cb789e8a" containerID="f0de70ba641985ce7b88f6798197c5f1fb231055b551d2ef84ffede3f344489b" exitCode=1 Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.101060 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.101664 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"47aef657-3e55-4afa-9b2f-5650cb789e8a","Type":"ContainerDied","Data":"9859043d9f89adb1179cfacf90e38db05af68485649e91692d7499cf83008d46"} Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.101694 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"47aef657-3e55-4afa-9b2f-5650cb789e8a","Type":"ContainerDied","Data":"f0de70ba641985ce7b88f6798197c5f1fb231055b551d2ef84ffede3f344489b"} Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.101705 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"47aef657-3e55-4afa-9b2f-5650cb789e8a","Type":"ContainerDied","Data":"2ec34286594a538e040b8e0121ad51759a803001ac959bd5cf4bcfd2db5edf9a"} Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.101719 4950 scope.go:117] "RemoveContainer" containerID="9859043d9f89adb1179cfacf90e38db05af68485649e91692d7499cf83008d46" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.116480 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-config-data-custom\") pod \"47aef657-3e55-4afa-9b2f-5650cb789e8a\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.116530 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pq94g\" (UniqueName: \"kubernetes.io/projected/47aef657-3e55-4afa-9b2f-5650cb789e8a-kube-api-access-pq94g\") pod \"47aef657-3e55-4afa-9b2f-5650cb789e8a\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.116551 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/47aef657-3e55-4afa-9b2f-5650cb789e8a-etc-machine-id\") pod \"47aef657-3e55-4afa-9b2f-5650cb789e8a\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.116597 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/47aef657-3e55-4afa-9b2f-5650cb789e8a-var-lib-manila\") pod \"47aef657-3e55-4afa-9b2f-5650cb789e8a\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.116659 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-scripts\") pod \"47aef657-3e55-4afa-9b2f-5650cb789e8a\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.116755 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-config-data\") pod \"47aef657-3e55-4afa-9b2f-5650cb789e8a\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.116816 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-combined-ca-bundle\") pod \"47aef657-3e55-4afa-9b2f-5650cb789e8a\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.116848 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/47aef657-3e55-4afa-9b2f-5650cb789e8a-ceph\") pod \"47aef657-3e55-4afa-9b2f-5650cb789e8a\" (UID: \"47aef657-3e55-4afa-9b2f-5650cb789e8a\") " Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.117100 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/47aef657-3e55-4afa-9b2f-5650cb789e8a-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "47aef657-3e55-4afa-9b2f-5650cb789e8a" (UID: "47aef657-3e55-4afa-9b2f-5650cb789e8a"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.117311 4950 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/47aef657-3e55-4afa-9b2f-5650cb789e8a-var-lib-manila\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.117511 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/47aef657-3e55-4afa-9b2f-5650cb789e8a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "47aef657-3e55-4afa-9b2f-5650cb789e8a" (UID: "47aef657-3e55-4afa-9b2f-5650cb789e8a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.123024 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47aef657-3e55-4afa-9b2f-5650cb789e8a-ceph" (OuterVolumeSpecName: "ceph") pod "47aef657-3e55-4afa-9b2f-5650cb789e8a" (UID: "47aef657-3e55-4afa-9b2f-5650cb789e8a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.124663 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47aef657-3e55-4afa-9b2f-5650cb789e8a-kube-api-access-pq94g" (OuterVolumeSpecName: "kube-api-access-pq94g") pod "47aef657-3e55-4afa-9b2f-5650cb789e8a" (UID: "47aef657-3e55-4afa-9b2f-5650cb789e8a"). InnerVolumeSpecName "kube-api-access-pq94g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.132677 4950 scope.go:117] "RemoveContainer" containerID="f0de70ba641985ce7b88f6798197c5f1fb231055b551d2ef84ffede3f344489b" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.132884 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-scripts" (OuterVolumeSpecName: "scripts") pod "47aef657-3e55-4afa-9b2f-5650cb789e8a" (UID: "47aef657-3e55-4afa-9b2f-5650cb789e8a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.132943 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "47aef657-3e55-4afa-9b2f-5650cb789e8a" (UID: "47aef657-3e55-4afa-9b2f-5650cb789e8a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.202938 4950 scope.go:117] "RemoveContainer" containerID="9859043d9f89adb1179cfacf90e38db05af68485649e91692d7499cf83008d46" Mar 18 21:00:50 crc kubenswrapper[4950]: E0318 21:00:50.204240 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9859043d9f89adb1179cfacf90e38db05af68485649e91692d7499cf83008d46\": container with ID starting with 9859043d9f89adb1179cfacf90e38db05af68485649e91692d7499cf83008d46 not found: ID does not exist" containerID="9859043d9f89adb1179cfacf90e38db05af68485649e91692d7499cf83008d46" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.204294 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9859043d9f89adb1179cfacf90e38db05af68485649e91692d7499cf83008d46"} err="failed to get container status \"9859043d9f89adb1179cfacf90e38db05af68485649e91692d7499cf83008d46\": rpc error: code = NotFound desc = could not find container \"9859043d9f89adb1179cfacf90e38db05af68485649e91692d7499cf83008d46\": container with ID starting with 9859043d9f89adb1179cfacf90e38db05af68485649e91692d7499cf83008d46 not found: ID does not exist" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.204325 4950 scope.go:117] "RemoveContainer" containerID="f0de70ba641985ce7b88f6798197c5f1fb231055b551d2ef84ffede3f344489b" Mar 18 21:00:50 crc kubenswrapper[4950]: E0318 21:00:50.205603 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0de70ba641985ce7b88f6798197c5f1fb231055b551d2ef84ffede3f344489b\": container with ID starting with f0de70ba641985ce7b88f6798197c5f1fb231055b551d2ef84ffede3f344489b not found: ID does not exist" containerID="f0de70ba641985ce7b88f6798197c5f1fb231055b551d2ef84ffede3f344489b" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.205631 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0de70ba641985ce7b88f6798197c5f1fb231055b551d2ef84ffede3f344489b"} err="failed to get container status \"f0de70ba641985ce7b88f6798197c5f1fb231055b551d2ef84ffede3f344489b\": rpc error: code = NotFound desc = could not find container \"f0de70ba641985ce7b88f6798197c5f1fb231055b551d2ef84ffede3f344489b\": container with ID starting with f0de70ba641985ce7b88f6798197c5f1fb231055b551d2ef84ffede3f344489b not found: ID does not exist" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.205653 4950 scope.go:117] "RemoveContainer" containerID="9859043d9f89adb1179cfacf90e38db05af68485649e91692d7499cf83008d46" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.206134 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9859043d9f89adb1179cfacf90e38db05af68485649e91692d7499cf83008d46"} err="failed to get container status \"9859043d9f89adb1179cfacf90e38db05af68485649e91692d7499cf83008d46\": rpc error: code = NotFound desc = could not find container \"9859043d9f89adb1179cfacf90e38db05af68485649e91692d7499cf83008d46\": container with ID starting with 9859043d9f89adb1179cfacf90e38db05af68485649e91692d7499cf83008d46 not found: ID does not exist" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.206177 4950 scope.go:117] "RemoveContainer" containerID="f0de70ba641985ce7b88f6798197c5f1fb231055b551d2ef84ffede3f344489b" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.206323 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47aef657-3e55-4afa-9b2f-5650cb789e8a" (UID: "47aef657-3e55-4afa-9b2f-5650cb789e8a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.206434 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0de70ba641985ce7b88f6798197c5f1fb231055b551d2ef84ffede3f344489b"} err="failed to get container status \"f0de70ba641985ce7b88f6798197c5f1fb231055b551d2ef84ffede3f344489b\": rpc error: code = NotFound desc = could not find container \"f0de70ba641985ce7b88f6798197c5f1fb231055b551d2ef84ffede3f344489b\": container with ID starting with f0de70ba641985ce7b88f6798197c5f1fb231055b551d2ef84ffede3f344489b not found: ID does not exist" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.219181 4950 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.219208 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pq94g\" (UniqueName: \"kubernetes.io/projected/47aef657-3e55-4afa-9b2f-5650cb789e8a-kube-api-access-pq94g\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.219219 4950 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/47aef657-3e55-4afa-9b2f-5650cb789e8a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.219227 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.219237 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.219256 4950 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/47aef657-3e55-4afa-9b2f-5650cb789e8a-ceph\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.258189 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-config-data" (OuterVolumeSpecName: "config-data") pod "47aef657-3e55-4afa-9b2f-5650cb789e8a" (UID: "47aef657-3e55-4afa-9b2f-5650cb789e8a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.321080 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47aef657-3e55-4afa-9b2f-5650cb789e8a-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.435285 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.447276 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.464881 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Mar 18 21:00:50 crc kubenswrapper[4950]: E0318 21:00:50.465294 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47aef657-3e55-4afa-9b2f-5650cb789e8a" containerName="manila-share" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.465310 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="47aef657-3e55-4afa-9b2f-5650cb789e8a" containerName="manila-share" Mar 18 21:00:50 crc kubenswrapper[4950]: E0318 21:00:50.465334 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47aef657-3e55-4afa-9b2f-5650cb789e8a" containerName="probe" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.465340 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="47aef657-3e55-4afa-9b2f-5650cb789e8a" containerName="probe" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.465531 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="47aef657-3e55-4afa-9b2f-5650cb789e8a" containerName="manila-share" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.465559 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="47aef657-3e55-4afa-9b2f-5650cb789e8a" containerName="probe" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.466619 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.471117 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.483249 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.543352 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/139d12a8-de25-41a7-a92d-440cbeec98eb-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.543401 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/139d12a8-de25-41a7-a92d-440cbeec98eb-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.543439 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/139d12a8-de25-41a7-a92d-440cbeec98eb-config-data\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.543545 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnxfd\" (UniqueName: \"kubernetes.io/projected/139d12a8-de25-41a7-a92d-440cbeec98eb-kube-api-access-xnxfd\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.543699 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/139d12a8-de25-41a7-a92d-440cbeec98eb-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.543888 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/139d12a8-de25-41a7-a92d-440cbeec98eb-scripts\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.543968 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/139d12a8-de25-41a7-a92d-440cbeec98eb-ceph\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.543985 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/139d12a8-de25-41a7-a92d-440cbeec98eb-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.646152 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/139d12a8-de25-41a7-a92d-440cbeec98eb-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.646593 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/139d12a8-de25-41a7-a92d-440cbeec98eb-scripts\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.646634 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/139d12a8-de25-41a7-a92d-440cbeec98eb-ceph\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.646651 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/139d12a8-de25-41a7-a92d-440cbeec98eb-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.646695 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/139d12a8-de25-41a7-a92d-440cbeec98eb-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.646723 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/139d12a8-de25-41a7-a92d-440cbeec98eb-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.646741 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/139d12a8-de25-41a7-a92d-440cbeec98eb-config-data\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.646764 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnxfd\" (UniqueName: \"kubernetes.io/projected/139d12a8-de25-41a7-a92d-440cbeec98eb-kube-api-access-xnxfd\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.647323 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/139d12a8-de25-41a7-a92d-440cbeec98eb-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.647393 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/139d12a8-de25-41a7-a92d-440cbeec98eb-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.652457 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/139d12a8-de25-41a7-a92d-440cbeec98eb-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.653285 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/139d12a8-de25-41a7-a92d-440cbeec98eb-ceph\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.654181 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/139d12a8-de25-41a7-a92d-440cbeec98eb-config-data\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.655896 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/139d12a8-de25-41a7-a92d-440cbeec98eb-scripts\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.656809 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/139d12a8-de25-41a7-a92d-440cbeec98eb-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.673764 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnxfd\" (UniqueName: \"kubernetes.io/projected/139d12a8-de25-41a7-a92d-440cbeec98eb-kube-api-access-xnxfd\") pod \"manila-share-share1-0\" (UID: \"139d12a8-de25-41a7-a92d-440cbeec98eb\") " pod="openstack/manila-share-share1-0" Mar 18 21:00:50 crc kubenswrapper[4950]: I0318 21:00:50.817704 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Mar 18 21:00:51 crc kubenswrapper[4950]: I0318 21:00:51.108704 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fkdtp" podUID="73b8c337-25a3-4f8a-88c9-1e24e96589b4" containerName="registry-server" containerID="cri-o://ae6b85bb22621397ef4605bb22edb4ba8730fea01af4f5b8477234037824fb44" gracePeriod=2 Mar 18 21:00:51 crc kubenswrapper[4950]: I0318 21:00:51.423086 4950 scope.go:117] "RemoveContainer" containerID="1eb8b7086479d629bb79852b059d6109d1860922ae7b4b4422572d2d9d2d89bc" Mar 18 21:00:51 crc kubenswrapper[4950]: I0318 21:00:51.450018 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Mar 18 21:00:51 crc kubenswrapper[4950]: I0318 21:00:51.494608 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47aef657-3e55-4afa-9b2f-5650cb789e8a" path="/var/lib/kubelet/pods/47aef657-3e55-4afa-9b2f-5650cb789e8a/volumes" Mar 18 21:00:51 crc kubenswrapper[4950]: I0318 21:00:51.535093 4950 scope.go:117] "RemoveContainer" containerID="72e6b8ccee81fc70d3acc8fc40745fd0adb6fac9070e7724d58751929bc1766a" Mar 18 21:00:51 crc kubenswrapper[4950]: I0318 21:00:51.641823 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fkdtp" Mar 18 21:00:51 crc kubenswrapper[4950]: I0318 21:00:51.670340 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73b8c337-25a3-4f8a-88c9-1e24e96589b4-utilities\") pod \"73b8c337-25a3-4f8a-88c9-1e24e96589b4\" (UID: \"73b8c337-25a3-4f8a-88c9-1e24e96589b4\") " Mar 18 21:00:51 crc kubenswrapper[4950]: I0318 21:00:51.670445 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73b8c337-25a3-4f8a-88c9-1e24e96589b4-catalog-content\") pod \"73b8c337-25a3-4f8a-88c9-1e24e96589b4\" (UID: \"73b8c337-25a3-4f8a-88c9-1e24e96589b4\") " Mar 18 21:00:51 crc kubenswrapper[4950]: I0318 21:00:51.670674 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-565xn\" (UniqueName: \"kubernetes.io/projected/73b8c337-25a3-4f8a-88c9-1e24e96589b4-kube-api-access-565xn\") pod \"73b8c337-25a3-4f8a-88c9-1e24e96589b4\" (UID: \"73b8c337-25a3-4f8a-88c9-1e24e96589b4\") " Mar 18 21:00:51 crc kubenswrapper[4950]: I0318 21:00:51.674700 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73b8c337-25a3-4f8a-88c9-1e24e96589b4-utilities" (OuterVolumeSpecName: "utilities") pod "73b8c337-25a3-4f8a-88c9-1e24e96589b4" (UID: "73b8c337-25a3-4f8a-88c9-1e24e96589b4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:00:51 crc kubenswrapper[4950]: I0318 21:00:51.692209 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73b8c337-25a3-4f8a-88c9-1e24e96589b4-kube-api-access-565xn" (OuterVolumeSpecName: "kube-api-access-565xn") pod "73b8c337-25a3-4f8a-88c9-1e24e96589b4" (UID: "73b8c337-25a3-4f8a-88c9-1e24e96589b4"). InnerVolumeSpecName "kube-api-access-565xn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:00:51 crc kubenswrapper[4950]: I0318 21:00:51.773369 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-565xn\" (UniqueName: \"kubernetes.io/projected/73b8c337-25a3-4f8a-88c9-1e24e96589b4-kube-api-access-565xn\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:51 crc kubenswrapper[4950]: I0318 21:00:51.773674 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73b8c337-25a3-4f8a-88c9-1e24e96589b4-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:51 crc kubenswrapper[4950]: I0318 21:00:51.823464 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73b8c337-25a3-4f8a-88c9-1e24e96589b4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73b8c337-25a3-4f8a-88c9-1e24e96589b4" (UID: "73b8c337-25a3-4f8a-88c9-1e24e96589b4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:00:51 crc kubenswrapper[4950]: I0318 21:00:51.876399 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73b8c337-25a3-4f8a-88c9-1e24e96589b4-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 21:00:52 crc kubenswrapper[4950]: I0318 21:00:52.120352 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b5ef7308-b687-45b3-bcf1-1ef804092712","Type":"ContainerStarted","Data":"eb04a09a4ddc720149ad47e15602ccdd704ad4375f0df85b04ae7548f533248e"} Mar 18 21:00:52 crc kubenswrapper[4950]: I0318 21:00:52.120728 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 18 21:00:52 crc kubenswrapper[4950]: I0318 21:00:52.129078 4950 generic.go:334] "Generic (PLEG): container finished" podID="73b8c337-25a3-4f8a-88c9-1e24e96589b4" containerID="ae6b85bb22621397ef4605bb22edb4ba8730fea01af4f5b8477234037824fb44" exitCode=0 Mar 18 21:00:52 crc kubenswrapper[4950]: I0318 21:00:52.129120 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fkdtp" Mar 18 21:00:52 crc kubenswrapper[4950]: I0318 21:00:52.129130 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fkdtp" event={"ID":"73b8c337-25a3-4f8a-88c9-1e24e96589b4","Type":"ContainerDied","Data":"ae6b85bb22621397ef4605bb22edb4ba8730fea01af4f5b8477234037824fb44"} Mar 18 21:00:52 crc kubenswrapper[4950]: I0318 21:00:52.129373 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fkdtp" event={"ID":"73b8c337-25a3-4f8a-88c9-1e24e96589b4","Type":"ContainerDied","Data":"8d7e9b93360b003447d1bab1f39fb654bb5a33c0ce261952638f1ea90800f101"} Mar 18 21:00:52 crc kubenswrapper[4950]: I0318 21:00:52.129448 4950 scope.go:117] "RemoveContainer" containerID="ae6b85bb22621397ef4605bb22edb4ba8730fea01af4f5b8477234037824fb44" Mar 18 21:00:52 crc kubenswrapper[4950]: I0318 21:00:52.132468 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"139d12a8-de25-41a7-a92d-440cbeec98eb","Type":"ContainerStarted","Data":"078737dec132fc3b98f9b4b92701bfbece74e7a28c34e7ccc7580e9ae526e9a1"} Mar 18 21:00:52 crc kubenswrapper[4950]: I0318 21:00:52.159314 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.115766747 podStartE2EDuration="7.159289672s" podCreationTimestamp="2026-03-18 21:00:45 +0000 UTC" firstStartedPulling="2026-03-18 21:00:46.279287741 +0000 UTC m=+3259.520129609" lastFinishedPulling="2026-03-18 21:00:51.322810666 +0000 UTC m=+3264.563652534" observedRunningTime="2026-03-18 21:00:52.143541579 +0000 UTC m=+3265.384383467" watchObservedRunningTime="2026-03-18 21:00:52.159289672 +0000 UTC m=+3265.400131550" Mar 18 21:00:52 crc kubenswrapper[4950]: I0318 21:00:52.165001 4950 scope.go:117] "RemoveContainer" containerID="cdc902dcf161ce80f7fa862793b85274322032d048cecea57a76141aa626f347" Mar 18 21:00:52 crc kubenswrapper[4950]: I0318 21:00:52.192605 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fkdtp"] Mar 18 21:00:52 crc kubenswrapper[4950]: I0318 21:00:52.200198 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fkdtp"] Mar 18 21:00:52 crc kubenswrapper[4950]: I0318 21:00:52.215563 4950 scope.go:117] "RemoveContainer" containerID="356c48d5bbcb0fdb45b1fae26db49dc704d48a05cb46f0487049a2aa6c0be20a" Mar 18 21:00:52 crc kubenswrapper[4950]: I0318 21:00:52.263666 4950 scope.go:117] "RemoveContainer" containerID="ae6b85bb22621397ef4605bb22edb4ba8730fea01af4f5b8477234037824fb44" Mar 18 21:00:52 crc kubenswrapper[4950]: E0318 21:00:52.265506 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae6b85bb22621397ef4605bb22edb4ba8730fea01af4f5b8477234037824fb44\": container with ID starting with ae6b85bb22621397ef4605bb22edb4ba8730fea01af4f5b8477234037824fb44 not found: ID does not exist" containerID="ae6b85bb22621397ef4605bb22edb4ba8730fea01af4f5b8477234037824fb44" Mar 18 21:00:52 crc kubenswrapper[4950]: I0318 21:00:52.265538 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae6b85bb22621397ef4605bb22edb4ba8730fea01af4f5b8477234037824fb44"} err="failed to get container status \"ae6b85bb22621397ef4605bb22edb4ba8730fea01af4f5b8477234037824fb44\": rpc error: code = NotFound desc = could not find container \"ae6b85bb22621397ef4605bb22edb4ba8730fea01af4f5b8477234037824fb44\": container with ID starting with ae6b85bb22621397ef4605bb22edb4ba8730fea01af4f5b8477234037824fb44 not found: ID does not exist" Mar 18 21:00:52 crc kubenswrapper[4950]: I0318 21:00:52.265558 4950 scope.go:117] "RemoveContainer" containerID="cdc902dcf161ce80f7fa862793b85274322032d048cecea57a76141aa626f347" Mar 18 21:00:52 crc kubenswrapper[4950]: E0318 21:00:52.268362 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdc902dcf161ce80f7fa862793b85274322032d048cecea57a76141aa626f347\": container with ID starting with cdc902dcf161ce80f7fa862793b85274322032d048cecea57a76141aa626f347 not found: ID does not exist" containerID="cdc902dcf161ce80f7fa862793b85274322032d048cecea57a76141aa626f347" Mar 18 21:00:52 crc kubenswrapper[4950]: I0318 21:00:52.268397 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdc902dcf161ce80f7fa862793b85274322032d048cecea57a76141aa626f347"} err="failed to get container status \"cdc902dcf161ce80f7fa862793b85274322032d048cecea57a76141aa626f347\": rpc error: code = NotFound desc = could not find container \"cdc902dcf161ce80f7fa862793b85274322032d048cecea57a76141aa626f347\": container with ID starting with cdc902dcf161ce80f7fa862793b85274322032d048cecea57a76141aa626f347 not found: ID does not exist" Mar 18 21:00:52 crc kubenswrapper[4950]: I0318 21:00:52.268433 4950 scope.go:117] "RemoveContainer" containerID="356c48d5bbcb0fdb45b1fae26db49dc704d48a05cb46f0487049a2aa6c0be20a" Mar 18 21:00:52 crc kubenswrapper[4950]: E0318 21:00:52.270806 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"356c48d5bbcb0fdb45b1fae26db49dc704d48a05cb46f0487049a2aa6c0be20a\": container with ID starting with 356c48d5bbcb0fdb45b1fae26db49dc704d48a05cb46f0487049a2aa6c0be20a not found: ID does not exist" containerID="356c48d5bbcb0fdb45b1fae26db49dc704d48a05cb46f0487049a2aa6c0be20a" Mar 18 21:00:52 crc kubenswrapper[4950]: I0318 21:00:52.270839 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"356c48d5bbcb0fdb45b1fae26db49dc704d48a05cb46f0487049a2aa6c0be20a"} err="failed to get container status \"356c48d5bbcb0fdb45b1fae26db49dc704d48a05cb46f0487049a2aa6c0be20a\": rpc error: code = NotFound desc = could not find container \"356c48d5bbcb0fdb45b1fae26db49dc704d48a05cb46f0487049a2aa6c0be20a\": container with ID starting with 356c48d5bbcb0fdb45b1fae26db49dc704d48a05cb46f0487049a2aa6c0be20a not found: ID does not exist" Mar 18 21:00:52 crc kubenswrapper[4950]: I0318 21:00:52.413022 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Mar 18 21:00:53 crc kubenswrapper[4950]: I0318 21:00:53.142900 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"139d12a8-de25-41a7-a92d-440cbeec98eb","Type":"ContainerStarted","Data":"4bd683de93d9db812892a34e54f4d6cffa5c1e2804776a8d8b3d635cbbcbc567"} Mar 18 21:00:53 crc kubenswrapper[4950]: I0318 21:00:53.143181 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"139d12a8-de25-41a7-a92d-440cbeec98eb","Type":"ContainerStarted","Data":"73840f90d6383be0da73bc07bbea7baa16e51033c226ec2c5c258ce5b43c5021"} Mar 18 21:00:53 crc kubenswrapper[4950]: I0318 21:00:53.161812 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.161794697 podStartE2EDuration="3.161794697s" podCreationTimestamp="2026-03-18 21:00:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 21:00:53.160292378 +0000 UTC m=+3266.401134246" watchObservedRunningTime="2026-03-18 21:00:53.161794697 +0000 UTC m=+3266.402636565" Mar 18 21:00:53 crc kubenswrapper[4950]: I0318 21:00:53.503948 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73b8c337-25a3-4f8a-88c9-1e24e96589b4" path="/var/lib/kubelet/pods/73b8c337-25a3-4f8a-88c9-1e24e96589b4/volumes" Mar 18 21:00:55 crc kubenswrapper[4950]: I0318 21:00:55.480300 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:00:55 crc kubenswrapper[4950]: E0318 21:00:55.481003 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:00:56 crc kubenswrapper[4950]: I0318 21:00:56.403926 4950 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5bfb97c7d-h56mk" podUID="26e882de-ea5e-4a22-889d-85c5d4347b1e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.9:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.9:8443: connect: connection refused" Mar 18 21:00:56 crc kubenswrapper[4950]: I0318 21:00:56.404074 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 21:01:00 crc kubenswrapper[4950]: I0318 21:01:00.149404 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29564461-jwfx5"] Mar 18 21:01:00 crc kubenswrapper[4950]: E0318 21:01:00.150445 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73b8c337-25a3-4f8a-88c9-1e24e96589b4" containerName="registry-server" Mar 18 21:01:00 crc kubenswrapper[4950]: I0318 21:01:00.150463 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="73b8c337-25a3-4f8a-88c9-1e24e96589b4" containerName="registry-server" Mar 18 21:01:00 crc kubenswrapper[4950]: E0318 21:01:00.150502 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73b8c337-25a3-4f8a-88c9-1e24e96589b4" containerName="extract-content" Mar 18 21:01:00 crc kubenswrapper[4950]: I0318 21:01:00.150509 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="73b8c337-25a3-4f8a-88c9-1e24e96589b4" containerName="extract-content" Mar 18 21:01:00 crc kubenswrapper[4950]: E0318 21:01:00.150529 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73b8c337-25a3-4f8a-88c9-1e24e96589b4" containerName="extract-utilities" Mar 18 21:01:00 crc kubenswrapper[4950]: I0318 21:01:00.150539 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="73b8c337-25a3-4f8a-88c9-1e24e96589b4" containerName="extract-utilities" Mar 18 21:01:00 crc kubenswrapper[4950]: I0318 21:01:00.150780 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="73b8c337-25a3-4f8a-88c9-1e24e96589b4" containerName="registry-server" Mar 18 21:01:00 crc kubenswrapper[4950]: I0318 21:01:00.151375 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29564461-jwfx5" Mar 18 21:01:00 crc kubenswrapper[4950]: I0318 21:01:00.164930 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29564461-jwfx5"] Mar 18 21:01:00 crc kubenswrapper[4950]: I0318 21:01:00.253977 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0d07dee-3b03-4b75-b150-85010f0bcf29-config-data\") pod \"keystone-cron-29564461-jwfx5\" (UID: \"e0d07dee-3b03-4b75-b150-85010f0bcf29\") " pod="openstack/keystone-cron-29564461-jwfx5" Mar 18 21:01:00 crc kubenswrapper[4950]: I0318 21:01:00.254264 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0d07dee-3b03-4b75-b150-85010f0bcf29-combined-ca-bundle\") pod \"keystone-cron-29564461-jwfx5\" (UID: \"e0d07dee-3b03-4b75-b150-85010f0bcf29\") " pod="openstack/keystone-cron-29564461-jwfx5" Mar 18 21:01:00 crc kubenswrapper[4950]: I0318 21:01:00.254315 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsjbk\" (UniqueName: \"kubernetes.io/projected/e0d07dee-3b03-4b75-b150-85010f0bcf29-kube-api-access-tsjbk\") pod \"keystone-cron-29564461-jwfx5\" (UID: \"e0d07dee-3b03-4b75-b150-85010f0bcf29\") " pod="openstack/keystone-cron-29564461-jwfx5" Mar 18 21:01:00 crc kubenswrapper[4950]: I0318 21:01:00.254585 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e0d07dee-3b03-4b75-b150-85010f0bcf29-fernet-keys\") pod \"keystone-cron-29564461-jwfx5\" (UID: \"e0d07dee-3b03-4b75-b150-85010f0bcf29\") " pod="openstack/keystone-cron-29564461-jwfx5" Mar 18 21:01:00 crc kubenswrapper[4950]: I0318 21:01:00.356682 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0d07dee-3b03-4b75-b150-85010f0bcf29-config-data\") pod \"keystone-cron-29564461-jwfx5\" (UID: \"e0d07dee-3b03-4b75-b150-85010f0bcf29\") " pod="openstack/keystone-cron-29564461-jwfx5" Mar 18 21:01:00 crc kubenswrapper[4950]: I0318 21:01:00.356725 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0d07dee-3b03-4b75-b150-85010f0bcf29-combined-ca-bundle\") pod \"keystone-cron-29564461-jwfx5\" (UID: \"e0d07dee-3b03-4b75-b150-85010f0bcf29\") " pod="openstack/keystone-cron-29564461-jwfx5" Mar 18 21:01:00 crc kubenswrapper[4950]: I0318 21:01:00.356778 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsjbk\" (UniqueName: \"kubernetes.io/projected/e0d07dee-3b03-4b75-b150-85010f0bcf29-kube-api-access-tsjbk\") pod \"keystone-cron-29564461-jwfx5\" (UID: \"e0d07dee-3b03-4b75-b150-85010f0bcf29\") " pod="openstack/keystone-cron-29564461-jwfx5" Mar 18 21:01:00 crc kubenswrapper[4950]: I0318 21:01:00.356882 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e0d07dee-3b03-4b75-b150-85010f0bcf29-fernet-keys\") pod \"keystone-cron-29564461-jwfx5\" (UID: \"e0d07dee-3b03-4b75-b150-85010f0bcf29\") " pod="openstack/keystone-cron-29564461-jwfx5" Mar 18 21:01:00 crc kubenswrapper[4950]: I0318 21:01:00.366548 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0d07dee-3b03-4b75-b150-85010f0bcf29-config-data\") pod \"keystone-cron-29564461-jwfx5\" (UID: \"e0d07dee-3b03-4b75-b150-85010f0bcf29\") " pod="openstack/keystone-cron-29564461-jwfx5" Mar 18 21:01:00 crc kubenswrapper[4950]: I0318 21:01:00.367682 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e0d07dee-3b03-4b75-b150-85010f0bcf29-fernet-keys\") pod \"keystone-cron-29564461-jwfx5\" (UID: \"e0d07dee-3b03-4b75-b150-85010f0bcf29\") " pod="openstack/keystone-cron-29564461-jwfx5" Mar 18 21:01:00 crc kubenswrapper[4950]: I0318 21:01:00.368006 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0d07dee-3b03-4b75-b150-85010f0bcf29-combined-ca-bundle\") pod \"keystone-cron-29564461-jwfx5\" (UID: \"e0d07dee-3b03-4b75-b150-85010f0bcf29\") " pod="openstack/keystone-cron-29564461-jwfx5" Mar 18 21:01:00 crc kubenswrapper[4950]: I0318 21:01:00.379698 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsjbk\" (UniqueName: \"kubernetes.io/projected/e0d07dee-3b03-4b75-b150-85010f0bcf29-kube-api-access-tsjbk\") pod \"keystone-cron-29564461-jwfx5\" (UID: \"e0d07dee-3b03-4b75-b150-85010f0bcf29\") " pod="openstack/keystone-cron-29564461-jwfx5" Mar 18 21:01:00 crc kubenswrapper[4950]: I0318 21:01:00.478734 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29564461-jwfx5" Mar 18 21:01:00 crc kubenswrapper[4950]: I0318 21:01:00.819641 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Mar 18 21:01:00 crc kubenswrapper[4950]: I0318 21:01:00.942723 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29564461-jwfx5"] Mar 18 21:01:00 crc kubenswrapper[4950]: W0318 21:01:00.960144 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0d07dee_3b03_4b75_b150_85010f0bcf29.slice/crio-7589d09b8c3513040db28ab296d0056b6c7ca65bbda580d6adaee82df74787d8 WatchSource:0}: Error finding container 7589d09b8c3513040db28ab296d0056b6c7ca65bbda580d6adaee82df74787d8: Status 404 returned error can't find the container with id 7589d09b8c3513040db28ab296d0056b6c7ca65bbda580d6adaee82df74787d8 Mar 18 21:01:01 crc kubenswrapper[4950]: I0318 21:01:01.223745 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29564461-jwfx5" event={"ID":"e0d07dee-3b03-4b75-b150-85010f0bcf29","Type":"ContainerStarted","Data":"b7cc5b2159065065c6cd64a16851ecd601f43efc74816a2b8bade5a281d7c8d6"} Mar 18 21:01:01 crc kubenswrapper[4950]: I0318 21:01:01.224087 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29564461-jwfx5" event={"ID":"e0d07dee-3b03-4b75-b150-85010f0bcf29","Type":"ContainerStarted","Data":"7589d09b8c3513040db28ab296d0056b6c7ca65bbda580d6adaee82df74787d8"} Mar 18 21:01:02 crc kubenswrapper[4950]: I0318 21:01:02.990100 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.009930 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29564461-jwfx5" podStartSLOduration=3.009910623 podStartE2EDuration="3.009910623s" podCreationTimestamp="2026-03-18 21:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 21:01:01.245719824 +0000 UTC m=+3274.486561692" watchObservedRunningTime="2026-03-18 21:01:03.009910623 +0000 UTC m=+3276.250752501" Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.036254 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2rfj\" (UniqueName: \"kubernetes.io/projected/26e882de-ea5e-4a22-889d-85c5d4347b1e-kube-api-access-p2rfj\") pod \"26e882de-ea5e-4a22-889d-85c5d4347b1e\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.036328 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/26e882de-ea5e-4a22-889d-85c5d4347b1e-horizon-secret-key\") pod \"26e882de-ea5e-4a22-889d-85c5d4347b1e\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.036378 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26e882de-ea5e-4a22-889d-85c5d4347b1e-scripts\") pod \"26e882de-ea5e-4a22-889d-85c5d4347b1e\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.036631 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26e882de-ea5e-4a22-889d-85c5d4347b1e-logs\") pod \"26e882de-ea5e-4a22-889d-85c5d4347b1e\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.036658 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26e882de-ea5e-4a22-889d-85c5d4347b1e-combined-ca-bundle\") pod \"26e882de-ea5e-4a22-889d-85c5d4347b1e\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.036678 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/26e882de-ea5e-4a22-889d-85c5d4347b1e-horizon-tls-certs\") pod \"26e882de-ea5e-4a22-889d-85c5d4347b1e\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.036729 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/26e882de-ea5e-4a22-889d-85c5d4347b1e-config-data\") pod \"26e882de-ea5e-4a22-889d-85c5d4347b1e\" (UID: \"26e882de-ea5e-4a22-889d-85c5d4347b1e\") " Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.041895 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26e882de-ea5e-4a22-889d-85c5d4347b1e-logs" (OuterVolumeSpecName: "logs") pod "26e882de-ea5e-4a22-889d-85c5d4347b1e" (UID: "26e882de-ea5e-4a22-889d-85c5d4347b1e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.066772 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26e882de-ea5e-4a22-889d-85c5d4347b1e-kube-api-access-p2rfj" (OuterVolumeSpecName: "kube-api-access-p2rfj") pod "26e882de-ea5e-4a22-889d-85c5d4347b1e" (UID: "26e882de-ea5e-4a22-889d-85c5d4347b1e"). InnerVolumeSpecName "kube-api-access-p2rfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.068293 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26e882de-ea5e-4a22-889d-85c5d4347b1e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "26e882de-ea5e-4a22-889d-85c5d4347b1e" (UID: "26e882de-ea5e-4a22-889d-85c5d4347b1e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.083386 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26e882de-ea5e-4a22-889d-85c5d4347b1e-config-data" (OuterVolumeSpecName: "config-data") pod "26e882de-ea5e-4a22-889d-85c5d4347b1e" (UID: "26e882de-ea5e-4a22-889d-85c5d4347b1e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.084704 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26e882de-ea5e-4a22-889d-85c5d4347b1e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "26e882de-ea5e-4a22-889d-85c5d4347b1e" (UID: "26e882de-ea5e-4a22-889d-85c5d4347b1e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.085191 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26e882de-ea5e-4a22-889d-85c5d4347b1e-scripts" (OuterVolumeSpecName: "scripts") pod "26e882de-ea5e-4a22-889d-85c5d4347b1e" (UID: "26e882de-ea5e-4a22-889d-85c5d4347b1e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.110232 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26e882de-ea5e-4a22-889d-85c5d4347b1e-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "26e882de-ea5e-4a22-889d-85c5d4347b1e" (UID: "26e882de-ea5e-4a22-889d-85c5d4347b1e"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.138469 4950 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26e882de-ea5e-4a22-889d-85c5d4347b1e-logs\") on node \"crc\" DevicePath \"\"" Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.138501 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26e882de-ea5e-4a22-889d-85c5d4347b1e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.138511 4950 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/26e882de-ea5e-4a22-889d-85c5d4347b1e-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.138519 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/26e882de-ea5e-4a22-889d-85c5d4347b1e-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.138529 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2rfj\" (UniqueName: \"kubernetes.io/projected/26e882de-ea5e-4a22-889d-85c5d4347b1e-kube-api-access-p2rfj\") on node \"crc\" DevicePath \"\"" Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.138537 4950 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/26e882de-ea5e-4a22-889d-85c5d4347b1e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.138545 4950 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26e882de-ea5e-4a22-889d-85c5d4347b1e-scripts\") on node \"crc\" DevicePath \"\"" Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.242216 4950 generic.go:334] "Generic (PLEG): container finished" podID="26e882de-ea5e-4a22-889d-85c5d4347b1e" containerID="36339553eb5a8810d3a99d963f412a51a7f943a23f9e3e57db6fa61a732993dc" exitCode=137 Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.242294 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bfb97c7d-h56mk" Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.242322 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bfb97c7d-h56mk" event={"ID":"26e882de-ea5e-4a22-889d-85c5d4347b1e","Type":"ContainerDied","Data":"36339553eb5a8810d3a99d963f412a51a7f943a23f9e3e57db6fa61a732993dc"} Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.242759 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bfb97c7d-h56mk" event={"ID":"26e882de-ea5e-4a22-889d-85c5d4347b1e","Type":"ContainerDied","Data":"9ecdb31d73ee163b6f37e5c983320971ef50a3c1ef517e946e9560a1c748782c"} Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.242789 4950 scope.go:117] "RemoveContainer" containerID="aa50a9c7f3ed240952a5dacf482e2676dcb49617b9f6bddc43a68f1aa9938ed8" Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.280446 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5bfb97c7d-h56mk"] Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.286161 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5bfb97c7d-h56mk"] Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.410866 4950 scope.go:117] "RemoveContainer" containerID="36339553eb5a8810d3a99d963f412a51a7f943a23f9e3e57db6fa61a732993dc" Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.430084 4950 scope.go:117] "RemoveContainer" containerID="aa50a9c7f3ed240952a5dacf482e2676dcb49617b9f6bddc43a68f1aa9938ed8" Mar 18 21:01:03 crc kubenswrapper[4950]: E0318 21:01:03.430507 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa50a9c7f3ed240952a5dacf482e2676dcb49617b9f6bddc43a68f1aa9938ed8\": container with ID starting with aa50a9c7f3ed240952a5dacf482e2676dcb49617b9f6bddc43a68f1aa9938ed8 not found: ID does not exist" containerID="aa50a9c7f3ed240952a5dacf482e2676dcb49617b9f6bddc43a68f1aa9938ed8" Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.430537 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa50a9c7f3ed240952a5dacf482e2676dcb49617b9f6bddc43a68f1aa9938ed8"} err="failed to get container status \"aa50a9c7f3ed240952a5dacf482e2676dcb49617b9f6bddc43a68f1aa9938ed8\": rpc error: code = NotFound desc = could not find container \"aa50a9c7f3ed240952a5dacf482e2676dcb49617b9f6bddc43a68f1aa9938ed8\": container with ID starting with aa50a9c7f3ed240952a5dacf482e2676dcb49617b9f6bddc43a68f1aa9938ed8 not found: ID does not exist" Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.430557 4950 scope.go:117] "RemoveContainer" containerID="36339553eb5a8810d3a99d963f412a51a7f943a23f9e3e57db6fa61a732993dc" Mar 18 21:01:03 crc kubenswrapper[4950]: E0318 21:01:03.430922 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36339553eb5a8810d3a99d963f412a51a7f943a23f9e3e57db6fa61a732993dc\": container with ID starting with 36339553eb5a8810d3a99d963f412a51a7f943a23f9e3e57db6fa61a732993dc not found: ID does not exist" containerID="36339553eb5a8810d3a99d963f412a51a7f943a23f9e3e57db6fa61a732993dc" Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.430943 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36339553eb5a8810d3a99d963f412a51a7f943a23f9e3e57db6fa61a732993dc"} err="failed to get container status \"36339553eb5a8810d3a99d963f412a51a7f943a23f9e3e57db6fa61a732993dc\": rpc error: code = NotFound desc = could not find container \"36339553eb5a8810d3a99d963f412a51a7f943a23f9e3e57db6fa61a732993dc\": container with ID starting with 36339553eb5a8810d3a99d963f412a51a7f943a23f9e3e57db6fa61a732993dc not found: ID does not exist" Mar 18 21:01:03 crc kubenswrapper[4950]: I0318 21:01:03.495266 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26e882de-ea5e-4a22-889d-85c5d4347b1e" path="/var/lib/kubelet/pods/26e882de-ea5e-4a22-889d-85c5d4347b1e/volumes" Mar 18 21:01:04 crc kubenswrapper[4950]: I0318 21:01:04.096322 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Mar 18 21:01:04 crc kubenswrapper[4950]: I0318 21:01:04.253950 4950 generic.go:334] "Generic (PLEG): container finished" podID="e0d07dee-3b03-4b75-b150-85010f0bcf29" containerID="b7cc5b2159065065c6cd64a16851ecd601f43efc74816a2b8bade5a281d7c8d6" exitCode=0 Mar 18 21:01:04 crc kubenswrapper[4950]: I0318 21:01:04.254022 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29564461-jwfx5" event={"ID":"e0d07dee-3b03-4b75-b150-85010f0bcf29","Type":"ContainerDied","Data":"b7cc5b2159065065c6cd64a16851ecd601f43efc74816a2b8bade5a281d7c8d6"} Mar 18 21:01:05 crc kubenswrapper[4950]: I0318 21:01:05.723056 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29564461-jwfx5" Mar 18 21:01:05 crc kubenswrapper[4950]: I0318 21:01:05.774617 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0d07dee-3b03-4b75-b150-85010f0bcf29-combined-ca-bundle\") pod \"e0d07dee-3b03-4b75-b150-85010f0bcf29\" (UID: \"e0d07dee-3b03-4b75-b150-85010f0bcf29\") " Mar 18 21:01:05 crc kubenswrapper[4950]: I0318 21:01:05.774836 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e0d07dee-3b03-4b75-b150-85010f0bcf29-fernet-keys\") pod \"e0d07dee-3b03-4b75-b150-85010f0bcf29\" (UID: \"e0d07dee-3b03-4b75-b150-85010f0bcf29\") " Mar 18 21:01:05 crc kubenswrapper[4950]: I0318 21:01:05.774979 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsjbk\" (UniqueName: \"kubernetes.io/projected/e0d07dee-3b03-4b75-b150-85010f0bcf29-kube-api-access-tsjbk\") pod \"e0d07dee-3b03-4b75-b150-85010f0bcf29\" (UID: \"e0d07dee-3b03-4b75-b150-85010f0bcf29\") " Mar 18 21:01:05 crc kubenswrapper[4950]: I0318 21:01:05.775061 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0d07dee-3b03-4b75-b150-85010f0bcf29-config-data\") pod \"e0d07dee-3b03-4b75-b150-85010f0bcf29\" (UID: \"e0d07dee-3b03-4b75-b150-85010f0bcf29\") " Mar 18 21:01:05 crc kubenswrapper[4950]: I0318 21:01:05.781956 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0d07dee-3b03-4b75-b150-85010f0bcf29-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e0d07dee-3b03-4b75-b150-85010f0bcf29" (UID: "e0d07dee-3b03-4b75-b150-85010f0bcf29"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:01:05 crc kubenswrapper[4950]: I0318 21:01:05.782822 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0d07dee-3b03-4b75-b150-85010f0bcf29-kube-api-access-tsjbk" (OuterVolumeSpecName: "kube-api-access-tsjbk") pod "e0d07dee-3b03-4b75-b150-85010f0bcf29" (UID: "e0d07dee-3b03-4b75-b150-85010f0bcf29"). InnerVolumeSpecName "kube-api-access-tsjbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:01:05 crc kubenswrapper[4950]: I0318 21:01:05.818309 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0d07dee-3b03-4b75-b150-85010f0bcf29-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e0d07dee-3b03-4b75-b150-85010f0bcf29" (UID: "e0d07dee-3b03-4b75-b150-85010f0bcf29"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:01:05 crc kubenswrapper[4950]: I0318 21:01:05.854861 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0d07dee-3b03-4b75-b150-85010f0bcf29-config-data" (OuterVolumeSpecName: "config-data") pod "e0d07dee-3b03-4b75-b150-85010f0bcf29" (UID: "e0d07dee-3b03-4b75-b150-85010f0bcf29"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:01:05 crc kubenswrapper[4950]: I0318 21:01:05.877790 4950 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0d07dee-3b03-4b75-b150-85010f0bcf29-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 18 21:01:05 crc kubenswrapper[4950]: I0318 21:01:05.877917 4950 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e0d07dee-3b03-4b75-b150-85010f0bcf29-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 18 21:01:05 crc kubenswrapper[4950]: I0318 21:01:05.877979 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsjbk\" (UniqueName: \"kubernetes.io/projected/e0d07dee-3b03-4b75-b150-85010f0bcf29-kube-api-access-tsjbk\") on node \"crc\" DevicePath \"\"" Mar 18 21:01:05 crc kubenswrapper[4950]: I0318 21:01:05.878038 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0d07dee-3b03-4b75-b150-85010f0bcf29-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 21:01:06 crc kubenswrapper[4950]: I0318 21:01:06.288797 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29564461-jwfx5" event={"ID":"e0d07dee-3b03-4b75-b150-85010f0bcf29","Type":"ContainerDied","Data":"7589d09b8c3513040db28ab296d0056b6c7ca65bbda580d6adaee82df74787d8"} Mar 18 21:01:06 crc kubenswrapper[4950]: I0318 21:01:06.288879 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7589d09b8c3513040db28ab296d0056b6c7ca65bbda580d6adaee82df74787d8" Mar 18 21:01:06 crc kubenswrapper[4950]: I0318 21:01:06.288940 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29564461-jwfx5" Mar 18 21:01:07 crc kubenswrapper[4950]: I0318 21:01:07.490347 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:01:07 crc kubenswrapper[4950]: E0318 21:01:07.491010 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:01:12 crc kubenswrapper[4950]: I0318 21:01:12.426716 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Mar 18 21:01:15 crc kubenswrapper[4950]: I0318 21:01:15.806234 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 18 21:01:20 crc kubenswrapper[4950]: I0318 21:01:20.480137 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:01:20 crc kubenswrapper[4950]: E0318 21:01:20.480928 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:01:34 crc kubenswrapper[4950]: I0318 21:01:34.479612 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:01:34 crc kubenswrapper[4950]: E0318 21:01:34.480378 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:01:47 crc kubenswrapper[4950]: I0318 21:01:47.481938 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:01:47 crc kubenswrapper[4950]: E0318 21:01:47.483157 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:02:00 crc kubenswrapper[4950]: I0318 21:02:00.180748 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564462-ftngf"] Mar 18 21:02:00 crc kubenswrapper[4950]: E0318 21:02:00.181867 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26e882de-ea5e-4a22-889d-85c5d4347b1e" containerName="horizon" Mar 18 21:02:00 crc kubenswrapper[4950]: I0318 21:02:00.181885 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="26e882de-ea5e-4a22-889d-85c5d4347b1e" containerName="horizon" Mar 18 21:02:00 crc kubenswrapper[4950]: E0318 21:02:00.181936 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26e882de-ea5e-4a22-889d-85c5d4347b1e" containerName="horizon-log" Mar 18 21:02:00 crc kubenswrapper[4950]: I0318 21:02:00.181950 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="26e882de-ea5e-4a22-889d-85c5d4347b1e" containerName="horizon-log" Mar 18 21:02:00 crc kubenswrapper[4950]: E0318 21:02:00.181980 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0d07dee-3b03-4b75-b150-85010f0bcf29" containerName="keystone-cron" Mar 18 21:02:00 crc kubenswrapper[4950]: I0318 21:02:00.181990 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0d07dee-3b03-4b75-b150-85010f0bcf29" containerName="keystone-cron" Mar 18 21:02:00 crc kubenswrapper[4950]: I0318 21:02:00.182231 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0d07dee-3b03-4b75-b150-85010f0bcf29" containerName="keystone-cron" Mar 18 21:02:00 crc kubenswrapper[4950]: I0318 21:02:00.182263 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="26e882de-ea5e-4a22-889d-85c5d4347b1e" containerName="horizon-log" Mar 18 21:02:00 crc kubenswrapper[4950]: I0318 21:02:00.182285 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="26e882de-ea5e-4a22-889d-85c5d4347b1e" containerName="horizon" Mar 18 21:02:00 crc kubenswrapper[4950]: I0318 21:02:00.183092 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564462-ftngf" Mar 18 21:02:00 crc kubenswrapper[4950]: I0318 21:02:00.186527 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 21:02:00 crc kubenswrapper[4950]: I0318 21:02:00.186554 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 21:02:00 crc kubenswrapper[4950]: I0318 21:02:00.187151 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 21:02:00 crc kubenswrapper[4950]: I0318 21:02:00.189782 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564462-ftngf"] Mar 18 21:02:00 crc kubenswrapper[4950]: I0318 21:02:00.236906 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9b57\" (UniqueName: \"kubernetes.io/projected/74848e1c-681b-4e2f-8d52-141547d440eb-kube-api-access-w9b57\") pod \"auto-csr-approver-29564462-ftngf\" (UID: \"74848e1c-681b-4e2f-8d52-141547d440eb\") " pod="openshift-infra/auto-csr-approver-29564462-ftngf" Mar 18 21:02:00 crc kubenswrapper[4950]: I0318 21:02:00.339324 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9b57\" (UniqueName: \"kubernetes.io/projected/74848e1c-681b-4e2f-8d52-141547d440eb-kube-api-access-w9b57\") pod \"auto-csr-approver-29564462-ftngf\" (UID: \"74848e1c-681b-4e2f-8d52-141547d440eb\") " pod="openshift-infra/auto-csr-approver-29564462-ftngf" Mar 18 21:02:00 crc kubenswrapper[4950]: I0318 21:02:00.367911 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9b57\" (UniqueName: \"kubernetes.io/projected/74848e1c-681b-4e2f-8d52-141547d440eb-kube-api-access-w9b57\") pod \"auto-csr-approver-29564462-ftngf\" (UID: \"74848e1c-681b-4e2f-8d52-141547d440eb\") " pod="openshift-infra/auto-csr-approver-29564462-ftngf" Mar 18 21:02:00 crc kubenswrapper[4950]: I0318 21:02:00.508563 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564462-ftngf" Mar 18 21:02:01 crc kubenswrapper[4950]: I0318 21:02:01.010648 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564462-ftngf"] Mar 18 21:02:01 crc kubenswrapper[4950]: W0318 21:02:01.012661 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74848e1c_681b_4e2f_8d52_141547d440eb.slice/crio-15569e1f82778716af1a6902ab0fc2786a59439283de5c4e27fbee7bc0e6f73d WatchSource:0}: Error finding container 15569e1f82778716af1a6902ab0fc2786a59439283de5c4e27fbee7bc0e6f73d: Status 404 returned error can't find the container with id 15569e1f82778716af1a6902ab0fc2786a59439283de5c4e27fbee7bc0e6f73d Mar 18 21:02:01 crc kubenswrapper[4950]: I0318 21:02:01.488816 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:02:01 crc kubenswrapper[4950]: E0318 21:02:01.489228 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:02:01 crc kubenswrapper[4950]: I0318 21:02:01.852379 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564462-ftngf" event={"ID":"74848e1c-681b-4e2f-8d52-141547d440eb","Type":"ContainerStarted","Data":"15569e1f82778716af1a6902ab0fc2786a59439283de5c4e27fbee7bc0e6f73d"} Mar 18 21:02:02 crc kubenswrapper[4950]: I0318 21:02:02.860575 4950 generic.go:334] "Generic (PLEG): container finished" podID="74848e1c-681b-4e2f-8d52-141547d440eb" containerID="23a80d84c715f39a3f35546576b36927d8c0ca75ec87201176e81085fc89804a" exitCode=0 Mar 18 21:02:02 crc kubenswrapper[4950]: I0318 21:02:02.860662 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564462-ftngf" event={"ID":"74848e1c-681b-4e2f-8d52-141547d440eb","Type":"ContainerDied","Data":"23a80d84c715f39a3f35546576b36927d8c0ca75ec87201176e81085fc89804a"} Mar 18 21:02:04 crc kubenswrapper[4950]: I0318 21:02:04.257563 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564462-ftngf" Mar 18 21:02:04 crc kubenswrapper[4950]: I0318 21:02:04.359048 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9b57\" (UniqueName: \"kubernetes.io/projected/74848e1c-681b-4e2f-8d52-141547d440eb-kube-api-access-w9b57\") pod \"74848e1c-681b-4e2f-8d52-141547d440eb\" (UID: \"74848e1c-681b-4e2f-8d52-141547d440eb\") " Mar 18 21:02:04 crc kubenswrapper[4950]: I0318 21:02:04.363991 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74848e1c-681b-4e2f-8d52-141547d440eb-kube-api-access-w9b57" (OuterVolumeSpecName: "kube-api-access-w9b57") pod "74848e1c-681b-4e2f-8d52-141547d440eb" (UID: "74848e1c-681b-4e2f-8d52-141547d440eb"). InnerVolumeSpecName "kube-api-access-w9b57". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:02:04 crc kubenswrapper[4950]: I0318 21:02:04.461492 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9b57\" (UniqueName: \"kubernetes.io/projected/74848e1c-681b-4e2f-8d52-141547d440eb-kube-api-access-w9b57\") on node \"crc\" DevicePath \"\"" Mar 18 21:02:04 crc kubenswrapper[4950]: I0318 21:02:04.878007 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564462-ftngf" event={"ID":"74848e1c-681b-4e2f-8d52-141547d440eb","Type":"ContainerDied","Data":"15569e1f82778716af1a6902ab0fc2786a59439283de5c4e27fbee7bc0e6f73d"} Mar 18 21:02:04 crc kubenswrapper[4950]: I0318 21:02:04.878647 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15569e1f82778716af1a6902ab0fc2786a59439283de5c4e27fbee7bc0e6f73d" Mar 18 21:02:04 crc kubenswrapper[4950]: I0318 21:02:04.878041 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564462-ftngf" Mar 18 21:02:06 crc kubenswrapper[4950]: I0318 21:02:06.166079 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564456-mpkd8"] Mar 18 21:02:06 crc kubenswrapper[4950]: I0318 21:02:06.176938 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564456-mpkd8"] Mar 18 21:02:07 crc kubenswrapper[4950]: I0318 21:02:07.502973 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9960f16-254e-4b51-8c28-5985808c6f84" path="/var/lib/kubelet/pods/f9960f16-254e-4b51-8c28-5985808c6f84/volumes" Mar 18 21:02:12 crc kubenswrapper[4950]: I0318 21:02:12.481780 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:02:12 crc kubenswrapper[4950]: E0318 21:02:12.483129 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.184037 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Mar 18 21:02:22 crc kubenswrapper[4950]: E0318 21:02:22.185641 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74848e1c-681b-4e2f-8d52-141547d440eb" containerName="oc" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.185661 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="74848e1c-681b-4e2f-8d52-141547d440eb" containerName="oc" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.186226 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="74848e1c-681b-4e2f-8d52-141547d440eb" containerName="oc" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.187474 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.191786 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.192003 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.191808 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-z2mjq" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.192520 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.214432 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.236431 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c1bc0a9-e172-4749-b998-710b5a306618-config-data\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.236513 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3c1bc0a9-e172-4749-b998-710b5a306618-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.236575 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3c1bc0a9-e172-4749-b998-710b5a306618-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.338305 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2wxj\" (UniqueName: \"kubernetes.io/projected/3c1bc0a9-e172-4749-b998-710b5a306618-kube-api-access-b2wxj\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.338370 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3c1bc0a9-e172-4749-b998-710b5a306618-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.338426 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3c1bc0a9-e172-4749-b998-710b5a306618-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.338452 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3c1bc0a9-e172-4749-b998-710b5a306618-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.338511 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3c1bc0a9-e172-4749-b998-710b5a306618-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.338537 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c1bc0a9-e172-4749-b998-710b5a306618-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.338579 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3c1bc0a9-e172-4749-b998-710b5a306618-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.338601 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.338626 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c1bc0a9-e172-4749-b998-710b5a306618-config-data\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.339799 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c1bc0a9-e172-4749-b998-710b5a306618-config-data\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.339869 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3c1bc0a9-e172-4749-b998-710b5a306618-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.350081 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3c1bc0a9-e172-4749-b998-710b5a306618-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.441445 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.441496 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2wxj\" (UniqueName: \"kubernetes.io/projected/3c1bc0a9-e172-4749-b998-710b5a306618-kube-api-access-b2wxj\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.441557 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3c1bc0a9-e172-4749-b998-710b5a306618-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.441620 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3c1bc0a9-e172-4749-b998-710b5a306618-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.441644 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c1bc0a9-e172-4749-b998-710b5a306618-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.441687 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3c1bc0a9-e172-4749-b998-710b5a306618-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.442108 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3c1bc0a9-e172-4749-b998-710b5a306618-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.442163 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3c1bc0a9-e172-4749-b998-710b5a306618-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.442975 4950 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.447185 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c1bc0a9-e172-4749-b998-710b5a306618-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.449477 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3c1bc0a9-e172-4749-b998-710b5a306618-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.461685 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2wxj\" (UniqueName: \"kubernetes.io/projected/3c1bc0a9-e172-4749-b998-710b5a306618-kube-api-access-b2wxj\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.486553 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " pod="openstack/tempest-tests-tempest" Mar 18 21:02:22 crc kubenswrapper[4950]: I0318 21:02:22.529829 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 18 21:02:23 crc kubenswrapper[4950]: I0318 21:02:23.073806 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Mar 18 21:02:24 crc kubenswrapper[4950]: I0318 21:02:24.106349 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"3c1bc0a9-e172-4749-b998-710b5a306618","Type":"ContainerStarted","Data":"1a89b537bbaea1d20489be69908fd1cc6ee3aeb1f33ed9e0b129c450548abb7c"} Mar 18 21:02:26 crc kubenswrapper[4950]: I0318 21:02:26.484674 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:02:26 crc kubenswrapper[4950]: E0318 21:02:26.486445 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:02:40 crc kubenswrapper[4950]: I0318 21:02:40.480904 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:02:40 crc kubenswrapper[4950]: E0318 21:02:40.482303 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:02:51 crc kubenswrapper[4950]: I0318 21:02:51.947969 4950 scope.go:117] "RemoveContainer" containerID="8d96716591a3c62d75372d015cbe66a76b30ba5103cd9171fdaf45bf213ca81c" Mar 18 21:02:52 crc kubenswrapper[4950]: I0318 21:02:52.480979 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:02:52 crc kubenswrapper[4950]: E0318 21:02:52.481230 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:02:59 crc kubenswrapper[4950]: E0318 21:02:59.628811 4950 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Mar 18 21:02:59 crc kubenswrapper[4950]: E0318 21:02:59.633665 4950 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b2wxj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(3c1bc0a9-e172-4749-b998-710b5a306618): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 18 21:02:59 crc kubenswrapper[4950]: E0318 21:02:59.635100 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="3c1bc0a9-e172-4749-b998-710b5a306618" Mar 18 21:03:00 crc kubenswrapper[4950]: E0318 21:03:00.480794 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="3c1bc0a9-e172-4749-b998-710b5a306618" Mar 18 21:03:06 crc kubenswrapper[4950]: I0318 21:03:06.480637 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:03:06 crc kubenswrapper[4950]: E0318 21:03:06.481210 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:03:11 crc kubenswrapper[4950]: I0318 21:03:11.904530 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Mar 18 21:03:13 crc kubenswrapper[4950]: I0318 21:03:13.621299 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"3c1bc0a9-e172-4749-b998-710b5a306618","Type":"ContainerStarted","Data":"cef779de1979e305cdbe9749cc9f3f279e8b05db43fcbd358178c77a68d20ca7"} Mar 18 21:03:13 crc kubenswrapper[4950]: I0318 21:03:13.639874 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.808367917 podStartE2EDuration="52.639860418s" podCreationTimestamp="2026-03-18 21:02:21 +0000 UTC" firstStartedPulling="2026-03-18 21:02:23.070202003 +0000 UTC m=+3356.311043871" lastFinishedPulling="2026-03-18 21:03:11.901694504 +0000 UTC m=+3405.142536372" observedRunningTime="2026-03-18 21:03:13.636030157 +0000 UTC m=+3406.876872025" watchObservedRunningTime="2026-03-18 21:03:13.639860418 +0000 UTC m=+3406.880702286" Mar 18 21:03:17 crc kubenswrapper[4950]: I0318 21:03:17.485533 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:03:17 crc kubenswrapper[4950]: E0318 21:03:17.488271 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:03:29 crc kubenswrapper[4950]: I0318 21:03:29.480436 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:03:29 crc kubenswrapper[4950]: E0318 21:03:29.483193 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:03:43 crc kubenswrapper[4950]: I0318 21:03:43.480253 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:03:43 crc kubenswrapper[4950]: E0318 21:03:43.481288 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:03:57 crc kubenswrapper[4950]: I0318 21:03:57.488742 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:03:57 crc kubenswrapper[4950]: E0318 21:03:57.489842 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:04:00 crc kubenswrapper[4950]: I0318 21:04:00.160191 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564464-qtkkw"] Mar 18 21:04:00 crc kubenswrapper[4950]: I0318 21:04:00.161608 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564464-qtkkw" Mar 18 21:04:00 crc kubenswrapper[4950]: I0318 21:04:00.170082 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 21:04:00 crc kubenswrapper[4950]: I0318 21:04:00.170378 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 21:04:00 crc kubenswrapper[4950]: I0318 21:04:00.170904 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 21:04:00 crc kubenswrapper[4950]: I0318 21:04:00.176970 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564464-qtkkw"] Mar 18 21:04:00 crc kubenswrapper[4950]: I0318 21:04:00.271505 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbsxp\" (UniqueName: \"kubernetes.io/projected/5de9d39c-cf8e-401d-84e4-95d1bbee331f-kube-api-access-vbsxp\") pod \"auto-csr-approver-29564464-qtkkw\" (UID: \"5de9d39c-cf8e-401d-84e4-95d1bbee331f\") " pod="openshift-infra/auto-csr-approver-29564464-qtkkw" Mar 18 21:04:00 crc kubenswrapper[4950]: I0318 21:04:00.374146 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbsxp\" (UniqueName: \"kubernetes.io/projected/5de9d39c-cf8e-401d-84e4-95d1bbee331f-kube-api-access-vbsxp\") pod \"auto-csr-approver-29564464-qtkkw\" (UID: \"5de9d39c-cf8e-401d-84e4-95d1bbee331f\") " pod="openshift-infra/auto-csr-approver-29564464-qtkkw" Mar 18 21:04:00 crc kubenswrapper[4950]: I0318 21:04:00.395787 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbsxp\" (UniqueName: \"kubernetes.io/projected/5de9d39c-cf8e-401d-84e4-95d1bbee331f-kube-api-access-vbsxp\") pod \"auto-csr-approver-29564464-qtkkw\" (UID: \"5de9d39c-cf8e-401d-84e4-95d1bbee331f\") " pod="openshift-infra/auto-csr-approver-29564464-qtkkw" Mar 18 21:04:00 crc kubenswrapper[4950]: I0318 21:04:00.483828 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564464-qtkkw" Mar 18 21:04:01 crc kubenswrapper[4950]: I0318 21:04:01.003691 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564464-qtkkw"] Mar 18 21:04:01 crc kubenswrapper[4950]: I0318 21:04:01.082772 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564464-qtkkw" event={"ID":"5de9d39c-cf8e-401d-84e4-95d1bbee331f","Type":"ContainerStarted","Data":"09b69be97ead13c496d614588bb7600db5efcd1e425cbab4855823dd9bfb1b76"} Mar 18 21:04:03 crc kubenswrapper[4950]: I0318 21:04:03.102045 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564464-qtkkw" event={"ID":"5de9d39c-cf8e-401d-84e4-95d1bbee331f","Type":"ContainerStarted","Data":"ecbd941aa773430e85ac00d884f2e32bd0a9f5c3dd92723c7c236aac7e3125a7"} Mar 18 21:04:03 crc kubenswrapper[4950]: I0318 21:04:03.115270 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29564464-qtkkw" podStartSLOduration=1.64007814 podStartE2EDuration="3.115250537s" podCreationTimestamp="2026-03-18 21:04:00 +0000 UTC" firstStartedPulling="2026-03-18 21:04:01.013549674 +0000 UTC m=+3454.254391582" lastFinishedPulling="2026-03-18 21:04:02.488722101 +0000 UTC m=+3455.729563979" observedRunningTime="2026-03-18 21:04:03.113168202 +0000 UTC m=+3456.354010070" watchObservedRunningTime="2026-03-18 21:04:03.115250537 +0000 UTC m=+3456.356092405" Mar 18 21:04:05 crc kubenswrapper[4950]: I0318 21:04:05.151189 4950 generic.go:334] "Generic (PLEG): container finished" podID="5de9d39c-cf8e-401d-84e4-95d1bbee331f" containerID="ecbd941aa773430e85ac00d884f2e32bd0a9f5c3dd92723c7c236aac7e3125a7" exitCode=0 Mar 18 21:04:05 crc kubenswrapper[4950]: I0318 21:04:05.151595 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564464-qtkkw" event={"ID":"5de9d39c-cf8e-401d-84e4-95d1bbee331f","Type":"ContainerDied","Data":"ecbd941aa773430e85ac00d884f2e32bd0a9f5c3dd92723c7c236aac7e3125a7"} Mar 18 21:04:06 crc kubenswrapper[4950]: I0318 21:04:06.694066 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564464-qtkkw" Mar 18 21:04:06 crc kubenswrapper[4950]: I0318 21:04:06.866262 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbsxp\" (UniqueName: \"kubernetes.io/projected/5de9d39c-cf8e-401d-84e4-95d1bbee331f-kube-api-access-vbsxp\") pod \"5de9d39c-cf8e-401d-84e4-95d1bbee331f\" (UID: \"5de9d39c-cf8e-401d-84e4-95d1bbee331f\") " Mar 18 21:04:06 crc kubenswrapper[4950]: I0318 21:04:06.872824 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5de9d39c-cf8e-401d-84e4-95d1bbee331f-kube-api-access-vbsxp" (OuterVolumeSpecName: "kube-api-access-vbsxp") pod "5de9d39c-cf8e-401d-84e4-95d1bbee331f" (UID: "5de9d39c-cf8e-401d-84e4-95d1bbee331f"). InnerVolumeSpecName "kube-api-access-vbsxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:04:06 crc kubenswrapper[4950]: I0318 21:04:06.968899 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbsxp\" (UniqueName: \"kubernetes.io/projected/5de9d39c-cf8e-401d-84e4-95d1bbee331f-kube-api-access-vbsxp\") on node \"crc\" DevicePath \"\"" Mar 18 21:04:07 crc kubenswrapper[4950]: I0318 21:04:07.169512 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564464-qtkkw" event={"ID":"5de9d39c-cf8e-401d-84e4-95d1bbee331f","Type":"ContainerDied","Data":"09b69be97ead13c496d614588bb7600db5efcd1e425cbab4855823dd9bfb1b76"} Mar 18 21:04:07 crc kubenswrapper[4950]: I0318 21:04:07.169556 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09b69be97ead13c496d614588bb7600db5efcd1e425cbab4855823dd9bfb1b76" Mar 18 21:04:07 crc kubenswrapper[4950]: I0318 21:04:07.169599 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564464-qtkkw" Mar 18 21:04:07 crc kubenswrapper[4950]: I0318 21:04:07.266753 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564458-fpv4w"] Mar 18 21:04:07 crc kubenswrapper[4950]: I0318 21:04:07.277786 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564458-fpv4w"] Mar 18 21:04:07 crc kubenswrapper[4950]: I0318 21:04:07.490492 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdf6ffec-9a44-461d-b30c-d59e5ad41801" path="/var/lib/kubelet/pods/fdf6ffec-9a44-461d-b30c-d59e5ad41801/volumes" Mar 18 21:04:11 crc kubenswrapper[4950]: I0318 21:04:11.483230 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:04:11 crc kubenswrapper[4950]: E0318 21:04:11.484290 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:04:25 crc kubenswrapper[4950]: I0318 21:04:25.480298 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:04:25 crc kubenswrapper[4950]: E0318 21:04:25.481803 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:04:40 crc kubenswrapper[4950]: I0318 21:04:40.480192 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:04:40 crc kubenswrapper[4950]: E0318 21:04:40.481223 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:04:54 crc kubenswrapper[4950]: I0318 21:04:54.480249 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:04:54 crc kubenswrapper[4950]: E0318 21:04:54.481038 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:04:59 crc kubenswrapper[4950]: I0318 21:04:59.686929 4950 scope.go:117] "RemoveContainer" containerID="af2bfc4c6a4a4d1c3d66471c2ce354b5460e43bab98d08892df2a6d3f639eabf" Mar 18 21:05:05 crc kubenswrapper[4950]: I0318 21:05:05.480343 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:05:05 crc kubenswrapper[4950]: I0318 21:05:05.726136 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerStarted","Data":"16b4cab9ed932e2b4b0bb08ce747a1dbd93ab41574373696a03e18ef0ef41c79"} Mar 18 21:05:49 crc kubenswrapper[4950]: I0318 21:05:49.763362 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-r57t6"] Mar 18 21:05:49 crc kubenswrapper[4950]: E0318 21:05:49.764468 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5de9d39c-cf8e-401d-84e4-95d1bbee331f" containerName="oc" Mar 18 21:05:49 crc kubenswrapper[4950]: I0318 21:05:49.764487 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="5de9d39c-cf8e-401d-84e4-95d1bbee331f" containerName="oc" Mar 18 21:05:49 crc kubenswrapper[4950]: I0318 21:05:49.764746 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="5de9d39c-cf8e-401d-84e4-95d1bbee331f" containerName="oc" Mar 18 21:05:49 crc kubenswrapper[4950]: I0318 21:05:49.766385 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r57t6" Mar 18 21:05:49 crc kubenswrapper[4950]: I0318 21:05:49.780552 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r57t6"] Mar 18 21:05:49 crc kubenswrapper[4950]: I0318 21:05:49.811396 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1154298b-5139-494c-9ab1-982eff50806b-catalog-content\") pod \"redhat-marketplace-r57t6\" (UID: \"1154298b-5139-494c-9ab1-982eff50806b\") " pod="openshift-marketplace/redhat-marketplace-r57t6" Mar 18 21:05:49 crc kubenswrapper[4950]: I0318 21:05:49.811557 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1154298b-5139-494c-9ab1-982eff50806b-utilities\") pod \"redhat-marketplace-r57t6\" (UID: \"1154298b-5139-494c-9ab1-982eff50806b\") " pod="openshift-marketplace/redhat-marketplace-r57t6" Mar 18 21:05:49 crc kubenswrapper[4950]: I0318 21:05:49.811602 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2rgr\" (UniqueName: \"kubernetes.io/projected/1154298b-5139-494c-9ab1-982eff50806b-kube-api-access-n2rgr\") pod \"redhat-marketplace-r57t6\" (UID: \"1154298b-5139-494c-9ab1-982eff50806b\") " pod="openshift-marketplace/redhat-marketplace-r57t6" Mar 18 21:05:49 crc kubenswrapper[4950]: I0318 21:05:49.913219 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1154298b-5139-494c-9ab1-982eff50806b-catalog-content\") pod \"redhat-marketplace-r57t6\" (UID: \"1154298b-5139-494c-9ab1-982eff50806b\") " pod="openshift-marketplace/redhat-marketplace-r57t6" Mar 18 21:05:49 crc kubenswrapper[4950]: I0318 21:05:49.913441 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1154298b-5139-494c-9ab1-982eff50806b-utilities\") pod \"redhat-marketplace-r57t6\" (UID: \"1154298b-5139-494c-9ab1-982eff50806b\") " pod="openshift-marketplace/redhat-marketplace-r57t6" Mar 18 21:05:49 crc kubenswrapper[4950]: I0318 21:05:49.913503 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2rgr\" (UniqueName: \"kubernetes.io/projected/1154298b-5139-494c-9ab1-982eff50806b-kube-api-access-n2rgr\") pod \"redhat-marketplace-r57t6\" (UID: \"1154298b-5139-494c-9ab1-982eff50806b\") " pod="openshift-marketplace/redhat-marketplace-r57t6" Mar 18 21:05:49 crc kubenswrapper[4950]: I0318 21:05:49.914113 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1154298b-5139-494c-9ab1-982eff50806b-utilities\") pod \"redhat-marketplace-r57t6\" (UID: \"1154298b-5139-494c-9ab1-982eff50806b\") " pod="openshift-marketplace/redhat-marketplace-r57t6" Mar 18 21:05:49 crc kubenswrapper[4950]: I0318 21:05:49.914278 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1154298b-5139-494c-9ab1-982eff50806b-catalog-content\") pod \"redhat-marketplace-r57t6\" (UID: \"1154298b-5139-494c-9ab1-982eff50806b\") " pod="openshift-marketplace/redhat-marketplace-r57t6" Mar 18 21:05:49 crc kubenswrapper[4950]: I0318 21:05:49.932482 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2rgr\" (UniqueName: \"kubernetes.io/projected/1154298b-5139-494c-9ab1-982eff50806b-kube-api-access-n2rgr\") pod \"redhat-marketplace-r57t6\" (UID: \"1154298b-5139-494c-9ab1-982eff50806b\") " pod="openshift-marketplace/redhat-marketplace-r57t6" Mar 18 21:05:50 crc kubenswrapper[4950]: I0318 21:05:50.087499 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r57t6" Mar 18 21:05:50 crc kubenswrapper[4950]: I0318 21:05:50.578379 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r57t6"] Mar 18 21:05:51 crc kubenswrapper[4950]: I0318 21:05:51.173046 4950 generic.go:334] "Generic (PLEG): container finished" podID="1154298b-5139-494c-9ab1-982eff50806b" containerID="a252a8fff4b4c5ca69efaeb77df0543df05316775248083a605497f65c2479c7" exitCode=0 Mar 18 21:05:51 crc kubenswrapper[4950]: I0318 21:05:51.174331 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r57t6" event={"ID":"1154298b-5139-494c-9ab1-982eff50806b","Type":"ContainerDied","Data":"a252a8fff4b4c5ca69efaeb77df0543df05316775248083a605497f65c2479c7"} Mar 18 21:05:51 crc kubenswrapper[4950]: I0318 21:05:51.174489 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r57t6" event={"ID":"1154298b-5139-494c-9ab1-982eff50806b","Type":"ContainerStarted","Data":"65d19bd1cc0ab8d899a5e8c13ff8d103da5f7abcbb846c8ed033faff189543fa"} Mar 18 21:05:51 crc kubenswrapper[4950]: I0318 21:05:51.179818 4950 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 21:05:52 crc kubenswrapper[4950]: I0318 21:05:52.187669 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r57t6" event={"ID":"1154298b-5139-494c-9ab1-982eff50806b","Type":"ContainerStarted","Data":"9d400774b9f0dcb2ae26571c6cdb2b791cd0ae60e32d5ae8d07757ce094b629e"} Mar 18 21:05:53 crc kubenswrapper[4950]: I0318 21:05:53.198445 4950 generic.go:334] "Generic (PLEG): container finished" podID="1154298b-5139-494c-9ab1-982eff50806b" containerID="9d400774b9f0dcb2ae26571c6cdb2b791cd0ae60e32d5ae8d07757ce094b629e" exitCode=0 Mar 18 21:05:53 crc kubenswrapper[4950]: I0318 21:05:53.198508 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r57t6" event={"ID":"1154298b-5139-494c-9ab1-982eff50806b","Type":"ContainerDied","Data":"9d400774b9f0dcb2ae26571c6cdb2b791cd0ae60e32d5ae8d07757ce094b629e"} Mar 18 21:05:54 crc kubenswrapper[4950]: I0318 21:05:54.212010 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r57t6" event={"ID":"1154298b-5139-494c-9ab1-982eff50806b","Type":"ContainerStarted","Data":"3a1e50952ac6f5d5b05ab8d7981e0c236178f105722e5505ef085a4073f15406"} Mar 18 21:05:54 crc kubenswrapper[4950]: I0318 21:05:54.236838 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-r57t6" podStartSLOduration=2.717161801 podStartE2EDuration="5.236812418s" podCreationTimestamp="2026-03-18 21:05:49 +0000 UTC" firstStartedPulling="2026-03-18 21:05:51.176802236 +0000 UTC m=+3564.417644114" lastFinishedPulling="2026-03-18 21:05:53.696452863 +0000 UTC m=+3566.937294731" observedRunningTime="2026-03-18 21:05:54.229336462 +0000 UTC m=+3567.470178330" watchObservedRunningTime="2026-03-18 21:05:54.236812418 +0000 UTC m=+3567.477654286" Mar 18 21:05:59 crc kubenswrapper[4950]: I0318 21:05:59.796156 4950 scope.go:117] "RemoveContainer" containerID="90a9304b3a5865bdd4309ff5762b12b95017e024bb9737ef4071bc8e4dfd75b5" Mar 18 21:05:59 crc kubenswrapper[4950]: I0318 21:05:59.831070 4950 scope.go:117] "RemoveContainer" containerID="0106c60c4df34db29fa73a405c56841abaf2806b56b6495af29c200350a42839" Mar 18 21:06:00 crc kubenswrapper[4950]: I0318 21:06:00.087685 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-r57t6" Mar 18 21:06:00 crc kubenswrapper[4950]: I0318 21:06:00.087790 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-r57t6" Mar 18 21:06:00 crc kubenswrapper[4950]: I0318 21:06:00.143280 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-r57t6" Mar 18 21:06:00 crc kubenswrapper[4950]: I0318 21:06:00.151053 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564466-vhzmh"] Mar 18 21:06:00 crc kubenswrapper[4950]: I0318 21:06:00.152625 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564466-vhzmh" Mar 18 21:06:00 crc kubenswrapper[4950]: I0318 21:06:00.154634 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 21:06:00 crc kubenswrapper[4950]: I0318 21:06:00.154798 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 21:06:00 crc kubenswrapper[4950]: I0318 21:06:00.155067 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 21:06:00 crc kubenswrapper[4950]: I0318 21:06:00.174017 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564466-vhzmh"] Mar 18 21:06:00 crc kubenswrapper[4950]: I0318 21:06:00.258723 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkzw9\" (UniqueName: \"kubernetes.io/projected/218bd624-88ee-45a0-8d81-756534613103-kube-api-access-jkzw9\") pod \"auto-csr-approver-29564466-vhzmh\" (UID: \"218bd624-88ee-45a0-8d81-756534613103\") " pod="openshift-infra/auto-csr-approver-29564466-vhzmh" Mar 18 21:06:00 crc kubenswrapper[4950]: I0318 21:06:00.331355 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-r57t6" Mar 18 21:06:00 crc kubenswrapper[4950]: I0318 21:06:00.361154 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkzw9\" (UniqueName: \"kubernetes.io/projected/218bd624-88ee-45a0-8d81-756534613103-kube-api-access-jkzw9\") pod \"auto-csr-approver-29564466-vhzmh\" (UID: \"218bd624-88ee-45a0-8d81-756534613103\") " pod="openshift-infra/auto-csr-approver-29564466-vhzmh" Mar 18 21:06:00 crc kubenswrapper[4950]: I0318 21:06:00.381491 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkzw9\" (UniqueName: \"kubernetes.io/projected/218bd624-88ee-45a0-8d81-756534613103-kube-api-access-jkzw9\") pod \"auto-csr-approver-29564466-vhzmh\" (UID: \"218bd624-88ee-45a0-8d81-756534613103\") " pod="openshift-infra/auto-csr-approver-29564466-vhzmh" Mar 18 21:06:00 crc kubenswrapper[4950]: I0318 21:06:00.399242 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r57t6"] Mar 18 21:06:00 crc kubenswrapper[4950]: I0318 21:06:00.484481 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564466-vhzmh" Mar 18 21:06:00 crc kubenswrapper[4950]: I0318 21:06:00.935561 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564466-vhzmh"] Mar 18 21:06:01 crc kubenswrapper[4950]: I0318 21:06:01.297477 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564466-vhzmh" event={"ID":"218bd624-88ee-45a0-8d81-756534613103","Type":"ContainerStarted","Data":"d4f2d9fd3b63c5d98aae16d2e4591395a2e1dda29dfa90361ef1a9fdefec54fd"} Mar 18 21:06:02 crc kubenswrapper[4950]: I0318 21:06:02.304503 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-r57t6" podUID="1154298b-5139-494c-9ab1-982eff50806b" containerName="registry-server" containerID="cri-o://3a1e50952ac6f5d5b05ab8d7981e0c236178f105722e5505ef085a4073f15406" gracePeriod=2 Mar 18 21:06:02 crc kubenswrapper[4950]: I0318 21:06:02.830216 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r57t6" Mar 18 21:06:02 crc kubenswrapper[4950]: I0318 21:06:02.925458 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2rgr\" (UniqueName: \"kubernetes.io/projected/1154298b-5139-494c-9ab1-982eff50806b-kube-api-access-n2rgr\") pod \"1154298b-5139-494c-9ab1-982eff50806b\" (UID: \"1154298b-5139-494c-9ab1-982eff50806b\") " Mar 18 21:06:02 crc kubenswrapper[4950]: I0318 21:06:02.925705 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1154298b-5139-494c-9ab1-982eff50806b-utilities\") pod \"1154298b-5139-494c-9ab1-982eff50806b\" (UID: \"1154298b-5139-494c-9ab1-982eff50806b\") " Mar 18 21:06:02 crc kubenswrapper[4950]: I0318 21:06:02.925741 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1154298b-5139-494c-9ab1-982eff50806b-catalog-content\") pod \"1154298b-5139-494c-9ab1-982eff50806b\" (UID: \"1154298b-5139-494c-9ab1-982eff50806b\") " Mar 18 21:06:02 crc kubenswrapper[4950]: I0318 21:06:02.926905 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1154298b-5139-494c-9ab1-982eff50806b-utilities" (OuterVolumeSpecName: "utilities") pod "1154298b-5139-494c-9ab1-982eff50806b" (UID: "1154298b-5139-494c-9ab1-982eff50806b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:06:02 crc kubenswrapper[4950]: I0318 21:06:02.930993 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1154298b-5139-494c-9ab1-982eff50806b-kube-api-access-n2rgr" (OuterVolumeSpecName: "kube-api-access-n2rgr") pod "1154298b-5139-494c-9ab1-982eff50806b" (UID: "1154298b-5139-494c-9ab1-982eff50806b"). InnerVolumeSpecName "kube-api-access-n2rgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:06:02 crc kubenswrapper[4950]: I0318 21:06:02.945364 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1154298b-5139-494c-9ab1-982eff50806b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1154298b-5139-494c-9ab1-982eff50806b" (UID: "1154298b-5139-494c-9ab1-982eff50806b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:06:03 crc kubenswrapper[4950]: I0318 21:06:03.027600 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1154298b-5139-494c-9ab1-982eff50806b-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 21:06:03 crc kubenswrapper[4950]: I0318 21:06:03.027633 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1154298b-5139-494c-9ab1-982eff50806b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 21:06:03 crc kubenswrapper[4950]: I0318 21:06:03.027646 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2rgr\" (UniqueName: \"kubernetes.io/projected/1154298b-5139-494c-9ab1-982eff50806b-kube-api-access-n2rgr\") on node \"crc\" DevicePath \"\"" Mar 18 21:06:03 crc kubenswrapper[4950]: I0318 21:06:03.317498 4950 generic.go:334] "Generic (PLEG): container finished" podID="218bd624-88ee-45a0-8d81-756534613103" containerID="c6d2053d2c61428bc84b0fce73ba188ed92f3c0a4dfe1c721f4077c524d9abf5" exitCode=0 Mar 18 21:06:03 crc kubenswrapper[4950]: I0318 21:06:03.317620 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564466-vhzmh" event={"ID":"218bd624-88ee-45a0-8d81-756534613103","Type":"ContainerDied","Data":"c6d2053d2c61428bc84b0fce73ba188ed92f3c0a4dfe1c721f4077c524d9abf5"} Mar 18 21:06:03 crc kubenswrapper[4950]: I0318 21:06:03.321664 4950 generic.go:334] "Generic (PLEG): container finished" podID="1154298b-5139-494c-9ab1-982eff50806b" containerID="3a1e50952ac6f5d5b05ab8d7981e0c236178f105722e5505ef085a4073f15406" exitCode=0 Mar 18 21:06:03 crc kubenswrapper[4950]: I0318 21:06:03.321733 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r57t6" event={"ID":"1154298b-5139-494c-9ab1-982eff50806b","Type":"ContainerDied","Data":"3a1e50952ac6f5d5b05ab8d7981e0c236178f105722e5505ef085a4073f15406"} Mar 18 21:06:03 crc kubenswrapper[4950]: I0318 21:06:03.321752 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r57t6" Mar 18 21:06:03 crc kubenswrapper[4950]: I0318 21:06:03.321827 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r57t6" event={"ID":"1154298b-5139-494c-9ab1-982eff50806b","Type":"ContainerDied","Data":"65d19bd1cc0ab8d899a5e8c13ff8d103da5f7abcbb846c8ed033faff189543fa"} Mar 18 21:06:03 crc kubenswrapper[4950]: I0318 21:06:03.321916 4950 scope.go:117] "RemoveContainer" containerID="3a1e50952ac6f5d5b05ab8d7981e0c236178f105722e5505ef085a4073f15406" Mar 18 21:06:03 crc kubenswrapper[4950]: I0318 21:06:03.371176 4950 scope.go:117] "RemoveContainer" containerID="9d400774b9f0dcb2ae26571c6cdb2b791cd0ae60e32d5ae8d07757ce094b629e" Mar 18 21:06:03 crc kubenswrapper[4950]: I0318 21:06:03.381242 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r57t6"] Mar 18 21:06:03 crc kubenswrapper[4950]: I0318 21:06:03.392629 4950 scope.go:117] "RemoveContainer" containerID="a252a8fff4b4c5ca69efaeb77df0543df05316775248083a605497f65c2479c7" Mar 18 21:06:03 crc kubenswrapper[4950]: I0318 21:06:03.399492 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-r57t6"] Mar 18 21:06:03 crc kubenswrapper[4950]: I0318 21:06:03.435226 4950 scope.go:117] "RemoveContainer" containerID="3a1e50952ac6f5d5b05ab8d7981e0c236178f105722e5505ef085a4073f15406" Mar 18 21:06:03 crc kubenswrapper[4950]: E0318 21:06:03.438142 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a1e50952ac6f5d5b05ab8d7981e0c236178f105722e5505ef085a4073f15406\": container with ID starting with 3a1e50952ac6f5d5b05ab8d7981e0c236178f105722e5505ef085a4073f15406 not found: ID does not exist" containerID="3a1e50952ac6f5d5b05ab8d7981e0c236178f105722e5505ef085a4073f15406" Mar 18 21:06:03 crc kubenswrapper[4950]: I0318 21:06:03.438191 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a1e50952ac6f5d5b05ab8d7981e0c236178f105722e5505ef085a4073f15406"} err="failed to get container status \"3a1e50952ac6f5d5b05ab8d7981e0c236178f105722e5505ef085a4073f15406\": rpc error: code = NotFound desc = could not find container \"3a1e50952ac6f5d5b05ab8d7981e0c236178f105722e5505ef085a4073f15406\": container with ID starting with 3a1e50952ac6f5d5b05ab8d7981e0c236178f105722e5505ef085a4073f15406 not found: ID does not exist" Mar 18 21:06:03 crc kubenswrapper[4950]: I0318 21:06:03.438224 4950 scope.go:117] "RemoveContainer" containerID="9d400774b9f0dcb2ae26571c6cdb2b791cd0ae60e32d5ae8d07757ce094b629e" Mar 18 21:06:03 crc kubenswrapper[4950]: E0318 21:06:03.438614 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d400774b9f0dcb2ae26571c6cdb2b791cd0ae60e32d5ae8d07757ce094b629e\": container with ID starting with 9d400774b9f0dcb2ae26571c6cdb2b791cd0ae60e32d5ae8d07757ce094b629e not found: ID does not exist" containerID="9d400774b9f0dcb2ae26571c6cdb2b791cd0ae60e32d5ae8d07757ce094b629e" Mar 18 21:06:03 crc kubenswrapper[4950]: I0318 21:06:03.438651 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d400774b9f0dcb2ae26571c6cdb2b791cd0ae60e32d5ae8d07757ce094b629e"} err="failed to get container status \"9d400774b9f0dcb2ae26571c6cdb2b791cd0ae60e32d5ae8d07757ce094b629e\": rpc error: code = NotFound desc = could not find container \"9d400774b9f0dcb2ae26571c6cdb2b791cd0ae60e32d5ae8d07757ce094b629e\": container with ID starting with 9d400774b9f0dcb2ae26571c6cdb2b791cd0ae60e32d5ae8d07757ce094b629e not found: ID does not exist" Mar 18 21:06:03 crc kubenswrapper[4950]: I0318 21:06:03.438674 4950 scope.go:117] "RemoveContainer" containerID="a252a8fff4b4c5ca69efaeb77df0543df05316775248083a605497f65c2479c7" Mar 18 21:06:03 crc kubenswrapper[4950]: E0318 21:06:03.439338 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a252a8fff4b4c5ca69efaeb77df0543df05316775248083a605497f65c2479c7\": container with ID starting with a252a8fff4b4c5ca69efaeb77df0543df05316775248083a605497f65c2479c7 not found: ID does not exist" containerID="a252a8fff4b4c5ca69efaeb77df0543df05316775248083a605497f65c2479c7" Mar 18 21:06:03 crc kubenswrapper[4950]: I0318 21:06:03.439364 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a252a8fff4b4c5ca69efaeb77df0543df05316775248083a605497f65c2479c7"} err="failed to get container status \"a252a8fff4b4c5ca69efaeb77df0543df05316775248083a605497f65c2479c7\": rpc error: code = NotFound desc = could not find container \"a252a8fff4b4c5ca69efaeb77df0543df05316775248083a605497f65c2479c7\": container with ID starting with a252a8fff4b4c5ca69efaeb77df0543df05316775248083a605497f65c2479c7 not found: ID does not exist" Mar 18 21:06:03 crc kubenswrapper[4950]: I0318 21:06:03.492578 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1154298b-5139-494c-9ab1-982eff50806b" path="/var/lib/kubelet/pods/1154298b-5139-494c-9ab1-982eff50806b/volumes" Mar 18 21:06:04 crc kubenswrapper[4950]: I0318 21:06:04.677838 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564466-vhzmh" Mar 18 21:06:04 crc kubenswrapper[4950]: I0318 21:06:04.758095 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkzw9\" (UniqueName: \"kubernetes.io/projected/218bd624-88ee-45a0-8d81-756534613103-kube-api-access-jkzw9\") pod \"218bd624-88ee-45a0-8d81-756534613103\" (UID: \"218bd624-88ee-45a0-8d81-756534613103\") " Mar 18 21:06:04 crc kubenswrapper[4950]: I0318 21:06:04.764343 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/218bd624-88ee-45a0-8d81-756534613103-kube-api-access-jkzw9" (OuterVolumeSpecName: "kube-api-access-jkzw9") pod "218bd624-88ee-45a0-8d81-756534613103" (UID: "218bd624-88ee-45a0-8d81-756534613103"). InnerVolumeSpecName "kube-api-access-jkzw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:06:04 crc kubenswrapper[4950]: I0318 21:06:04.860953 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkzw9\" (UniqueName: \"kubernetes.io/projected/218bd624-88ee-45a0-8d81-756534613103-kube-api-access-jkzw9\") on node \"crc\" DevicePath \"\"" Mar 18 21:06:05 crc kubenswrapper[4950]: I0318 21:06:05.343648 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564466-vhzmh" event={"ID":"218bd624-88ee-45a0-8d81-756534613103","Type":"ContainerDied","Data":"d4f2d9fd3b63c5d98aae16d2e4591395a2e1dda29dfa90361ef1a9fdefec54fd"} Mar 18 21:06:05 crc kubenswrapper[4950]: I0318 21:06:05.343755 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4f2d9fd3b63c5d98aae16d2e4591395a2e1dda29dfa90361ef1a9fdefec54fd" Mar 18 21:06:05 crc kubenswrapper[4950]: I0318 21:06:05.343707 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564466-vhzmh" Mar 18 21:06:05 crc kubenswrapper[4950]: I0318 21:06:05.780590 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564460-krxkx"] Mar 18 21:06:05 crc kubenswrapper[4950]: I0318 21:06:05.796516 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564460-krxkx"] Mar 18 21:06:07 crc kubenswrapper[4950]: I0318 21:06:07.492716 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="703737d3-cdc5-4db1-8b1f-2714f5261d4e" path="/var/lib/kubelet/pods/703737d3-cdc5-4db1-8b1f-2714f5261d4e/volumes" Mar 18 21:06:27 crc kubenswrapper[4950]: I0318 21:06:27.822185 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-j2ksp"] Mar 18 21:06:27 crc kubenswrapper[4950]: E0318 21:06:27.824834 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1154298b-5139-494c-9ab1-982eff50806b" containerName="registry-server" Mar 18 21:06:27 crc kubenswrapper[4950]: I0318 21:06:27.824958 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="1154298b-5139-494c-9ab1-982eff50806b" containerName="registry-server" Mar 18 21:06:27 crc kubenswrapper[4950]: E0318 21:06:27.825080 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="218bd624-88ee-45a0-8d81-756534613103" containerName="oc" Mar 18 21:06:27 crc kubenswrapper[4950]: I0318 21:06:27.825170 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="218bd624-88ee-45a0-8d81-756534613103" containerName="oc" Mar 18 21:06:27 crc kubenswrapper[4950]: E0318 21:06:27.825261 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1154298b-5139-494c-9ab1-982eff50806b" containerName="extract-content" Mar 18 21:06:27 crc kubenswrapper[4950]: I0318 21:06:27.825340 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="1154298b-5139-494c-9ab1-982eff50806b" containerName="extract-content" Mar 18 21:06:27 crc kubenswrapper[4950]: E0318 21:06:27.825467 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1154298b-5139-494c-9ab1-982eff50806b" containerName="extract-utilities" Mar 18 21:06:27 crc kubenswrapper[4950]: I0318 21:06:27.825560 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="1154298b-5139-494c-9ab1-982eff50806b" containerName="extract-utilities" Mar 18 21:06:27 crc kubenswrapper[4950]: I0318 21:06:27.825868 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="218bd624-88ee-45a0-8d81-756534613103" containerName="oc" Mar 18 21:06:27 crc kubenswrapper[4950]: I0318 21:06:27.825979 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="1154298b-5139-494c-9ab1-982eff50806b" containerName="registry-server" Mar 18 21:06:27 crc kubenswrapper[4950]: I0318 21:06:27.827758 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j2ksp" Mar 18 21:06:27 crc kubenswrapper[4950]: I0318 21:06:27.871372 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j2ksp"] Mar 18 21:06:27 crc kubenswrapper[4950]: I0318 21:06:27.958275 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd-catalog-content\") pod \"community-operators-j2ksp\" (UID: \"8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd\") " pod="openshift-marketplace/community-operators-j2ksp" Mar 18 21:06:27 crc kubenswrapper[4950]: I0318 21:06:27.958588 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fbt6\" (UniqueName: \"kubernetes.io/projected/8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd-kube-api-access-2fbt6\") pod \"community-operators-j2ksp\" (UID: \"8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd\") " pod="openshift-marketplace/community-operators-j2ksp" Mar 18 21:06:27 crc kubenswrapper[4950]: I0318 21:06:27.958730 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd-utilities\") pod \"community-operators-j2ksp\" (UID: \"8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd\") " pod="openshift-marketplace/community-operators-j2ksp" Mar 18 21:06:28 crc kubenswrapper[4950]: I0318 21:06:28.060673 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fbt6\" (UniqueName: \"kubernetes.io/projected/8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd-kube-api-access-2fbt6\") pod \"community-operators-j2ksp\" (UID: \"8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd\") " pod="openshift-marketplace/community-operators-j2ksp" Mar 18 21:06:28 crc kubenswrapper[4950]: I0318 21:06:28.061106 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd-utilities\") pod \"community-operators-j2ksp\" (UID: \"8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd\") " pod="openshift-marketplace/community-operators-j2ksp" Mar 18 21:06:28 crc kubenswrapper[4950]: I0318 21:06:28.061382 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd-catalog-content\") pod \"community-operators-j2ksp\" (UID: \"8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd\") " pod="openshift-marketplace/community-operators-j2ksp" Mar 18 21:06:28 crc kubenswrapper[4950]: I0318 21:06:28.061623 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd-utilities\") pod \"community-operators-j2ksp\" (UID: \"8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd\") " pod="openshift-marketplace/community-operators-j2ksp" Mar 18 21:06:28 crc kubenswrapper[4950]: I0318 21:06:28.061868 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd-catalog-content\") pod \"community-operators-j2ksp\" (UID: \"8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd\") " pod="openshift-marketplace/community-operators-j2ksp" Mar 18 21:06:28 crc kubenswrapper[4950]: I0318 21:06:28.092516 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fbt6\" (UniqueName: \"kubernetes.io/projected/8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd-kube-api-access-2fbt6\") pod \"community-operators-j2ksp\" (UID: \"8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd\") " pod="openshift-marketplace/community-operators-j2ksp" Mar 18 21:06:28 crc kubenswrapper[4950]: I0318 21:06:28.157858 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j2ksp" Mar 18 21:06:28 crc kubenswrapper[4950]: I0318 21:06:28.757678 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j2ksp"] Mar 18 21:06:28 crc kubenswrapper[4950]: W0318 21:06:28.770074 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8eb1f2f9_cdfb_4656_9dfa_1fe8803b6bfd.slice/crio-14a6eb61056e04f0481efcf4ab531da293036cec635cca9f0ab2c19cef5c9ee8 WatchSource:0}: Error finding container 14a6eb61056e04f0481efcf4ab531da293036cec635cca9f0ab2c19cef5c9ee8: Status 404 returned error can't find the container with id 14a6eb61056e04f0481efcf4ab531da293036cec635cca9f0ab2c19cef5c9ee8 Mar 18 21:06:29 crc kubenswrapper[4950]: I0318 21:06:29.584747 4950 generic.go:334] "Generic (PLEG): container finished" podID="8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd" containerID="71e59e7c266d718e89f0d9638e5208e31d2f6d158974a2dca32f075d245b6833" exitCode=0 Mar 18 21:06:29 crc kubenswrapper[4950]: I0318 21:06:29.585101 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j2ksp" event={"ID":"8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd","Type":"ContainerDied","Data":"71e59e7c266d718e89f0d9638e5208e31d2f6d158974a2dca32f075d245b6833"} Mar 18 21:06:29 crc kubenswrapper[4950]: I0318 21:06:29.585129 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j2ksp" event={"ID":"8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd","Type":"ContainerStarted","Data":"14a6eb61056e04f0481efcf4ab531da293036cec635cca9f0ab2c19cef5c9ee8"} Mar 18 21:06:30 crc kubenswrapper[4950]: I0318 21:06:30.596174 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j2ksp" event={"ID":"8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd","Type":"ContainerStarted","Data":"0c1ac377abdac70c139328f8dcd6fe4e4d30d1d27be4b4c75b2ffe4297bbfd51"} Mar 18 21:06:32 crc kubenswrapper[4950]: I0318 21:06:32.630614 4950 generic.go:334] "Generic (PLEG): container finished" podID="8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd" containerID="0c1ac377abdac70c139328f8dcd6fe4e4d30d1d27be4b4c75b2ffe4297bbfd51" exitCode=0 Mar 18 21:06:32 crc kubenswrapper[4950]: I0318 21:06:32.630678 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j2ksp" event={"ID":"8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd","Type":"ContainerDied","Data":"0c1ac377abdac70c139328f8dcd6fe4e4d30d1d27be4b4c75b2ffe4297bbfd51"} Mar 18 21:06:32 crc kubenswrapper[4950]: E0318 21:06:32.683299 4950 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8eb1f2f9_cdfb_4656_9dfa_1fe8803b6bfd.slice/crio-0c1ac377abdac70c139328f8dcd6fe4e4d30d1d27be4b4c75b2ffe4297bbfd51.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8eb1f2f9_cdfb_4656_9dfa_1fe8803b6bfd.slice/crio-conmon-0c1ac377abdac70c139328f8dcd6fe4e4d30d1d27be4b4c75b2ffe4297bbfd51.scope\": RecentStats: unable to find data in memory cache]" Mar 18 21:06:33 crc kubenswrapper[4950]: I0318 21:06:33.644703 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j2ksp" event={"ID":"8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd","Type":"ContainerStarted","Data":"70fae480116bd2aa04913823d4d2c29e553bf19c7296cca1b5c51698f6b351e1"} Mar 18 21:06:33 crc kubenswrapper[4950]: I0318 21:06:33.669617 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-j2ksp" podStartSLOduration=3.229452136 podStartE2EDuration="6.669599861s" podCreationTimestamp="2026-03-18 21:06:27 +0000 UTC" firstStartedPulling="2026-03-18 21:06:29.590755091 +0000 UTC m=+3602.831596959" lastFinishedPulling="2026-03-18 21:06:33.030902816 +0000 UTC m=+3606.271744684" observedRunningTime="2026-03-18 21:06:33.667918967 +0000 UTC m=+3606.908760855" watchObservedRunningTime="2026-03-18 21:06:33.669599861 +0000 UTC m=+3606.910441739" Mar 18 21:06:38 crc kubenswrapper[4950]: I0318 21:06:38.158708 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-j2ksp" Mar 18 21:06:38 crc kubenswrapper[4950]: I0318 21:06:38.159143 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-j2ksp" Mar 18 21:06:39 crc kubenswrapper[4950]: I0318 21:06:39.200425 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-j2ksp" podUID="8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd" containerName="registry-server" probeResult="failure" output=< Mar 18 21:06:39 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 21:06:39 crc kubenswrapper[4950]: > Mar 18 21:06:48 crc kubenswrapper[4950]: I0318 21:06:48.215249 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-j2ksp" Mar 18 21:06:48 crc kubenswrapper[4950]: I0318 21:06:48.280073 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-j2ksp" Mar 18 21:06:48 crc kubenswrapper[4950]: I0318 21:06:48.466678 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j2ksp"] Mar 18 21:06:49 crc kubenswrapper[4950]: I0318 21:06:49.795629 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-j2ksp" podUID="8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd" containerName="registry-server" containerID="cri-o://70fae480116bd2aa04913823d4d2c29e553bf19c7296cca1b5c51698f6b351e1" gracePeriod=2 Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.328504 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j2ksp" Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.377054 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fbt6\" (UniqueName: \"kubernetes.io/projected/8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd-kube-api-access-2fbt6\") pod \"8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd\" (UID: \"8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd\") " Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.377287 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd-catalog-content\") pod \"8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd\" (UID: \"8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd\") " Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.377354 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd-utilities\") pod \"8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd\" (UID: \"8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd\") " Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.378698 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd-utilities" (OuterVolumeSpecName: "utilities") pod "8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd" (UID: "8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.389645 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd-kube-api-access-2fbt6" (OuterVolumeSpecName: "kube-api-access-2fbt6") pod "8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd" (UID: "8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd"). InnerVolumeSpecName "kube-api-access-2fbt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.440987 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd" (UID: "8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.479984 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.480028 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fbt6\" (UniqueName: \"kubernetes.io/projected/8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd-kube-api-access-2fbt6\") on node \"crc\" DevicePath \"\"" Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.480045 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.807459 4950 generic.go:334] "Generic (PLEG): container finished" podID="8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd" containerID="70fae480116bd2aa04913823d4d2c29e553bf19c7296cca1b5c51698f6b351e1" exitCode=0 Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.807500 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j2ksp" event={"ID":"8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd","Type":"ContainerDied","Data":"70fae480116bd2aa04913823d4d2c29e553bf19c7296cca1b5c51698f6b351e1"} Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.807526 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j2ksp" event={"ID":"8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd","Type":"ContainerDied","Data":"14a6eb61056e04f0481efcf4ab531da293036cec635cca9f0ab2c19cef5c9ee8"} Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.807524 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j2ksp" Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.807544 4950 scope.go:117] "RemoveContainer" containerID="70fae480116bd2aa04913823d4d2c29e553bf19c7296cca1b5c51698f6b351e1" Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.884595 4950 scope.go:117] "RemoveContainer" containerID="0c1ac377abdac70c139328f8dcd6fe4e4d30d1d27be4b4c75b2ffe4297bbfd51" Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.896725 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j2ksp"] Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.907459 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-j2ksp"] Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.918472 4950 scope.go:117] "RemoveContainer" containerID="71e59e7c266d718e89f0d9638e5208e31d2f6d158974a2dca32f075d245b6833" Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.980049 4950 scope.go:117] "RemoveContainer" containerID="70fae480116bd2aa04913823d4d2c29e553bf19c7296cca1b5c51698f6b351e1" Mar 18 21:06:50 crc kubenswrapper[4950]: E0318 21:06:50.980747 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70fae480116bd2aa04913823d4d2c29e553bf19c7296cca1b5c51698f6b351e1\": container with ID starting with 70fae480116bd2aa04913823d4d2c29e553bf19c7296cca1b5c51698f6b351e1 not found: ID does not exist" containerID="70fae480116bd2aa04913823d4d2c29e553bf19c7296cca1b5c51698f6b351e1" Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.980784 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70fae480116bd2aa04913823d4d2c29e553bf19c7296cca1b5c51698f6b351e1"} err="failed to get container status \"70fae480116bd2aa04913823d4d2c29e553bf19c7296cca1b5c51698f6b351e1\": rpc error: code = NotFound desc = could not find container \"70fae480116bd2aa04913823d4d2c29e553bf19c7296cca1b5c51698f6b351e1\": container with ID starting with 70fae480116bd2aa04913823d4d2c29e553bf19c7296cca1b5c51698f6b351e1 not found: ID does not exist" Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.980815 4950 scope.go:117] "RemoveContainer" containerID="0c1ac377abdac70c139328f8dcd6fe4e4d30d1d27be4b4c75b2ffe4297bbfd51" Mar 18 21:06:50 crc kubenswrapper[4950]: E0318 21:06:50.991389 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c1ac377abdac70c139328f8dcd6fe4e4d30d1d27be4b4c75b2ffe4297bbfd51\": container with ID starting with 0c1ac377abdac70c139328f8dcd6fe4e4d30d1d27be4b4c75b2ffe4297bbfd51 not found: ID does not exist" containerID="0c1ac377abdac70c139328f8dcd6fe4e4d30d1d27be4b4c75b2ffe4297bbfd51" Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.991447 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c1ac377abdac70c139328f8dcd6fe4e4d30d1d27be4b4c75b2ffe4297bbfd51"} err="failed to get container status \"0c1ac377abdac70c139328f8dcd6fe4e4d30d1d27be4b4c75b2ffe4297bbfd51\": rpc error: code = NotFound desc = could not find container \"0c1ac377abdac70c139328f8dcd6fe4e4d30d1d27be4b4c75b2ffe4297bbfd51\": container with ID starting with 0c1ac377abdac70c139328f8dcd6fe4e4d30d1d27be4b4c75b2ffe4297bbfd51 not found: ID does not exist" Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.991472 4950 scope.go:117] "RemoveContainer" containerID="71e59e7c266d718e89f0d9638e5208e31d2f6d158974a2dca32f075d245b6833" Mar 18 21:06:50 crc kubenswrapper[4950]: E0318 21:06:50.994714 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71e59e7c266d718e89f0d9638e5208e31d2f6d158974a2dca32f075d245b6833\": container with ID starting with 71e59e7c266d718e89f0d9638e5208e31d2f6d158974a2dca32f075d245b6833 not found: ID does not exist" containerID="71e59e7c266d718e89f0d9638e5208e31d2f6d158974a2dca32f075d245b6833" Mar 18 21:06:50 crc kubenswrapper[4950]: I0318 21:06:50.994769 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71e59e7c266d718e89f0d9638e5208e31d2f6d158974a2dca32f075d245b6833"} err="failed to get container status \"71e59e7c266d718e89f0d9638e5208e31d2f6d158974a2dca32f075d245b6833\": rpc error: code = NotFound desc = could not find container \"71e59e7c266d718e89f0d9638e5208e31d2f6d158974a2dca32f075d245b6833\": container with ID starting with 71e59e7c266d718e89f0d9638e5208e31d2f6d158974a2dca32f075d245b6833 not found: ID does not exist" Mar 18 21:06:51 crc kubenswrapper[4950]: I0318 21:06:51.502297 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd" path="/var/lib/kubelet/pods/8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd/volumes" Mar 18 21:06:59 crc kubenswrapper[4950]: I0318 21:06:59.907001 4950 scope.go:117] "RemoveContainer" containerID="b18dc7ae008b2e958bcd61cbb5faf4c4cf93ce26ea3399ab2cf053160cf355c1" Mar 18 21:07:33 crc kubenswrapper[4950]: I0318 21:07:33.836219 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 21:07:33 crc kubenswrapper[4950]: I0318 21:07:33.836998 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 21:08:00 crc kubenswrapper[4950]: I0318 21:08:00.162928 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564468-2cvt2"] Mar 18 21:08:00 crc kubenswrapper[4950]: E0318 21:08:00.163991 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd" containerName="registry-server" Mar 18 21:08:00 crc kubenswrapper[4950]: I0318 21:08:00.164004 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd" containerName="registry-server" Mar 18 21:08:00 crc kubenswrapper[4950]: E0318 21:08:00.164013 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd" containerName="extract-utilities" Mar 18 21:08:00 crc kubenswrapper[4950]: I0318 21:08:00.164021 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd" containerName="extract-utilities" Mar 18 21:08:00 crc kubenswrapper[4950]: E0318 21:08:00.164037 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd" containerName="extract-content" Mar 18 21:08:00 crc kubenswrapper[4950]: I0318 21:08:00.164044 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd" containerName="extract-content" Mar 18 21:08:00 crc kubenswrapper[4950]: I0318 21:08:00.164232 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="8eb1f2f9-cdfb-4656-9dfa-1fe8803b6bfd" containerName="registry-server" Mar 18 21:08:00 crc kubenswrapper[4950]: I0318 21:08:00.164934 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564468-2cvt2" Mar 18 21:08:00 crc kubenswrapper[4950]: I0318 21:08:00.167722 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 21:08:00 crc kubenswrapper[4950]: I0318 21:08:00.173486 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 21:08:00 crc kubenswrapper[4950]: I0318 21:08:00.178556 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564468-2cvt2"] Mar 18 21:08:00 crc kubenswrapper[4950]: I0318 21:08:00.186505 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 21:08:00 crc kubenswrapper[4950]: I0318 21:08:00.260702 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvpcf\" (UniqueName: \"kubernetes.io/projected/c1deaedc-119e-4e3b-ad75-86eda667d445-kube-api-access-rvpcf\") pod \"auto-csr-approver-29564468-2cvt2\" (UID: \"c1deaedc-119e-4e3b-ad75-86eda667d445\") " pod="openshift-infra/auto-csr-approver-29564468-2cvt2" Mar 18 21:08:00 crc kubenswrapper[4950]: I0318 21:08:00.362271 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvpcf\" (UniqueName: \"kubernetes.io/projected/c1deaedc-119e-4e3b-ad75-86eda667d445-kube-api-access-rvpcf\") pod \"auto-csr-approver-29564468-2cvt2\" (UID: \"c1deaedc-119e-4e3b-ad75-86eda667d445\") " pod="openshift-infra/auto-csr-approver-29564468-2cvt2" Mar 18 21:08:00 crc kubenswrapper[4950]: I0318 21:08:00.388215 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvpcf\" (UniqueName: \"kubernetes.io/projected/c1deaedc-119e-4e3b-ad75-86eda667d445-kube-api-access-rvpcf\") pod \"auto-csr-approver-29564468-2cvt2\" (UID: \"c1deaedc-119e-4e3b-ad75-86eda667d445\") " pod="openshift-infra/auto-csr-approver-29564468-2cvt2" Mar 18 21:08:00 crc kubenswrapper[4950]: I0318 21:08:00.485187 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564468-2cvt2" Mar 18 21:08:00 crc kubenswrapper[4950]: I0318 21:08:00.978592 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564468-2cvt2"] Mar 18 21:08:01 crc kubenswrapper[4950]: I0318 21:08:01.462835 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564468-2cvt2" event={"ID":"c1deaedc-119e-4e3b-ad75-86eda667d445","Type":"ContainerStarted","Data":"80cb19836f07475e249e99036e1b5b54035c8eb34ec78844328bc7092fa441cd"} Mar 18 21:08:02 crc kubenswrapper[4950]: I0318 21:08:02.497403 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564468-2cvt2" event={"ID":"c1deaedc-119e-4e3b-ad75-86eda667d445","Type":"ContainerStarted","Data":"e8fd6fd512f2ef0e49276b7bdf8a42dfbb9e0e1236732ace73cedd484d6bfa3e"} Mar 18 21:08:02 crc kubenswrapper[4950]: I0318 21:08:02.519357 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29564468-2cvt2" podStartSLOduration=1.5799061559999998 podStartE2EDuration="2.519336589s" podCreationTimestamp="2026-03-18 21:08:00 +0000 UTC" firstStartedPulling="2026-03-18 21:08:01.016905996 +0000 UTC m=+3694.257747864" lastFinishedPulling="2026-03-18 21:08:01.956336429 +0000 UTC m=+3695.197178297" observedRunningTime="2026-03-18 21:08:02.514845231 +0000 UTC m=+3695.755687099" watchObservedRunningTime="2026-03-18 21:08:02.519336589 +0000 UTC m=+3695.760178457" Mar 18 21:08:03 crc kubenswrapper[4950]: I0318 21:08:03.507760 4950 generic.go:334] "Generic (PLEG): container finished" podID="c1deaedc-119e-4e3b-ad75-86eda667d445" containerID="e8fd6fd512f2ef0e49276b7bdf8a42dfbb9e0e1236732ace73cedd484d6bfa3e" exitCode=0 Mar 18 21:08:03 crc kubenswrapper[4950]: I0318 21:08:03.508270 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564468-2cvt2" event={"ID":"c1deaedc-119e-4e3b-ad75-86eda667d445","Type":"ContainerDied","Data":"e8fd6fd512f2ef0e49276b7bdf8a42dfbb9e0e1236732ace73cedd484d6bfa3e"} Mar 18 21:08:03 crc kubenswrapper[4950]: I0318 21:08:03.836127 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 21:08:03 crc kubenswrapper[4950]: I0318 21:08:03.836191 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 21:08:04 crc kubenswrapper[4950]: I0318 21:08:04.994737 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564468-2cvt2" Mar 18 21:08:05 crc kubenswrapper[4950]: I0318 21:08:05.052378 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvpcf\" (UniqueName: \"kubernetes.io/projected/c1deaedc-119e-4e3b-ad75-86eda667d445-kube-api-access-rvpcf\") pod \"c1deaedc-119e-4e3b-ad75-86eda667d445\" (UID: \"c1deaedc-119e-4e3b-ad75-86eda667d445\") " Mar 18 21:08:05 crc kubenswrapper[4950]: I0318 21:08:05.058817 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1deaedc-119e-4e3b-ad75-86eda667d445-kube-api-access-rvpcf" (OuterVolumeSpecName: "kube-api-access-rvpcf") pod "c1deaedc-119e-4e3b-ad75-86eda667d445" (UID: "c1deaedc-119e-4e3b-ad75-86eda667d445"). InnerVolumeSpecName "kube-api-access-rvpcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:08:05 crc kubenswrapper[4950]: I0318 21:08:05.155580 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvpcf\" (UniqueName: \"kubernetes.io/projected/c1deaedc-119e-4e3b-ad75-86eda667d445-kube-api-access-rvpcf\") on node \"crc\" DevicePath \"\"" Mar 18 21:08:05 crc kubenswrapper[4950]: I0318 21:08:05.531358 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564468-2cvt2" event={"ID":"c1deaedc-119e-4e3b-ad75-86eda667d445","Type":"ContainerDied","Data":"80cb19836f07475e249e99036e1b5b54035c8eb34ec78844328bc7092fa441cd"} Mar 18 21:08:05 crc kubenswrapper[4950]: I0318 21:08:05.531748 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80cb19836f07475e249e99036e1b5b54035c8eb34ec78844328bc7092fa441cd" Mar 18 21:08:05 crc kubenswrapper[4950]: I0318 21:08:05.531508 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564468-2cvt2" Mar 18 21:08:05 crc kubenswrapper[4950]: I0318 21:08:05.604598 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564462-ftngf"] Mar 18 21:08:05 crc kubenswrapper[4950]: I0318 21:08:05.614195 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564462-ftngf"] Mar 18 21:08:07 crc kubenswrapper[4950]: I0318 21:08:07.505137 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74848e1c-681b-4e2f-8d52-141547d440eb" path="/var/lib/kubelet/pods/74848e1c-681b-4e2f-8d52-141547d440eb/volumes" Mar 18 21:08:33 crc kubenswrapper[4950]: I0318 21:08:33.836271 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 21:08:33 crc kubenswrapper[4950]: I0318 21:08:33.836976 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 21:08:33 crc kubenswrapper[4950]: I0318 21:08:33.837039 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 21:08:33 crc kubenswrapper[4950]: I0318 21:08:33.838215 4950 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"16b4cab9ed932e2b4b0bb08ce747a1dbd93ab41574373696a03e18ef0ef41c79"} pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 21:08:33 crc kubenswrapper[4950]: I0318 21:08:33.838380 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" containerID="cri-o://16b4cab9ed932e2b4b0bb08ce747a1dbd93ab41574373696a03e18ef0ef41c79" gracePeriod=600 Mar 18 21:08:34 crc kubenswrapper[4950]: I0318 21:08:34.806888 4950 generic.go:334] "Generic (PLEG): container finished" podID="4048b439-3266-46e7-9de0-22377efacc46" containerID="16b4cab9ed932e2b4b0bb08ce747a1dbd93ab41574373696a03e18ef0ef41c79" exitCode=0 Mar 18 21:08:34 crc kubenswrapper[4950]: I0318 21:08:34.806967 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerDied","Data":"16b4cab9ed932e2b4b0bb08ce747a1dbd93ab41574373696a03e18ef0ef41c79"} Mar 18 21:08:34 crc kubenswrapper[4950]: I0318 21:08:34.807824 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerStarted","Data":"8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9"} Mar 18 21:08:34 crc kubenswrapper[4950]: I0318 21:08:34.807845 4950 scope.go:117] "RemoveContainer" containerID="afd59ac4e153ee7e0c52b87cee7d28b979840b2470ad0e967f711113d0e8e0c1" Mar 18 21:09:00 crc kubenswrapper[4950]: I0318 21:09:00.037677 4950 scope.go:117] "RemoveContainer" containerID="23a80d84c715f39a3f35546576b36927d8c0ca75ec87201176e81085fc89804a" Mar 18 21:09:36 crc kubenswrapper[4950]: I0318 21:09:36.049529 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-4qv4l"] Mar 18 21:09:36 crc kubenswrapper[4950]: I0318 21:09:36.056838 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-4qv4l"] Mar 18 21:09:37 crc kubenswrapper[4950]: I0318 21:09:37.496234 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e05258ae-28d6-4c4c-8bc0-c20e6f112198" path="/var/lib/kubelet/pods/e05258ae-28d6-4c4c-8bc0-c20e6f112198/volumes" Mar 18 21:09:38 crc kubenswrapper[4950]: I0318 21:09:38.025806 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-ca5b-account-create-update-8l5r2"] Mar 18 21:09:38 crc kubenswrapper[4950]: I0318 21:09:38.034893 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-ca5b-account-create-update-8l5r2"] Mar 18 21:09:39 crc kubenswrapper[4950]: I0318 21:09:39.493028 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bf02c5a-528e-43e3-8f9d-faa5619091e8" path="/var/lib/kubelet/pods/4bf02c5a-528e-43e3-8f9d-faa5619091e8/volumes" Mar 18 21:09:51 crc kubenswrapper[4950]: I0318 21:09:51.964293 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-42xf9"] Mar 18 21:09:51 crc kubenswrapper[4950]: E0318 21:09:51.965283 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1deaedc-119e-4e3b-ad75-86eda667d445" containerName="oc" Mar 18 21:09:51 crc kubenswrapper[4950]: I0318 21:09:51.965299 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1deaedc-119e-4e3b-ad75-86eda667d445" containerName="oc" Mar 18 21:09:51 crc kubenswrapper[4950]: I0318 21:09:51.965592 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1deaedc-119e-4e3b-ad75-86eda667d445" containerName="oc" Mar 18 21:09:51 crc kubenswrapper[4950]: I0318 21:09:51.967260 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-42xf9" Mar 18 21:09:51 crc kubenswrapper[4950]: I0318 21:09:51.980668 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-42xf9"] Mar 18 21:09:52 crc kubenswrapper[4950]: I0318 21:09:52.086368 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl6mj\" (UniqueName: \"kubernetes.io/projected/d72524a5-b7b7-416f-b9ba-9527f98e787c-kube-api-access-jl6mj\") pod \"redhat-operators-42xf9\" (UID: \"d72524a5-b7b7-416f-b9ba-9527f98e787c\") " pod="openshift-marketplace/redhat-operators-42xf9" Mar 18 21:09:52 crc kubenswrapper[4950]: I0318 21:09:52.087065 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d72524a5-b7b7-416f-b9ba-9527f98e787c-catalog-content\") pod \"redhat-operators-42xf9\" (UID: \"d72524a5-b7b7-416f-b9ba-9527f98e787c\") " pod="openshift-marketplace/redhat-operators-42xf9" Mar 18 21:09:52 crc kubenswrapper[4950]: I0318 21:09:52.087191 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d72524a5-b7b7-416f-b9ba-9527f98e787c-utilities\") pod \"redhat-operators-42xf9\" (UID: \"d72524a5-b7b7-416f-b9ba-9527f98e787c\") " pod="openshift-marketplace/redhat-operators-42xf9" Mar 18 21:09:52 crc kubenswrapper[4950]: I0318 21:09:52.189108 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl6mj\" (UniqueName: \"kubernetes.io/projected/d72524a5-b7b7-416f-b9ba-9527f98e787c-kube-api-access-jl6mj\") pod \"redhat-operators-42xf9\" (UID: \"d72524a5-b7b7-416f-b9ba-9527f98e787c\") " pod="openshift-marketplace/redhat-operators-42xf9" Mar 18 21:09:52 crc kubenswrapper[4950]: I0318 21:09:52.189173 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d72524a5-b7b7-416f-b9ba-9527f98e787c-catalog-content\") pod \"redhat-operators-42xf9\" (UID: \"d72524a5-b7b7-416f-b9ba-9527f98e787c\") " pod="openshift-marketplace/redhat-operators-42xf9" Mar 18 21:09:52 crc kubenswrapper[4950]: I0318 21:09:52.189244 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d72524a5-b7b7-416f-b9ba-9527f98e787c-utilities\") pod \"redhat-operators-42xf9\" (UID: \"d72524a5-b7b7-416f-b9ba-9527f98e787c\") " pod="openshift-marketplace/redhat-operators-42xf9" Mar 18 21:09:52 crc kubenswrapper[4950]: I0318 21:09:52.189928 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d72524a5-b7b7-416f-b9ba-9527f98e787c-catalog-content\") pod \"redhat-operators-42xf9\" (UID: \"d72524a5-b7b7-416f-b9ba-9527f98e787c\") " pod="openshift-marketplace/redhat-operators-42xf9" Mar 18 21:09:52 crc kubenswrapper[4950]: I0318 21:09:52.189930 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d72524a5-b7b7-416f-b9ba-9527f98e787c-utilities\") pod \"redhat-operators-42xf9\" (UID: \"d72524a5-b7b7-416f-b9ba-9527f98e787c\") " pod="openshift-marketplace/redhat-operators-42xf9" Mar 18 21:09:52 crc kubenswrapper[4950]: I0318 21:09:52.225243 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl6mj\" (UniqueName: \"kubernetes.io/projected/d72524a5-b7b7-416f-b9ba-9527f98e787c-kube-api-access-jl6mj\") pod \"redhat-operators-42xf9\" (UID: \"d72524a5-b7b7-416f-b9ba-9527f98e787c\") " pod="openshift-marketplace/redhat-operators-42xf9" Mar 18 21:09:52 crc kubenswrapper[4950]: I0318 21:09:52.290848 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-42xf9" Mar 18 21:09:52 crc kubenswrapper[4950]: I0318 21:09:52.835480 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-42xf9"] Mar 18 21:09:53 crc kubenswrapper[4950]: I0318 21:09:53.582057 4950 generic.go:334] "Generic (PLEG): container finished" podID="d72524a5-b7b7-416f-b9ba-9527f98e787c" containerID="2a551b8d2375678022d16016f9f9f9556757dad44962c720880ede223c49e83c" exitCode=0 Mar 18 21:09:53 crc kubenswrapper[4950]: I0318 21:09:53.582258 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42xf9" event={"ID":"d72524a5-b7b7-416f-b9ba-9527f98e787c","Type":"ContainerDied","Data":"2a551b8d2375678022d16016f9f9f9556757dad44962c720880ede223c49e83c"} Mar 18 21:09:53 crc kubenswrapper[4950]: I0318 21:09:53.582359 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42xf9" event={"ID":"d72524a5-b7b7-416f-b9ba-9527f98e787c","Type":"ContainerStarted","Data":"f5ddee5d66abd661d82b8c83532ef88fc804ac00bc1b820a18d4ae47377e796a"} Mar 18 21:09:55 crc kubenswrapper[4950]: I0318 21:09:55.583032 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9x7pn"] Mar 18 21:09:55 crc kubenswrapper[4950]: I0318 21:09:55.585516 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9x7pn" Mar 18 21:09:55 crc kubenswrapper[4950]: I0318 21:09:55.615202 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9x7pn"] Mar 18 21:09:55 crc kubenswrapper[4950]: I0318 21:09:55.764211 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkhsr\" (UniqueName: \"kubernetes.io/projected/73e71e82-d4df-4df9-9c3b-35fc9f990a67-kube-api-access-kkhsr\") pod \"certified-operators-9x7pn\" (UID: \"73e71e82-d4df-4df9-9c3b-35fc9f990a67\") " pod="openshift-marketplace/certified-operators-9x7pn" Mar 18 21:09:55 crc kubenswrapper[4950]: I0318 21:09:55.764323 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73e71e82-d4df-4df9-9c3b-35fc9f990a67-catalog-content\") pod \"certified-operators-9x7pn\" (UID: \"73e71e82-d4df-4df9-9c3b-35fc9f990a67\") " pod="openshift-marketplace/certified-operators-9x7pn" Mar 18 21:09:55 crc kubenswrapper[4950]: I0318 21:09:55.764349 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73e71e82-d4df-4df9-9c3b-35fc9f990a67-utilities\") pod \"certified-operators-9x7pn\" (UID: \"73e71e82-d4df-4df9-9c3b-35fc9f990a67\") " pod="openshift-marketplace/certified-operators-9x7pn" Mar 18 21:09:55 crc kubenswrapper[4950]: I0318 21:09:55.865607 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkhsr\" (UniqueName: \"kubernetes.io/projected/73e71e82-d4df-4df9-9c3b-35fc9f990a67-kube-api-access-kkhsr\") pod \"certified-operators-9x7pn\" (UID: \"73e71e82-d4df-4df9-9c3b-35fc9f990a67\") " pod="openshift-marketplace/certified-operators-9x7pn" Mar 18 21:09:55 crc kubenswrapper[4950]: I0318 21:09:55.865727 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73e71e82-d4df-4df9-9c3b-35fc9f990a67-catalog-content\") pod \"certified-operators-9x7pn\" (UID: \"73e71e82-d4df-4df9-9c3b-35fc9f990a67\") " pod="openshift-marketplace/certified-operators-9x7pn" Mar 18 21:09:55 crc kubenswrapper[4950]: I0318 21:09:55.865761 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73e71e82-d4df-4df9-9c3b-35fc9f990a67-utilities\") pod \"certified-operators-9x7pn\" (UID: \"73e71e82-d4df-4df9-9c3b-35fc9f990a67\") " pod="openshift-marketplace/certified-operators-9x7pn" Mar 18 21:09:55 crc kubenswrapper[4950]: I0318 21:09:55.866336 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73e71e82-d4df-4df9-9c3b-35fc9f990a67-utilities\") pod \"certified-operators-9x7pn\" (UID: \"73e71e82-d4df-4df9-9c3b-35fc9f990a67\") " pod="openshift-marketplace/certified-operators-9x7pn" Mar 18 21:09:55 crc kubenswrapper[4950]: I0318 21:09:55.866453 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73e71e82-d4df-4df9-9c3b-35fc9f990a67-catalog-content\") pod \"certified-operators-9x7pn\" (UID: \"73e71e82-d4df-4df9-9c3b-35fc9f990a67\") " pod="openshift-marketplace/certified-operators-9x7pn" Mar 18 21:09:55 crc kubenswrapper[4950]: I0318 21:09:55.885383 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkhsr\" (UniqueName: \"kubernetes.io/projected/73e71e82-d4df-4df9-9c3b-35fc9f990a67-kube-api-access-kkhsr\") pod \"certified-operators-9x7pn\" (UID: \"73e71e82-d4df-4df9-9c3b-35fc9f990a67\") " pod="openshift-marketplace/certified-operators-9x7pn" Mar 18 21:09:55 crc kubenswrapper[4950]: I0318 21:09:55.915621 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9x7pn" Mar 18 21:09:56 crc kubenswrapper[4950]: I0318 21:09:56.680346 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9x7pn"] Mar 18 21:09:56 crc kubenswrapper[4950]: W0318 21:09:56.695673 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73e71e82_d4df_4df9_9c3b_35fc9f990a67.slice/crio-d7e91317d8f63c6eee32619a46a766878956dc6696210055bda7f4e9ea2b252b WatchSource:0}: Error finding container d7e91317d8f63c6eee32619a46a766878956dc6696210055bda7f4e9ea2b252b: Status 404 returned error can't find the container with id d7e91317d8f63c6eee32619a46a766878956dc6696210055bda7f4e9ea2b252b Mar 18 21:09:57 crc kubenswrapper[4950]: I0318 21:09:57.638765 4950 generic.go:334] "Generic (PLEG): container finished" podID="73e71e82-d4df-4df9-9c3b-35fc9f990a67" containerID="63dc5fa85995b3a4729f98640763abca5646338ed067cbb6c9ced6c29ca204f1" exitCode=0 Mar 18 21:09:57 crc kubenswrapper[4950]: I0318 21:09:57.638945 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9x7pn" event={"ID":"73e71e82-d4df-4df9-9c3b-35fc9f990a67","Type":"ContainerDied","Data":"63dc5fa85995b3a4729f98640763abca5646338ed067cbb6c9ced6c29ca204f1"} Mar 18 21:09:57 crc kubenswrapper[4950]: I0318 21:09:57.639072 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9x7pn" event={"ID":"73e71e82-d4df-4df9-9c3b-35fc9f990a67","Type":"ContainerStarted","Data":"d7e91317d8f63c6eee32619a46a766878956dc6696210055bda7f4e9ea2b252b"} Mar 18 21:09:58 crc kubenswrapper[4950]: I0318 21:09:58.652355 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9x7pn" event={"ID":"73e71e82-d4df-4df9-9c3b-35fc9f990a67","Type":"ContainerStarted","Data":"e62985af67b08d18284cda3e149c4040fcaeeb367cea820cacde4889dff86c02"} Mar 18 21:10:00 crc kubenswrapper[4950]: I0318 21:10:00.145440 4950 scope.go:117] "RemoveContainer" containerID="4f8a0098d6fda368230ff5bb863ce9e72a1a14336e6d63b90fa918bb309607fa" Mar 18 21:10:00 crc kubenswrapper[4950]: I0318 21:10:00.191563 4950 scope.go:117] "RemoveContainer" containerID="93b98e9080b2a203c6474765a7e3afd36f0f3a668d12adcbcd4e48111df5e0b4" Mar 18 21:10:00 crc kubenswrapper[4950]: I0318 21:10:00.206526 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564470-mhw8z"] Mar 18 21:10:00 crc kubenswrapper[4950]: I0318 21:10:00.208086 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564470-mhw8z" Mar 18 21:10:00 crc kubenswrapper[4950]: I0318 21:10:00.213971 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 21:10:00 crc kubenswrapper[4950]: I0318 21:10:00.214196 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 21:10:00 crc kubenswrapper[4950]: I0318 21:10:00.214396 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 21:10:00 crc kubenswrapper[4950]: I0318 21:10:00.234934 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564470-mhw8z"] Mar 18 21:10:00 crc kubenswrapper[4950]: I0318 21:10:00.368880 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8f54\" (UniqueName: \"kubernetes.io/projected/7142cc1f-1577-45b7-bc79-a4dd4416455e-kube-api-access-j8f54\") pod \"auto-csr-approver-29564470-mhw8z\" (UID: \"7142cc1f-1577-45b7-bc79-a4dd4416455e\") " pod="openshift-infra/auto-csr-approver-29564470-mhw8z" Mar 18 21:10:00 crc kubenswrapper[4950]: I0318 21:10:00.471254 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8f54\" (UniqueName: \"kubernetes.io/projected/7142cc1f-1577-45b7-bc79-a4dd4416455e-kube-api-access-j8f54\") pod \"auto-csr-approver-29564470-mhw8z\" (UID: \"7142cc1f-1577-45b7-bc79-a4dd4416455e\") " pod="openshift-infra/auto-csr-approver-29564470-mhw8z" Mar 18 21:10:00 crc kubenswrapper[4950]: I0318 21:10:00.499430 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8f54\" (UniqueName: \"kubernetes.io/projected/7142cc1f-1577-45b7-bc79-a4dd4416455e-kube-api-access-j8f54\") pod \"auto-csr-approver-29564470-mhw8z\" (UID: \"7142cc1f-1577-45b7-bc79-a4dd4416455e\") " pod="openshift-infra/auto-csr-approver-29564470-mhw8z" Mar 18 21:10:00 crc kubenswrapper[4950]: I0318 21:10:00.549445 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564470-mhw8z" Mar 18 21:10:00 crc kubenswrapper[4950]: I0318 21:10:00.675369 4950 generic.go:334] "Generic (PLEG): container finished" podID="73e71e82-d4df-4df9-9c3b-35fc9f990a67" containerID="e62985af67b08d18284cda3e149c4040fcaeeb367cea820cacde4889dff86c02" exitCode=0 Mar 18 21:10:00 crc kubenswrapper[4950]: I0318 21:10:00.675431 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9x7pn" event={"ID":"73e71e82-d4df-4df9-9c3b-35fc9f990a67","Type":"ContainerDied","Data":"e62985af67b08d18284cda3e149c4040fcaeeb367cea820cacde4889dff86c02"} Mar 18 21:10:01 crc kubenswrapper[4950]: I0318 21:10:01.252882 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564470-mhw8z"] Mar 18 21:10:01 crc kubenswrapper[4950]: I0318 21:10:01.683865 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564470-mhw8z" event={"ID":"7142cc1f-1577-45b7-bc79-a4dd4416455e","Type":"ContainerStarted","Data":"871baa173beabe1ad47172092285b00201d3024f42062005c9a8e6b9c743e347"} Mar 18 21:10:01 crc kubenswrapper[4950]: I0318 21:10:01.686620 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9x7pn" event={"ID":"73e71e82-d4df-4df9-9c3b-35fc9f990a67","Type":"ContainerStarted","Data":"f962de81a8bdda735e833a62d37ee650546560c30cfa8115f41bab1aef2fb361"} Mar 18 21:10:03 crc kubenswrapper[4950]: I0318 21:10:03.702938 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564470-mhw8z" event={"ID":"7142cc1f-1577-45b7-bc79-a4dd4416455e","Type":"ContainerStarted","Data":"0ca7d6412521119010b6f028cb103fb9b20d1e5515c87aa7a804389c40feee74"} Mar 18 21:10:03 crc kubenswrapper[4950]: I0318 21:10:03.721596 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9x7pn" podStartSLOduration=5.265783687 podStartE2EDuration="8.721578773s" podCreationTimestamp="2026-03-18 21:09:55 +0000 UTC" firstStartedPulling="2026-03-18 21:09:57.641928037 +0000 UTC m=+3810.882769905" lastFinishedPulling="2026-03-18 21:10:01.097723123 +0000 UTC m=+3814.338564991" observedRunningTime="2026-03-18 21:10:01.708831974 +0000 UTC m=+3814.949673862" watchObservedRunningTime="2026-03-18 21:10:03.721578773 +0000 UTC m=+3816.962420641" Mar 18 21:10:03 crc kubenswrapper[4950]: I0318 21:10:03.726504 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29564470-mhw8z" podStartSLOduration=2.031702972 podStartE2EDuration="3.726488882s" podCreationTimestamp="2026-03-18 21:10:00 +0000 UTC" firstStartedPulling="2026-03-18 21:10:01.265940335 +0000 UTC m=+3814.506782203" lastFinishedPulling="2026-03-18 21:10:02.960726245 +0000 UTC m=+3816.201568113" observedRunningTime="2026-03-18 21:10:03.720241258 +0000 UTC m=+3816.961083126" watchObservedRunningTime="2026-03-18 21:10:03.726488882 +0000 UTC m=+3816.967330750" Mar 18 21:10:04 crc kubenswrapper[4950]: I0318 21:10:04.714778 4950 generic.go:334] "Generic (PLEG): container finished" podID="7142cc1f-1577-45b7-bc79-a4dd4416455e" containerID="0ca7d6412521119010b6f028cb103fb9b20d1e5515c87aa7a804389c40feee74" exitCode=0 Mar 18 21:10:04 crc kubenswrapper[4950]: I0318 21:10:04.714865 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564470-mhw8z" event={"ID":"7142cc1f-1577-45b7-bc79-a4dd4416455e","Type":"ContainerDied","Data":"0ca7d6412521119010b6f028cb103fb9b20d1e5515c87aa7a804389c40feee74"} Mar 18 21:10:05 crc kubenswrapper[4950]: I0318 21:10:05.963907 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9x7pn" Mar 18 21:10:05 crc kubenswrapper[4950]: I0318 21:10:05.965578 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9x7pn" Mar 18 21:10:07 crc kubenswrapper[4950]: I0318 21:10:07.014814 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-9x7pn" podUID="73e71e82-d4df-4df9-9c3b-35fc9f990a67" containerName="registry-server" probeResult="failure" output=< Mar 18 21:10:07 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 21:10:07 crc kubenswrapper[4950]: > Mar 18 21:10:10 crc kubenswrapper[4950]: I0318 21:10:10.212939 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564470-mhw8z" Mar 18 21:10:10 crc kubenswrapper[4950]: I0318 21:10:10.259963 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8f54\" (UniqueName: \"kubernetes.io/projected/7142cc1f-1577-45b7-bc79-a4dd4416455e-kube-api-access-j8f54\") pod \"7142cc1f-1577-45b7-bc79-a4dd4416455e\" (UID: \"7142cc1f-1577-45b7-bc79-a4dd4416455e\") " Mar 18 21:10:10 crc kubenswrapper[4950]: I0318 21:10:10.285394 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7142cc1f-1577-45b7-bc79-a4dd4416455e-kube-api-access-j8f54" (OuterVolumeSpecName: "kube-api-access-j8f54") pod "7142cc1f-1577-45b7-bc79-a4dd4416455e" (UID: "7142cc1f-1577-45b7-bc79-a4dd4416455e"). InnerVolumeSpecName "kube-api-access-j8f54". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:10:10 crc kubenswrapper[4950]: I0318 21:10:10.362817 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8f54\" (UniqueName: \"kubernetes.io/projected/7142cc1f-1577-45b7-bc79-a4dd4416455e-kube-api-access-j8f54\") on node \"crc\" DevicePath \"\"" Mar 18 21:10:10 crc kubenswrapper[4950]: I0318 21:10:10.770778 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564470-mhw8z" Mar 18 21:10:10 crc kubenswrapper[4950]: I0318 21:10:10.771590 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564470-mhw8z" event={"ID":"7142cc1f-1577-45b7-bc79-a4dd4416455e","Type":"ContainerDied","Data":"871baa173beabe1ad47172092285b00201d3024f42062005c9a8e6b9c743e347"} Mar 18 21:10:10 crc kubenswrapper[4950]: I0318 21:10:10.771624 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="871baa173beabe1ad47172092285b00201d3024f42062005c9a8e6b9c743e347" Mar 18 21:10:10 crc kubenswrapper[4950]: I0318 21:10:10.775280 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42xf9" event={"ID":"d72524a5-b7b7-416f-b9ba-9527f98e787c","Type":"ContainerStarted","Data":"0a099375426e51507e45e724a186f9b3038c69f2a0341b9f40daf5647cc12acd"} Mar 18 21:10:11 crc kubenswrapper[4950]: I0318 21:10:11.377238 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564464-qtkkw"] Mar 18 21:10:11 crc kubenswrapper[4950]: I0318 21:10:11.394568 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564464-qtkkw"] Mar 18 21:10:11 crc kubenswrapper[4950]: I0318 21:10:11.492538 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5de9d39c-cf8e-401d-84e4-95d1bbee331f" path="/var/lib/kubelet/pods/5de9d39c-cf8e-401d-84e4-95d1bbee331f/volumes" Mar 18 21:10:13 crc kubenswrapper[4950]: I0318 21:10:13.810856 4950 generic.go:334] "Generic (PLEG): container finished" podID="d72524a5-b7b7-416f-b9ba-9527f98e787c" containerID="0a099375426e51507e45e724a186f9b3038c69f2a0341b9f40daf5647cc12acd" exitCode=0 Mar 18 21:10:13 crc kubenswrapper[4950]: I0318 21:10:13.811277 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42xf9" event={"ID":"d72524a5-b7b7-416f-b9ba-9527f98e787c","Type":"ContainerDied","Data":"0a099375426e51507e45e724a186f9b3038c69f2a0341b9f40daf5647cc12acd"} Mar 18 21:10:14 crc kubenswrapper[4950]: I0318 21:10:14.821999 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42xf9" event={"ID":"d72524a5-b7b7-416f-b9ba-9527f98e787c","Type":"ContainerStarted","Data":"24ff426d6a997da2523833cdf4dde9ba60e0e23ed1ec30b08f07d7a575f35f5f"} Mar 18 21:10:14 crc kubenswrapper[4950]: I0318 21:10:14.842949 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-42xf9" podStartSLOduration=3.201742962 podStartE2EDuration="23.842930168s" podCreationTimestamp="2026-03-18 21:09:51 +0000 UTC" firstStartedPulling="2026-03-18 21:09:53.584191742 +0000 UTC m=+3806.825033610" lastFinishedPulling="2026-03-18 21:10:14.225378948 +0000 UTC m=+3827.466220816" observedRunningTime="2026-03-18 21:10:14.842232049 +0000 UTC m=+3828.083073927" watchObservedRunningTime="2026-03-18 21:10:14.842930168 +0000 UTC m=+3828.083772036" Mar 18 21:10:16 crc kubenswrapper[4950]: I0318 21:10:16.045299 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-5szr9"] Mar 18 21:10:16 crc kubenswrapper[4950]: I0318 21:10:16.053804 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-5szr9"] Mar 18 21:10:16 crc kubenswrapper[4950]: I0318 21:10:16.961777 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-9x7pn" podUID="73e71e82-d4df-4df9-9c3b-35fc9f990a67" containerName="registry-server" probeResult="failure" output=< Mar 18 21:10:16 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 21:10:16 crc kubenswrapper[4950]: > Mar 18 21:10:17 crc kubenswrapper[4950]: I0318 21:10:17.497531 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3a7d0be-1cb1-47dc-845b-25c91adb00dd" path="/var/lib/kubelet/pods/a3a7d0be-1cb1-47dc-845b-25c91adb00dd/volumes" Mar 18 21:10:22 crc kubenswrapper[4950]: I0318 21:10:22.292165 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-42xf9" Mar 18 21:10:22 crc kubenswrapper[4950]: I0318 21:10:22.293282 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-42xf9" Mar 18 21:10:23 crc kubenswrapper[4950]: I0318 21:10:23.367224 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-42xf9" podUID="d72524a5-b7b7-416f-b9ba-9527f98e787c" containerName="registry-server" probeResult="failure" output=< Mar 18 21:10:23 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 21:10:23 crc kubenswrapper[4950]: > Mar 18 21:10:25 crc kubenswrapper[4950]: I0318 21:10:25.972067 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9x7pn" Mar 18 21:10:26 crc kubenswrapper[4950]: I0318 21:10:26.026098 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9x7pn" Mar 18 21:10:27 crc kubenswrapper[4950]: I0318 21:10:27.568007 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9x7pn"] Mar 18 21:10:27 crc kubenswrapper[4950]: I0318 21:10:27.954227 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9x7pn" podUID="73e71e82-d4df-4df9-9c3b-35fc9f990a67" containerName="registry-server" containerID="cri-o://f962de81a8bdda735e833a62d37ee650546560c30cfa8115f41bab1aef2fb361" gracePeriod=2 Mar 18 21:10:28 crc kubenswrapper[4950]: I0318 21:10:28.743693 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9x7pn" Mar 18 21:10:28 crc kubenswrapper[4950]: I0318 21:10:28.920356 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkhsr\" (UniqueName: \"kubernetes.io/projected/73e71e82-d4df-4df9-9c3b-35fc9f990a67-kube-api-access-kkhsr\") pod \"73e71e82-d4df-4df9-9c3b-35fc9f990a67\" (UID: \"73e71e82-d4df-4df9-9c3b-35fc9f990a67\") " Mar 18 21:10:28 crc kubenswrapper[4950]: I0318 21:10:28.920592 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73e71e82-d4df-4df9-9c3b-35fc9f990a67-catalog-content\") pod \"73e71e82-d4df-4df9-9c3b-35fc9f990a67\" (UID: \"73e71e82-d4df-4df9-9c3b-35fc9f990a67\") " Mar 18 21:10:28 crc kubenswrapper[4950]: I0318 21:10:28.920823 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73e71e82-d4df-4df9-9c3b-35fc9f990a67-utilities\") pod \"73e71e82-d4df-4df9-9c3b-35fc9f990a67\" (UID: \"73e71e82-d4df-4df9-9c3b-35fc9f990a67\") " Mar 18 21:10:28 crc kubenswrapper[4950]: I0318 21:10:28.921708 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73e71e82-d4df-4df9-9c3b-35fc9f990a67-utilities" (OuterVolumeSpecName: "utilities") pod "73e71e82-d4df-4df9-9c3b-35fc9f990a67" (UID: "73e71e82-d4df-4df9-9c3b-35fc9f990a67"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:10:28 crc kubenswrapper[4950]: I0318 21:10:28.940517 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73e71e82-d4df-4df9-9c3b-35fc9f990a67-kube-api-access-kkhsr" (OuterVolumeSpecName: "kube-api-access-kkhsr") pod "73e71e82-d4df-4df9-9c3b-35fc9f990a67" (UID: "73e71e82-d4df-4df9-9c3b-35fc9f990a67"). InnerVolumeSpecName "kube-api-access-kkhsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:10:28 crc kubenswrapper[4950]: I0318 21:10:28.979333 4950 generic.go:334] "Generic (PLEG): container finished" podID="73e71e82-d4df-4df9-9c3b-35fc9f990a67" containerID="f962de81a8bdda735e833a62d37ee650546560c30cfa8115f41bab1aef2fb361" exitCode=0 Mar 18 21:10:28 crc kubenswrapper[4950]: I0318 21:10:28.979416 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9x7pn" Mar 18 21:10:28 crc kubenswrapper[4950]: I0318 21:10:28.979397 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9x7pn" event={"ID":"73e71e82-d4df-4df9-9c3b-35fc9f990a67","Type":"ContainerDied","Data":"f962de81a8bdda735e833a62d37ee650546560c30cfa8115f41bab1aef2fb361"} Mar 18 21:10:28 crc kubenswrapper[4950]: I0318 21:10:28.979572 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9x7pn" event={"ID":"73e71e82-d4df-4df9-9c3b-35fc9f990a67","Type":"ContainerDied","Data":"d7e91317d8f63c6eee32619a46a766878956dc6696210055bda7f4e9ea2b252b"} Mar 18 21:10:28 crc kubenswrapper[4950]: I0318 21:10:28.979599 4950 scope.go:117] "RemoveContainer" containerID="f962de81a8bdda735e833a62d37ee650546560c30cfa8115f41bab1aef2fb361" Mar 18 21:10:29 crc kubenswrapper[4950]: I0318 21:10:29.007580 4950 scope.go:117] "RemoveContainer" containerID="e62985af67b08d18284cda3e149c4040fcaeeb367cea820cacde4889dff86c02" Mar 18 21:10:29 crc kubenswrapper[4950]: I0318 21:10:29.022794 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73e71e82-d4df-4df9-9c3b-35fc9f990a67-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73e71e82-d4df-4df9-9c3b-35fc9f990a67" (UID: "73e71e82-d4df-4df9-9c3b-35fc9f990a67"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:10:29 crc kubenswrapper[4950]: I0318 21:10:29.024134 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73e71e82-d4df-4df9-9c3b-35fc9f990a67-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 21:10:29 crc kubenswrapper[4950]: I0318 21:10:29.024167 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73e71e82-d4df-4df9-9c3b-35fc9f990a67-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 21:10:29 crc kubenswrapper[4950]: I0318 21:10:29.024177 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkhsr\" (UniqueName: \"kubernetes.io/projected/73e71e82-d4df-4df9-9c3b-35fc9f990a67-kube-api-access-kkhsr\") on node \"crc\" DevicePath \"\"" Mar 18 21:10:29 crc kubenswrapper[4950]: I0318 21:10:29.035835 4950 scope.go:117] "RemoveContainer" containerID="63dc5fa85995b3a4729f98640763abca5646338ed067cbb6c9ced6c29ca204f1" Mar 18 21:10:29 crc kubenswrapper[4950]: I0318 21:10:29.101985 4950 scope.go:117] "RemoveContainer" containerID="f962de81a8bdda735e833a62d37ee650546560c30cfa8115f41bab1aef2fb361" Mar 18 21:10:29 crc kubenswrapper[4950]: E0318 21:10:29.103269 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f962de81a8bdda735e833a62d37ee650546560c30cfa8115f41bab1aef2fb361\": container with ID starting with f962de81a8bdda735e833a62d37ee650546560c30cfa8115f41bab1aef2fb361 not found: ID does not exist" containerID="f962de81a8bdda735e833a62d37ee650546560c30cfa8115f41bab1aef2fb361" Mar 18 21:10:29 crc kubenswrapper[4950]: I0318 21:10:29.103309 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f962de81a8bdda735e833a62d37ee650546560c30cfa8115f41bab1aef2fb361"} err="failed to get container status \"f962de81a8bdda735e833a62d37ee650546560c30cfa8115f41bab1aef2fb361\": rpc error: code = NotFound desc = could not find container \"f962de81a8bdda735e833a62d37ee650546560c30cfa8115f41bab1aef2fb361\": container with ID starting with f962de81a8bdda735e833a62d37ee650546560c30cfa8115f41bab1aef2fb361 not found: ID does not exist" Mar 18 21:10:29 crc kubenswrapper[4950]: I0318 21:10:29.103331 4950 scope.go:117] "RemoveContainer" containerID="e62985af67b08d18284cda3e149c4040fcaeeb367cea820cacde4889dff86c02" Mar 18 21:10:29 crc kubenswrapper[4950]: E0318 21:10:29.103831 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e62985af67b08d18284cda3e149c4040fcaeeb367cea820cacde4889dff86c02\": container with ID starting with e62985af67b08d18284cda3e149c4040fcaeeb367cea820cacde4889dff86c02 not found: ID does not exist" containerID="e62985af67b08d18284cda3e149c4040fcaeeb367cea820cacde4889dff86c02" Mar 18 21:10:29 crc kubenswrapper[4950]: I0318 21:10:29.103853 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e62985af67b08d18284cda3e149c4040fcaeeb367cea820cacde4889dff86c02"} err="failed to get container status \"e62985af67b08d18284cda3e149c4040fcaeeb367cea820cacde4889dff86c02\": rpc error: code = NotFound desc = could not find container \"e62985af67b08d18284cda3e149c4040fcaeeb367cea820cacde4889dff86c02\": container with ID starting with e62985af67b08d18284cda3e149c4040fcaeeb367cea820cacde4889dff86c02 not found: ID does not exist" Mar 18 21:10:29 crc kubenswrapper[4950]: I0318 21:10:29.103896 4950 scope.go:117] "RemoveContainer" containerID="63dc5fa85995b3a4729f98640763abca5646338ed067cbb6c9ced6c29ca204f1" Mar 18 21:10:29 crc kubenswrapper[4950]: E0318 21:10:29.105384 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63dc5fa85995b3a4729f98640763abca5646338ed067cbb6c9ced6c29ca204f1\": container with ID starting with 63dc5fa85995b3a4729f98640763abca5646338ed067cbb6c9ced6c29ca204f1 not found: ID does not exist" containerID="63dc5fa85995b3a4729f98640763abca5646338ed067cbb6c9ced6c29ca204f1" Mar 18 21:10:29 crc kubenswrapper[4950]: I0318 21:10:29.105428 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63dc5fa85995b3a4729f98640763abca5646338ed067cbb6c9ced6c29ca204f1"} err="failed to get container status \"63dc5fa85995b3a4729f98640763abca5646338ed067cbb6c9ced6c29ca204f1\": rpc error: code = NotFound desc = could not find container \"63dc5fa85995b3a4729f98640763abca5646338ed067cbb6c9ced6c29ca204f1\": container with ID starting with 63dc5fa85995b3a4729f98640763abca5646338ed067cbb6c9ced6c29ca204f1 not found: ID does not exist" Mar 18 21:10:29 crc kubenswrapper[4950]: I0318 21:10:29.319099 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9x7pn"] Mar 18 21:10:29 crc kubenswrapper[4950]: I0318 21:10:29.330048 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9x7pn"] Mar 18 21:10:29 crc kubenswrapper[4950]: I0318 21:10:29.490525 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73e71e82-d4df-4df9-9c3b-35fc9f990a67" path="/var/lib/kubelet/pods/73e71e82-d4df-4df9-9c3b-35fc9f990a67/volumes" Mar 18 21:10:33 crc kubenswrapper[4950]: I0318 21:10:33.341381 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-42xf9" podUID="d72524a5-b7b7-416f-b9ba-9527f98e787c" containerName="registry-server" probeResult="failure" output=< Mar 18 21:10:33 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 21:10:33 crc kubenswrapper[4950]: > Mar 18 21:10:42 crc kubenswrapper[4950]: I0318 21:10:42.340401 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-42xf9" Mar 18 21:10:42 crc kubenswrapper[4950]: I0318 21:10:42.393189 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-42xf9" Mar 18 21:10:42 crc kubenswrapper[4950]: I0318 21:10:42.492938 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-42xf9"] Mar 18 21:10:42 crc kubenswrapper[4950]: I0318 21:10:42.582089 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ldrm5"] Mar 18 21:10:42 crc kubenswrapper[4950]: I0318 21:10:42.582592 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ldrm5" podUID="81fa390c-72e6-471b-ac97-8e5d5fc3b4d7" containerName="registry-server" containerID="cri-o://faa09adf7d326e2afe7514093466e5cd0922ed36ed761b54afb1c4b3a8ad9764" gracePeriod=2 Mar 18 21:10:43 crc kubenswrapper[4950]: I0318 21:10:43.123996 4950 generic.go:334] "Generic (PLEG): container finished" podID="81fa390c-72e6-471b-ac97-8e5d5fc3b4d7" containerID="faa09adf7d326e2afe7514093466e5cd0922ed36ed761b54afb1c4b3a8ad9764" exitCode=0 Mar 18 21:10:43 crc kubenswrapper[4950]: I0318 21:10:43.124168 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldrm5" event={"ID":"81fa390c-72e6-471b-ac97-8e5d5fc3b4d7","Type":"ContainerDied","Data":"faa09adf7d326e2afe7514093466e5cd0922ed36ed761b54afb1c4b3a8ad9764"} Mar 18 21:10:43 crc kubenswrapper[4950]: I0318 21:10:43.501804 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ldrm5" Mar 18 21:10:43 crc kubenswrapper[4950]: I0318 21:10:43.623184 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81fa390c-72e6-471b-ac97-8e5d5fc3b4d7-catalog-content\") pod \"81fa390c-72e6-471b-ac97-8e5d5fc3b4d7\" (UID: \"81fa390c-72e6-471b-ac97-8e5d5fc3b4d7\") " Mar 18 21:10:43 crc kubenswrapper[4950]: I0318 21:10:43.623377 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81fa390c-72e6-471b-ac97-8e5d5fc3b4d7-utilities\") pod \"81fa390c-72e6-471b-ac97-8e5d5fc3b4d7\" (UID: \"81fa390c-72e6-471b-ac97-8e5d5fc3b4d7\") " Mar 18 21:10:43 crc kubenswrapper[4950]: I0318 21:10:43.623401 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbllm\" (UniqueName: \"kubernetes.io/projected/81fa390c-72e6-471b-ac97-8e5d5fc3b4d7-kube-api-access-rbllm\") pod \"81fa390c-72e6-471b-ac97-8e5d5fc3b4d7\" (UID: \"81fa390c-72e6-471b-ac97-8e5d5fc3b4d7\") " Mar 18 21:10:43 crc kubenswrapper[4950]: I0318 21:10:43.623822 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81fa390c-72e6-471b-ac97-8e5d5fc3b4d7-utilities" (OuterVolumeSpecName: "utilities") pod "81fa390c-72e6-471b-ac97-8e5d5fc3b4d7" (UID: "81fa390c-72e6-471b-ac97-8e5d5fc3b4d7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:10:43 crc kubenswrapper[4950]: I0318 21:10:43.625441 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81fa390c-72e6-471b-ac97-8e5d5fc3b4d7-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 21:10:43 crc kubenswrapper[4950]: I0318 21:10:43.629729 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81fa390c-72e6-471b-ac97-8e5d5fc3b4d7-kube-api-access-rbllm" (OuterVolumeSpecName: "kube-api-access-rbllm") pod "81fa390c-72e6-471b-ac97-8e5d5fc3b4d7" (UID: "81fa390c-72e6-471b-ac97-8e5d5fc3b4d7"). InnerVolumeSpecName "kube-api-access-rbllm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:10:43 crc kubenswrapper[4950]: I0318 21:10:43.727657 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbllm\" (UniqueName: \"kubernetes.io/projected/81fa390c-72e6-471b-ac97-8e5d5fc3b4d7-kube-api-access-rbllm\") on node \"crc\" DevicePath \"\"" Mar 18 21:10:43 crc kubenswrapper[4950]: I0318 21:10:43.778318 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81fa390c-72e6-471b-ac97-8e5d5fc3b4d7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "81fa390c-72e6-471b-ac97-8e5d5fc3b4d7" (UID: "81fa390c-72e6-471b-ac97-8e5d5fc3b4d7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:10:43 crc kubenswrapper[4950]: I0318 21:10:43.828949 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81fa390c-72e6-471b-ac97-8e5d5fc3b4d7-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 21:10:44 crc kubenswrapper[4950]: I0318 21:10:44.134332 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ldrm5" event={"ID":"81fa390c-72e6-471b-ac97-8e5d5fc3b4d7","Type":"ContainerDied","Data":"112ea6205059f5c043f1fd27d088e60bf5cfc48ffa3a7aa1d5395821da8cda98"} Mar 18 21:10:44 crc kubenswrapper[4950]: I0318 21:10:44.134819 4950 scope.go:117] "RemoveContainer" containerID="faa09adf7d326e2afe7514093466e5cd0922ed36ed761b54afb1c4b3a8ad9764" Mar 18 21:10:44 crc kubenswrapper[4950]: I0318 21:10:44.134361 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ldrm5" Mar 18 21:10:44 crc kubenswrapper[4950]: I0318 21:10:44.162655 4950 scope.go:117] "RemoveContainer" containerID="346c36d7293d3fafd4d85fd5487209b524537b640bce764bf8408c27935c20c5" Mar 18 21:10:44 crc kubenswrapper[4950]: I0318 21:10:44.167485 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ldrm5"] Mar 18 21:10:44 crc kubenswrapper[4950]: I0318 21:10:44.187616 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ldrm5"] Mar 18 21:10:44 crc kubenswrapper[4950]: I0318 21:10:44.198696 4950 scope.go:117] "RemoveContainer" containerID="f3289f682a363386458f56a7c14afd5877d045a384db2ec36ea2e8fc0796af34" Mar 18 21:10:45 crc kubenswrapper[4950]: I0318 21:10:45.490703 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81fa390c-72e6-471b-ac97-8e5d5fc3b4d7" path="/var/lib/kubelet/pods/81fa390c-72e6-471b-ac97-8e5d5fc3b4d7/volumes" Mar 18 21:11:00 crc kubenswrapper[4950]: I0318 21:11:00.412090 4950 scope.go:117] "RemoveContainer" containerID="a1b7b36b64c69a4bc5dbf5bc4c35c0f6a0d2d705da4804ce955847c102f01be3" Mar 18 21:11:00 crc kubenswrapper[4950]: I0318 21:11:00.444329 4950 scope.go:117] "RemoveContainer" containerID="ecbd941aa773430e85ac00d884f2e32bd0a9f5c3dd92723c7c236aac7e3125a7" Mar 18 21:11:03 crc kubenswrapper[4950]: I0318 21:11:03.836898 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 21:11:03 crc kubenswrapper[4950]: I0318 21:11:03.837428 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 21:11:33 crc kubenswrapper[4950]: I0318 21:11:33.836779 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 21:11:33 crc kubenswrapper[4950]: I0318 21:11:33.837353 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 21:12:00 crc kubenswrapper[4950]: I0318 21:12:00.170089 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564472-fjh27"] Mar 18 21:12:00 crc kubenswrapper[4950]: E0318 21:12:00.172780 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73e71e82-d4df-4df9-9c3b-35fc9f990a67" containerName="registry-server" Mar 18 21:12:00 crc kubenswrapper[4950]: I0318 21:12:00.172947 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="73e71e82-d4df-4df9-9c3b-35fc9f990a67" containerName="registry-server" Mar 18 21:12:00 crc kubenswrapper[4950]: E0318 21:12:00.173073 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73e71e82-d4df-4df9-9c3b-35fc9f990a67" containerName="extract-utilities" Mar 18 21:12:00 crc kubenswrapper[4950]: I0318 21:12:00.173185 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="73e71e82-d4df-4df9-9c3b-35fc9f990a67" containerName="extract-utilities" Mar 18 21:12:00 crc kubenswrapper[4950]: E0318 21:12:00.173329 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81fa390c-72e6-471b-ac97-8e5d5fc3b4d7" containerName="extract-utilities" Mar 18 21:12:00 crc kubenswrapper[4950]: I0318 21:12:00.173478 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="81fa390c-72e6-471b-ac97-8e5d5fc3b4d7" containerName="extract-utilities" Mar 18 21:12:00 crc kubenswrapper[4950]: E0318 21:12:00.173627 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81fa390c-72e6-471b-ac97-8e5d5fc3b4d7" containerName="registry-server" Mar 18 21:12:00 crc kubenswrapper[4950]: I0318 21:12:00.173718 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="81fa390c-72e6-471b-ac97-8e5d5fc3b4d7" containerName="registry-server" Mar 18 21:12:00 crc kubenswrapper[4950]: E0318 21:12:00.173830 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73e71e82-d4df-4df9-9c3b-35fc9f990a67" containerName="extract-content" Mar 18 21:12:00 crc kubenswrapper[4950]: I0318 21:12:00.173933 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="73e71e82-d4df-4df9-9c3b-35fc9f990a67" containerName="extract-content" Mar 18 21:12:00 crc kubenswrapper[4950]: E0318 21:12:00.174041 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81fa390c-72e6-471b-ac97-8e5d5fc3b4d7" containerName="extract-content" Mar 18 21:12:00 crc kubenswrapper[4950]: I0318 21:12:00.174153 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="81fa390c-72e6-471b-ac97-8e5d5fc3b4d7" containerName="extract-content" Mar 18 21:12:00 crc kubenswrapper[4950]: E0318 21:12:00.174290 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7142cc1f-1577-45b7-bc79-a4dd4416455e" containerName="oc" Mar 18 21:12:00 crc kubenswrapper[4950]: I0318 21:12:00.174550 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="7142cc1f-1577-45b7-bc79-a4dd4416455e" containerName="oc" Mar 18 21:12:00 crc kubenswrapper[4950]: I0318 21:12:00.175046 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="7142cc1f-1577-45b7-bc79-a4dd4416455e" containerName="oc" Mar 18 21:12:00 crc kubenswrapper[4950]: I0318 21:12:00.175188 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="81fa390c-72e6-471b-ac97-8e5d5fc3b4d7" containerName="registry-server" Mar 18 21:12:00 crc kubenswrapper[4950]: I0318 21:12:00.175306 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="73e71e82-d4df-4df9-9c3b-35fc9f990a67" containerName="registry-server" Mar 18 21:12:00 crc kubenswrapper[4950]: I0318 21:12:00.176603 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564472-fjh27" Mar 18 21:12:00 crc kubenswrapper[4950]: I0318 21:12:00.179816 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 21:12:00 crc kubenswrapper[4950]: I0318 21:12:00.180083 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 21:12:00 crc kubenswrapper[4950]: I0318 21:12:00.180223 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 21:12:00 crc kubenswrapper[4950]: I0318 21:12:00.198755 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564472-fjh27"] Mar 18 21:12:00 crc kubenswrapper[4950]: I0318 21:12:00.210716 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5sqm\" (UniqueName: \"kubernetes.io/projected/cdd5815f-30c3-40b6-89fe-9fa8d7707928-kube-api-access-c5sqm\") pod \"auto-csr-approver-29564472-fjh27\" (UID: \"cdd5815f-30c3-40b6-89fe-9fa8d7707928\") " pod="openshift-infra/auto-csr-approver-29564472-fjh27" Mar 18 21:12:00 crc kubenswrapper[4950]: I0318 21:12:00.312678 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5sqm\" (UniqueName: \"kubernetes.io/projected/cdd5815f-30c3-40b6-89fe-9fa8d7707928-kube-api-access-c5sqm\") pod \"auto-csr-approver-29564472-fjh27\" (UID: \"cdd5815f-30c3-40b6-89fe-9fa8d7707928\") " pod="openshift-infra/auto-csr-approver-29564472-fjh27" Mar 18 21:12:00 crc kubenswrapper[4950]: I0318 21:12:00.333724 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5sqm\" (UniqueName: \"kubernetes.io/projected/cdd5815f-30c3-40b6-89fe-9fa8d7707928-kube-api-access-c5sqm\") pod \"auto-csr-approver-29564472-fjh27\" (UID: \"cdd5815f-30c3-40b6-89fe-9fa8d7707928\") " pod="openshift-infra/auto-csr-approver-29564472-fjh27" Mar 18 21:12:00 crc kubenswrapper[4950]: I0318 21:12:00.503950 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564472-fjh27" Mar 18 21:12:01 crc kubenswrapper[4950]: I0318 21:12:01.035169 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564472-fjh27"] Mar 18 21:12:01 crc kubenswrapper[4950]: I0318 21:12:01.051986 4950 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 21:12:01 crc kubenswrapper[4950]: I0318 21:12:01.922263 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564472-fjh27" event={"ID":"cdd5815f-30c3-40b6-89fe-9fa8d7707928","Type":"ContainerStarted","Data":"15f2d84a73d416d8b1e52b4ebe4be82eb0de1031e19dae3a42f2ce0f856f262b"} Mar 18 21:12:02 crc kubenswrapper[4950]: I0318 21:12:02.931474 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564472-fjh27" event={"ID":"cdd5815f-30c3-40b6-89fe-9fa8d7707928","Type":"ContainerStarted","Data":"8370b7e5b41bf9fbf5b4731d3ba51b27a9d591d7f40c6a149850c86d5f306c50"} Mar 18 21:12:02 crc kubenswrapper[4950]: I0318 21:12:02.956711 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29564472-fjh27" podStartSLOduration=2.049286142 podStartE2EDuration="2.956692925s" podCreationTimestamp="2026-03-18 21:12:00 +0000 UTC" firstStartedPulling="2026-03-18 21:12:01.05161478 +0000 UTC m=+3934.292456688" lastFinishedPulling="2026-03-18 21:12:01.959021583 +0000 UTC m=+3935.199863471" observedRunningTime="2026-03-18 21:12:02.9462 +0000 UTC m=+3936.187041868" watchObservedRunningTime="2026-03-18 21:12:02.956692925 +0000 UTC m=+3936.197534793" Mar 18 21:12:03 crc kubenswrapper[4950]: I0318 21:12:03.836204 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 21:12:03 crc kubenswrapper[4950]: I0318 21:12:03.836295 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 21:12:03 crc kubenswrapper[4950]: I0318 21:12:03.836378 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 21:12:03 crc kubenswrapper[4950]: I0318 21:12:03.836974 4950 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9"} pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 21:12:03 crc kubenswrapper[4950]: I0318 21:12:03.837040 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" containerID="cri-o://8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" gracePeriod=600 Mar 18 21:12:03 crc kubenswrapper[4950]: I0318 21:12:03.941551 4950 generic.go:334] "Generic (PLEG): container finished" podID="cdd5815f-30c3-40b6-89fe-9fa8d7707928" containerID="8370b7e5b41bf9fbf5b4731d3ba51b27a9d591d7f40c6a149850c86d5f306c50" exitCode=0 Mar 18 21:12:03 crc kubenswrapper[4950]: I0318 21:12:03.941764 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564472-fjh27" event={"ID":"cdd5815f-30c3-40b6-89fe-9fa8d7707928","Type":"ContainerDied","Data":"8370b7e5b41bf9fbf5b4731d3ba51b27a9d591d7f40c6a149850c86d5f306c50"} Mar 18 21:12:03 crc kubenswrapper[4950]: E0318 21:12:03.987074 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:12:04 crc kubenswrapper[4950]: I0318 21:12:04.958464 4950 generic.go:334] "Generic (PLEG): container finished" podID="4048b439-3266-46e7-9de0-22377efacc46" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" exitCode=0 Mar 18 21:12:04 crc kubenswrapper[4950]: I0318 21:12:04.958544 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerDied","Data":"8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9"} Mar 18 21:12:04 crc kubenswrapper[4950]: I0318 21:12:04.958593 4950 scope.go:117] "RemoveContainer" containerID="16b4cab9ed932e2b4b0bb08ce747a1dbd93ab41574373696a03e18ef0ef41c79" Mar 18 21:12:04 crc kubenswrapper[4950]: I0318 21:12:04.959364 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:12:04 crc kubenswrapper[4950]: E0318 21:12:04.959791 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:12:06 crc kubenswrapper[4950]: I0318 21:12:05.981243 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564472-fjh27" event={"ID":"cdd5815f-30c3-40b6-89fe-9fa8d7707928","Type":"ContainerDied","Data":"15f2d84a73d416d8b1e52b4ebe4be82eb0de1031e19dae3a42f2ce0f856f262b"} Mar 18 21:12:06 crc kubenswrapper[4950]: I0318 21:12:05.981547 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15f2d84a73d416d8b1e52b4ebe4be82eb0de1031e19dae3a42f2ce0f856f262b" Mar 18 21:12:06 crc kubenswrapper[4950]: I0318 21:12:06.062462 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564472-fjh27" Mar 18 21:12:06 crc kubenswrapper[4950]: I0318 21:12:06.186358 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5sqm\" (UniqueName: \"kubernetes.io/projected/cdd5815f-30c3-40b6-89fe-9fa8d7707928-kube-api-access-c5sqm\") pod \"cdd5815f-30c3-40b6-89fe-9fa8d7707928\" (UID: \"cdd5815f-30c3-40b6-89fe-9fa8d7707928\") " Mar 18 21:12:06 crc kubenswrapper[4950]: I0318 21:12:06.195634 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdd5815f-30c3-40b6-89fe-9fa8d7707928-kube-api-access-c5sqm" (OuterVolumeSpecName: "kube-api-access-c5sqm") pod "cdd5815f-30c3-40b6-89fe-9fa8d7707928" (UID: "cdd5815f-30c3-40b6-89fe-9fa8d7707928"). InnerVolumeSpecName "kube-api-access-c5sqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:12:06 crc kubenswrapper[4950]: I0318 21:12:06.288738 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5sqm\" (UniqueName: \"kubernetes.io/projected/cdd5815f-30c3-40b6-89fe-9fa8d7707928-kube-api-access-c5sqm\") on node \"crc\" DevicePath \"\"" Mar 18 21:12:06 crc kubenswrapper[4950]: I0318 21:12:06.987611 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564472-fjh27" Mar 18 21:12:07 crc kubenswrapper[4950]: I0318 21:12:07.169342 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564466-vhzmh"] Mar 18 21:12:07 crc kubenswrapper[4950]: I0318 21:12:07.177930 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564466-vhzmh"] Mar 18 21:12:07 crc kubenswrapper[4950]: I0318 21:12:07.522347 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="218bd624-88ee-45a0-8d81-756534613103" path="/var/lib/kubelet/pods/218bd624-88ee-45a0-8d81-756534613103/volumes" Mar 18 21:12:17 crc kubenswrapper[4950]: I0318 21:12:17.485685 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:12:17 crc kubenswrapper[4950]: E0318 21:12:17.486386 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:12:31 crc kubenswrapper[4950]: I0318 21:12:31.480311 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:12:31 crc kubenswrapper[4950]: E0318 21:12:31.480904 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:12:45 crc kubenswrapper[4950]: I0318 21:12:45.484797 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:12:45 crc kubenswrapper[4950]: E0318 21:12:45.486920 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:13:00 crc kubenswrapper[4950]: I0318 21:13:00.481060 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:13:00 crc kubenswrapper[4950]: E0318 21:13:00.481818 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:13:00 crc kubenswrapper[4950]: I0318 21:13:00.669342 4950 scope.go:117] "RemoveContainer" containerID="c6d2053d2c61428bc84b0fce73ba188ed92f3c0a4dfe1c721f4077c524d9abf5" Mar 18 21:13:11 crc kubenswrapper[4950]: I0318 21:13:11.480558 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:13:11 crc kubenswrapper[4950]: E0318 21:13:11.481246 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:13:22 crc kubenswrapper[4950]: I0318 21:13:22.479996 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:13:22 crc kubenswrapper[4950]: E0318 21:13:22.480805 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:13:33 crc kubenswrapper[4950]: I0318 21:13:33.480177 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:13:33 crc kubenswrapper[4950]: E0318 21:13:33.482196 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:13:47 crc kubenswrapper[4950]: I0318 21:13:47.490281 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:13:47 crc kubenswrapper[4950]: E0318 21:13:47.492088 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:13:58 crc kubenswrapper[4950]: I0318 21:13:58.479983 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:13:58 crc kubenswrapper[4950]: E0318 21:13:58.480766 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:14:00 crc kubenswrapper[4950]: I0318 21:14:00.156808 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564474-tbsqb"] Mar 18 21:14:00 crc kubenswrapper[4950]: E0318 21:14:00.157316 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdd5815f-30c3-40b6-89fe-9fa8d7707928" containerName="oc" Mar 18 21:14:00 crc kubenswrapper[4950]: I0318 21:14:00.157331 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdd5815f-30c3-40b6-89fe-9fa8d7707928" containerName="oc" Mar 18 21:14:00 crc kubenswrapper[4950]: I0318 21:14:00.157645 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdd5815f-30c3-40b6-89fe-9fa8d7707928" containerName="oc" Mar 18 21:14:00 crc kubenswrapper[4950]: I0318 21:14:00.158447 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564474-tbsqb" Mar 18 21:14:00 crc kubenswrapper[4950]: I0318 21:14:00.162329 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 21:14:00 crc kubenswrapper[4950]: I0318 21:14:00.162912 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 21:14:00 crc kubenswrapper[4950]: I0318 21:14:00.163054 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 21:14:00 crc kubenswrapper[4950]: I0318 21:14:00.167137 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564474-tbsqb"] Mar 18 21:14:00 crc kubenswrapper[4950]: I0318 21:14:00.264228 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fslnm\" (UniqueName: \"kubernetes.io/projected/6d38b254-beb1-45fa-b20d-2b9c082c00b4-kube-api-access-fslnm\") pod \"auto-csr-approver-29564474-tbsqb\" (UID: \"6d38b254-beb1-45fa-b20d-2b9c082c00b4\") " pod="openshift-infra/auto-csr-approver-29564474-tbsqb" Mar 18 21:14:00 crc kubenswrapper[4950]: I0318 21:14:00.367795 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fslnm\" (UniqueName: \"kubernetes.io/projected/6d38b254-beb1-45fa-b20d-2b9c082c00b4-kube-api-access-fslnm\") pod \"auto-csr-approver-29564474-tbsqb\" (UID: \"6d38b254-beb1-45fa-b20d-2b9c082c00b4\") " pod="openshift-infra/auto-csr-approver-29564474-tbsqb" Mar 18 21:14:00 crc kubenswrapper[4950]: I0318 21:14:00.387923 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fslnm\" (UniqueName: \"kubernetes.io/projected/6d38b254-beb1-45fa-b20d-2b9c082c00b4-kube-api-access-fslnm\") pod \"auto-csr-approver-29564474-tbsqb\" (UID: \"6d38b254-beb1-45fa-b20d-2b9c082c00b4\") " pod="openshift-infra/auto-csr-approver-29564474-tbsqb" Mar 18 21:14:00 crc kubenswrapper[4950]: I0318 21:14:00.483030 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564474-tbsqb" Mar 18 21:14:00 crc kubenswrapper[4950]: I0318 21:14:00.923742 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564474-tbsqb"] Mar 18 21:14:01 crc kubenswrapper[4950]: I0318 21:14:01.004566 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564474-tbsqb" event={"ID":"6d38b254-beb1-45fa-b20d-2b9c082c00b4","Type":"ContainerStarted","Data":"df65f198e675c6d3e2d0d5e8e73b93dff1430f442d8954a44e2ac07428806a3f"} Mar 18 21:14:03 crc kubenswrapper[4950]: I0318 21:14:03.021733 4950 generic.go:334] "Generic (PLEG): container finished" podID="6d38b254-beb1-45fa-b20d-2b9c082c00b4" containerID="5e4f099e712c5548236f505a3a9a4c4486f422245e990ee927e547af07b8e21a" exitCode=0 Mar 18 21:14:03 crc kubenswrapper[4950]: I0318 21:14:03.021852 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564474-tbsqb" event={"ID":"6d38b254-beb1-45fa-b20d-2b9c082c00b4","Type":"ContainerDied","Data":"5e4f099e712c5548236f505a3a9a4c4486f422245e990ee927e547af07b8e21a"} Mar 18 21:14:04 crc kubenswrapper[4950]: I0318 21:14:04.453659 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564474-tbsqb" Mar 18 21:14:04 crc kubenswrapper[4950]: I0318 21:14:04.561369 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fslnm\" (UniqueName: \"kubernetes.io/projected/6d38b254-beb1-45fa-b20d-2b9c082c00b4-kube-api-access-fslnm\") pod \"6d38b254-beb1-45fa-b20d-2b9c082c00b4\" (UID: \"6d38b254-beb1-45fa-b20d-2b9c082c00b4\") " Mar 18 21:14:04 crc kubenswrapper[4950]: I0318 21:14:04.570667 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d38b254-beb1-45fa-b20d-2b9c082c00b4-kube-api-access-fslnm" (OuterVolumeSpecName: "kube-api-access-fslnm") pod "6d38b254-beb1-45fa-b20d-2b9c082c00b4" (UID: "6d38b254-beb1-45fa-b20d-2b9c082c00b4"). InnerVolumeSpecName "kube-api-access-fslnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:14:04 crc kubenswrapper[4950]: I0318 21:14:04.665175 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fslnm\" (UniqueName: \"kubernetes.io/projected/6d38b254-beb1-45fa-b20d-2b9c082c00b4-kube-api-access-fslnm\") on node \"crc\" DevicePath \"\"" Mar 18 21:14:05 crc kubenswrapper[4950]: I0318 21:14:05.040550 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564474-tbsqb" event={"ID":"6d38b254-beb1-45fa-b20d-2b9c082c00b4","Type":"ContainerDied","Data":"df65f198e675c6d3e2d0d5e8e73b93dff1430f442d8954a44e2ac07428806a3f"} Mar 18 21:14:05 crc kubenswrapper[4950]: I0318 21:14:05.040589 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df65f198e675c6d3e2d0d5e8e73b93dff1430f442d8954a44e2ac07428806a3f" Mar 18 21:14:05 crc kubenswrapper[4950]: I0318 21:14:05.040664 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564474-tbsqb" Mar 18 21:14:05 crc kubenswrapper[4950]: I0318 21:14:05.573763 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564468-2cvt2"] Mar 18 21:14:05 crc kubenswrapper[4950]: I0318 21:14:05.583468 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564468-2cvt2"] Mar 18 21:14:07 crc kubenswrapper[4950]: I0318 21:14:07.493142 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1deaedc-119e-4e3b-ad75-86eda667d445" path="/var/lib/kubelet/pods/c1deaedc-119e-4e3b-ad75-86eda667d445/volumes" Mar 18 21:14:10 crc kubenswrapper[4950]: I0318 21:14:10.479657 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:14:10 crc kubenswrapper[4950]: E0318 21:14:10.480283 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:14:22 crc kubenswrapper[4950]: I0318 21:14:22.479652 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:14:22 crc kubenswrapper[4950]: E0318 21:14:22.480768 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:14:36 crc kubenswrapper[4950]: I0318 21:14:36.479554 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:14:36 crc kubenswrapper[4950]: E0318 21:14:36.480363 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:14:41 crc kubenswrapper[4950]: I0318 21:14:41.442813 4950 generic.go:334] "Generic (PLEG): container finished" podID="3c1bc0a9-e172-4749-b998-710b5a306618" containerID="cef779de1979e305cdbe9749cc9f3f279e8b05db43fcbd358178c77a68d20ca7" exitCode=0 Mar 18 21:14:41 crc kubenswrapper[4950]: I0318 21:14:41.442940 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"3c1bc0a9-e172-4749-b998-710b5a306618","Type":"ContainerDied","Data":"cef779de1979e305cdbe9749cc9f3f279e8b05db43fcbd358178c77a68d20ca7"} Mar 18 21:14:42 crc kubenswrapper[4950]: I0318 21:14:42.869538 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.003924 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3c1bc0a9-e172-4749-b998-710b5a306618-test-operator-ephemeral-temporary\") pod \"3c1bc0a9-e172-4749-b998-710b5a306618\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.004596 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3c1bc0a9-e172-4749-b998-710b5a306618-ca-certs\") pod \"3c1bc0a9-e172-4749-b998-710b5a306618\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.004754 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c1bc0a9-e172-4749-b998-710b5a306618-config-data\") pod \"3c1bc0a9-e172-4749-b998-710b5a306618\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.004953 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c1bc0a9-e172-4749-b998-710b5a306618-ssh-key\") pod \"3c1bc0a9-e172-4749-b998-710b5a306618\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.005105 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"3c1bc0a9-e172-4749-b998-710b5a306618\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.005210 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3c1bc0a9-e172-4749-b998-710b5a306618-openstack-config-secret\") pod \"3c1bc0a9-e172-4749-b998-710b5a306618\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.005462 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2wxj\" (UniqueName: \"kubernetes.io/projected/3c1bc0a9-e172-4749-b998-710b5a306618-kube-api-access-b2wxj\") pod \"3c1bc0a9-e172-4749-b998-710b5a306618\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.005561 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3c1bc0a9-e172-4749-b998-710b5a306618-test-operator-ephemeral-workdir\") pod \"3c1bc0a9-e172-4749-b998-710b5a306618\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.005984 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3c1bc0a9-e172-4749-b998-710b5a306618-openstack-config\") pod \"3c1bc0a9-e172-4749-b998-710b5a306618\" (UID: \"3c1bc0a9-e172-4749-b998-710b5a306618\") " Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.006574 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c1bc0a9-e172-4749-b998-710b5a306618-config-data" (OuterVolumeSpecName: "config-data") pod "3c1bc0a9-e172-4749-b998-710b5a306618" (UID: "3c1bc0a9-e172-4749-b998-710b5a306618"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.006803 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c1bc0a9-e172-4749-b998-710b5a306618-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "3c1bc0a9-e172-4749-b998-710b5a306618" (UID: "3c1bc0a9-e172-4749-b998-710b5a306618"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.006921 4950 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c1bc0a9-e172-4749-b998-710b5a306618-config-data\") on node \"crc\" DevicePath \"\"" Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.009929 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c1bc0a9-e172-4749-b998-710b5a306618-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "3c1bc0a9-e172-4749-b998-710b5a306618" (UID: "3c1bc0a9-e172-4749-b998-710b5a306618"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.016183 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "test-operator-logs") pod "3c1bc0a9-e172-4749-b998-710b5a306618" (UID: "3c1bc0a9-e172-4749-b998-710b5a306618"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.018223 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c1bc0a9-e172-4749-b998-710b5a306618-kube-api-access-b2wxj" (OuterVolumeSpecName: "kube-api-access-b2wxj") pod "3c1bc0a9-e172-4749-b998-710b5a306618" (UID: "3c1bc0a9-e172-4749-b998-710b5a306618"). InnerVolumeSpecName "kube-api-access-b2wxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.039634 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c1bc0a9-e172-4749-b998-710b5a306618-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3c1bc0a9-e172-4749-b998-710b5a306618" (UID: "3c1bc0a9-e172-4749-b998-710b5a306618"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.044235 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c1bc0a9-e172-4749-b998-710b5a306618-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "3c1bc0a9-e172-4749-b998-710b5a306618" (UID: "3c1bc0a9-e172-4749-b998-710b5a306618"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.045915 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c1bc0a9-e172-4749-b998-710b5a306618-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "3c1bc0a9-e172-4749-b998-710b5a306618" (UID: "3c1bc0a9-e172-4749-b998-710b5a306618"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.063180 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c1bc0a9-e172-4749-b998-710b5a306618-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "3c1bc0a9-e172-4749-b998-710b5a306618" (UID: "3c1bc0a9-e172-4749-b998-710b5a306618"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.108523 4950 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3c1bc0a9-e172-4749-b998-710b5a306618-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.108577 4950 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3c1bc0a9-e172-4749-b998-710b5a306618-ca-certs\") on node \"crc\" DevicePath \"\"" Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.108591 4950 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c1bc0a9-e172-4749-b998-710b5a306618-ssh-key\") on node \"crc\" DevicePath \"\"" Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.110056 4950 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.110079 4950 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3c1bc0a9-e172-4749-b998-710b5a306618-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.110089 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2wxj\" (UniqueName: \"kubernetes.io/projected/3c1bc0a9-e172-4749-b998-710b5a306618-kube-api-access-b2wxj\") on node \"crc\" DevicePath \"\"" Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.110099 4950 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3c1bc0a9-e172-4749-b998-710b5a306618-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.110111 4950 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3c1bc0a9-e172-4749-b998-710b5a306618-openstack-config\") on node \"crc\" DevicePath \"\"" Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.136141 4950 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.211759 4950 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.461490 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"3c1bc0a9-e172-4749-b998-710b5a306618","Type":"ContainerDied","Data":"1a89b537bbaea1d20489be69908fd1cc6ee3aeb1f33ed9e0b129c450548abb7c"} Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.461528 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a89b537bbaea1d20489be69908fd1cc6ee3aeb1f33ed9e0b129c450548abb7c" Mar 18 21:14:43 crc kubenswrapper[4950]: I0318 21:14:43.461582 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 18 21:14:47 crc kubenswrapper[4950]: I0318 21:14:47.486817 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:14:47 crc kubenswrapper[4950]: E0318 21:14:47.490022 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:14:52 crc kubenswrapper[4950]: I0318 21:14:52.396321 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 18 21:14:52 crc kubenswrapper[4950]: E0318 21:14:52.397473 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c1bc0a9-e172-4749-b998-710b5a306618" containerName="tempest-tests-tempest-tests-runner" Mar 18 21:14:52 crc kubenswrapper[4950]: I0318 21:14:52.397488 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c1bc0a9-e172-4749-b998-710b5a306618" containerName="tempest-tests-tempest-tests-runner" Mar 18 21:14:52 crc kubenswrapper[4950]: E0318 21:14:52.397524 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d38b254-beb1-45fa-b20d-2b9c082c00b4" containerName="oc" Mar 18 21:14:52 crc kubenswrapper[4950]: I0318 21:14:52.397530 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d38b254-beb1-45fa-b20d-2b9c082c00b4" containerName="oc" Mar 18 21:14:52 crc kubenswrapper[4950]: I0318 21:14:52.397724 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c1bc0a9-e172-4749-b998-710b5a306618" containerName="tempest-tests-tempest-tests-runner" Mar 18 21:14:52 crc kubenswrapper[4950]: I0318 21:14:52.397742 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d38b254-beb1-45fa-b20d-2b9c082c00b4" containerName="oc" Mar 18 21:14:52 crc kubenswrapper[4950]: I0318 21:14:52.398389 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 18 21:14:52 crc kubenswrapper[4950]: I0318 21:14:52.409496 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 18 21:14:52 crc kubenswrapper[4950]: I0318 21:14:52.445329 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-z2mjq" Mar 18 21:14:52 crc kubenswrapper[4950]: I0318 21:14:52.508087 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c5edf1a8-44a7-4cd4-ae5f-f1a77338d35c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 18 21:14:52 crc kubenswrapper[4950]: I0318 21:14:52.508400 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgnpv\" (UniqueName: \"kubernetes.io/projected/c5edf1a8-44a7-4cd4-ae5f-f1a77338d35c-kube-api-access-xgnpv\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c5edf1a8-44a7-4cd4-ae5f-f1a77338d35c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 18 21:14:52 crc kubenswrapper[4950]: I0318 21:14:52.609939 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgnpv\" (UniqueName: \"kubernetes.io/projected/c5edf1a8-44a7-4cd4-ae5f-f1a77338d35c-kube-api-access-xgnpv\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c5edf1a8-44a7-4cd4-ae5f-f1a77338d35c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 18 21:14:52 crc kubenswrapper[4950]: I0318 21:14:52.610192 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c5edf1a8-44a7-4cd4-ae5f-f1a77338d35c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 18 21:14:52 crc kubenswrapper[4950]: I0318 21:14:52.613367 4950 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c5edf1a8-44a7-4cd4-ae5f-f1a77338d35c\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 18 21:14:52 crc kubenswrapper[4950]: I0318 21:14:52.630526 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgnpv\" (UniqueName: \"kubernetes.io/projected/c5edf1a8-44a7-4cd4-ae5f-f1a77338d35c-kube-api-access-xgnpv\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c5edf1a8-44a7-4cd4-ae5f-f1a77338d35c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 18 21:14:52 crc kubenswrapper[4950]: I0318 21:14:52.638267 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c5edf1a8-44a7-4cd4-ae5f-f1a77338d35c\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 18 21:14:52 crc kubenswrapper[4950]: I0318 21:14:52.770948 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 18 21:14:53 crc kubenswrapper[4950]: I0318 21:14:53.227829 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 18 21:14:53 crc kubenswrapper[4950]: I0318 21:14:53.555394 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"c5edf1a8-44a7-4cd4-ae5f-f1a77338d35c","Type":"ContainerStarted","Data":"cd225012dbf47a91f04616cf763c037325f067024191066d33921e54e2434f57"} Mar 18 21:14:54 crc kubenswrapper[4950]: I0318 21:14:54.565530 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"c5edf1a8-44a7-4cd4-ae5f-f1a77338d35c","Type":"ContainerStarted","Data":"94d1fe42c1a5f754b99acbefaeb55f68d5caddd28980adefa69d5c4217343342"} Mar 18 21:14:54 crc kubenswrapper[4950]: I0318 21:14:54.587153 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.808827129 podStartE2EDuration="2.587134755s" podCreationTimestamp="2026-03-18 21:14:52 +0000 UTC" firstStartedPulling="2026-03-18 21:14:53.234254296 +0000 UTC m=+4106.475096174" lastFinishedPulling="2026-03-18 21:14:54.012561942 +0000 UTC m=+4107.253403800" observedRunningTime="2026-03-18 21:14:54.580221194 +0000 UTC m=+4107.821063062" watchObservedRunningTime="2026-03-18 21:14:54.587134755 +0000 UTC m=+4107.827976623" Mar 18 21:15:00 crc kubenswrapper[4950]: I0318 21:15:00.181329 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564475-rtr5n"] Mar 18 21:15:00 crc kubenswrapper[4950]: I0318 21:15:00.185031 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564475-rtr5n" Mar 18 21:15:00 crc kubenswrapper[4950]: I0318 21:15:00.188322 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 18 21:15:00 crc kubenswrapper[4950]: I0318 21:15:00.193675 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564475-rtr5n"] Mar 18 21:15:00 crc kubenswrapper[4950]: I0318 21:15:00.199156 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 18 21:15:00 crc kubenswrapper[4950]: I0318 21:15:00.271682 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2br54\" (UniqueName: \"kubernetes.io/projected/8356f317-bbd8-487f-9637-a85dd9025385-kube-api-access-2br54\") pod \"collect-profiles-29564475-rtr5n\" (UID: \"8356f317-bbd8-487f-9637-a85dd9025385\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564475-rtr5n" Mar 18 21:15:00 crc kubenswrapper[4950]: I0318 21:15:00.271899 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8356f317-bbd8-487f-9637-a85dd9025385-config-volume\") pod \"collect-profiles-29564475-rtr5n\" (UID: \"8356f317-bbd8-487f-9637-a85dd9025385\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564475-rtr5n" Mar 18 21:15:00 crc kubenswrapper[4950]: I0318 21:15:00.271959 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8356f317-bbd8-487f-9637-a85dd9025385-secret-volume\") pod \"collect-profiles-29564475-rtr5n\" (UID: \"8356f317-bbd8-487f-9637-a85dd9025385\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564475-rtr5n" Mar 18 21:15:00 crc kubenswrapper[4950]: I0318 21:15:00.374585 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2br54\" (UniqueName: \"kubernetes.io/projected/8356f317-bbd8-487f-9637-a85dd9025385-kube-api-access-2br54\") pod \"collect-profiles-29564475-rtr5n\" (UID: \"8356f317-bbd8-487f-9637-a85dd9025385\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564475-rtr5n" Mar 18 21:15:00 crc kubenswrapper[4950]: I0318 21:15:00.374650 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8356f317-bbd8-487f-9637-a85dd9025385-config-volume\") pod \"collect-profiles-29564475-rtr5n\" (UID: \"8356f317-bbd8-487f-9637-a85dd9025385\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564475-rtr5n" Mar 18 21:15:00 crc kubenswrapper[4950]: I0318 21:15:00.374678 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8356f317-bbd8-487f-9637-a85dd9025385-secret-volume\") pod \"collect-profiles-29564475-rtr5n\" (UID: \"8356f317-bbd8-487f-9637-a85dd9025385\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564475-rtr5n" Mar 18 21:15:00 crc kubenswrapper[4950]: I0318 21:15:00.376570 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8356f317-bbd8-487f-9637-a85dd9025385-config-volume\") pod \"collect-profiles-29564475-rtr5n\" (UID: \"8356f317-bbd8-487f-9637-a85dd9025385\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564475-rtr5n" Mar 18 21:15:00 crc kubenswrapper[4950]: I0318 21:15:00.388564 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8356f317-bbd8-487f-9637-a85dd9025385-secret-volume\") pod \"collect-profiles-29564475-rtr5n\" (UID: \"8356f317-bbd8-487f-9637-a85dd9025385\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564475-rtr5n" Mar 18 21:15:00 crc kubenswrapper[4950]: I0318 21:15:00.390476 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2br54\" (UniqueName: \"kubernetes.io/projected/8356f317-bbd8-487f-9637-a85dd9025385-kube-api-access-2br54\") pod \"collect-profiles-29564475-rtr5n\" (UID: \"8356f317-bbd8-487f-9637-a85dd9025385\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564475-rtr5n" Mar 18 21:15:00 crc kubenswrapper[4950]: I0318 21:15:00.525229 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564475-rtr5n" Mar 18 21:15:00 crc kubenswrapper[4950]: W0318 21:15:00.979766 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8356f317_bbd8_487f_9637_a85dd9025385.slice/crio-e06f335e68e18d97a0a63f42da5345231969afae4ec60dc293b50c430b73d94f WatchSource:0}: Error finding container e06f335e68e18d97a0a63f42da5345231969afae4ec60dc293b50c430b73d94f: Status 404 returned error can't find the container with id e06f335e68e18d97a0a63f42da5345231969afae4ec60dc293b50c430b73d94f Mar 18 21:15:00 crc kubenswrapper[4950]: I0318 21:15:00.982544 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564475-rtr5n"] Mar 18 21:15:01 crc kubenswrapper[4950]: I0318 21:15:01.229502 4950 scope.go:117] "RemoveContainer" containerID="e8fd6fd512f2ef0e49276b7bdf8a42dfbb9e0e1236732ace73cedd484d6bfa3e" Mar 18 21:15:01 crc kubenswrapper[4950]: I0318 21:15:01.634088 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564475-rtr5n" event={"ID":"8356f317-bbd8-487f-9637-a85dd9025385","Type":"ContainerStarted","Data":"31d087fb3631f91a259b882041df94adf17a93db978e22e89614d78c6a0a5e81"} Mar 18 21:15:01 crc kubenswrapper[4950]: I0318 21:15:01.634514 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564475-rtr5n" event={"ID":"8356f317-bbd8-487f-9637-a85dd9025385","Type":"ContainerStarted","Data":"e06f335e68e18d97a0a63f42da5345231969afae4ec60dc293b50c430b73d94f"} Mar 18 21:15:02 crc kubenswrapper[4950]: I0318 21:15:02.479790 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:15:02 crc kubenswrapper[4950]: E0318 21:15:02.480343 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:15:02 crc kubenswrapper[4950]: I0318 21:15:02.649512 4950 generic.go:334] "Generic (PLEG): container finished" podID="8356f317-bbd8-487f-9637-a85dd9025385" containerID="31d087fb3631f91a259b882041df94adf17a93db978e22e89614d78c6a0a5e81" exitCode=0 Mar 18 21:15:02 crc kubenswrapper[4950]: I0318 21:15:02.649559 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564475-rtr5n" event={"ID":"8356f317-bbd8-487f-9637-a85dd9025385","Type":"ContainerDied","Data":"31d087fb3631f91a259b882041df94adf17a93db978e22e89614d78c6a0a5e81"} Mar 18 21:15:04 crc kubenswrapper[4950]: I0318 21:15:04.045179 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564475-rtr5n" Mar 18 21:15:04 crc kubenswrapper[4950]: I0318 21:15:04.158521 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8356f317-bbd8-487f-9637-a85dd9025385-secret-volume\") pod \"8356f317-bbd8-487f-9637-a85dd9025385\" (UID: \"8356f317-bbd8-487f-9637-a85dd9025385\") " Mar 18 21:15:04 crc kubenswrapper[4950]: I0318 21:15:04.158648 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8356f317-bbd8-487f-9637-a85dd9025385-config-volume\") pod \"8356f317-bbd8-487f-9637-a85dd9025385\" (UID: \"8356f317-bbd8-487f-9637-a85dd9025385\") " Mar 18 21:15:04 crc kubenswrapper[4950]: I0318 21:15:04.158951 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2br54\" (UniqueName: \"kubernetes.io/projected/8356f317-bbd8-487f-9637-a85dd9025385-kube-api-access-2br54\") pod \"8356f317-bbd8-487f-9637-a85dd9025385\" (UID: \"8356f317-bbd8-487f-9637-a85dd9025385\") " Mar 18 21:15:04 crc kubenswrapper[4950]: I0318 21:15:04.159318 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8356f317-bbd8-487f-9637-a85dd9025385-config-volume" (OuterVolumeSpecName: "config-volume") pod "8356f317-bbd8-487f-9637-a85dd9025385" (UID: "8356f317-bbd8-487f-9637-a85dd9025385"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 21:15:04 crc kubenswrapper[4950]: I0318 21:15:04.160489 4950 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8356f317-bbd8-487f-9637-a85dd9025385-config-volume\") on node \"crc\" DevicePath \"\"" Mar 18 21:15:04 crc kubenswrapper[4950]: I0318 21:15:04.164231 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8356f317-bbd8-487f-9637-a85dd9025385-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8356f317-bbd8-487f-9637-a85dd9025385" (UID: "8356f317-bbd8-487f-9637-a85dd9025385"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:15:04 crc kubenswrapper[4950]: I0318 21:15:04.167065 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8356f317-bbd8-487f-9637-a85dd9025385-kube-api-access-2br54" (OuterVolumeSpecName: "kube-api-access-2br54") pod "8356f317-bbd8-487f-9637-a85dd9025385" (UID: "8356f317-bbd8-487f-9637-a85dd9025385"). InnerVolumeSpecName "kube-api-access-2br54". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:15:04 crc kubenswrapper[4950]: I0318 21:15:04.261737 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2br54\" (UniqueName: \"kubernetes.io/projected/8356f317-bbd8-487f-9637-a85dd9025385-kube-api-access-2br54\") on node \"crc\" DevicePath \"\"" Mar 18 21:15:04 crc kubenswrapper[4950]: I0318 21:15:04.261767 4950 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8356f317-bbd8-487f-9637-a85dd9025385-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 18 21:15:04 crc kubenswrapper[4950]: I0318 21:15:04.669855 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564475-rtr5n" event={"ID":"8356f317-bbd8-487f-9637-a85dd9025385","Type":"ContainerDied","Data":"e06f335e68e18d97a0a63f42da5345231969afae4ec60dc293b50c430b73d94f"} Mar 18 21:15:04 crc kubenswrapper[4950]: I0318 21:15:04.669907 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564475-rtr5n" Mar 18 21:15:04 crc kubenswrapper[4950]: I0318 21:15:04.669913 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e06f335e68e18d97a0a63f42da5345231969afae4ec60dc293b50c430b73d94f" Mar 18 21:15:04 crc kubenswrapper[4950]: I0318 21:15:04.747036 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564430-hrm5q"] Mar 18 21:15:04 crc kubenswrapper[4950]: I0318 21:15:04.760625 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564430-hrm5q"] Mar 18 21:15:04 crc kubenswrapper[4950]: E0318 21:15:04.881468 4950 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8356f317_bbd8_487f_9637_a85dd9025385.slice\": RecentStats: unable to find data in memory cache]" Mar 18 21:15:05 crc kubenswrapper[4950]: I0318 21:15:05.514843 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87931a42-2463-4633-a253-f7a01eee1935" path="/var/lib/kubelet/pods/87931a42-2463-4633-a253-f7a01eee1935/volumes" Mar 18 21:15:13 crc kubenswrapper[4950]: I0318 21:15:13.480316 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:15:13 crc kubenswrapper[4950]: E0318 21:15:13.481072 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:15:21 crc kubenswrapper[4950]: I0318 21:15:21.220972 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hjm8k/must-gather-p5s84"] Mar 18 21:15:21 crc kubenswrapper[4950]: E0318 21:15:21.222706 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8356f317-bbd8-487f-9637-a85dd9025385" containerName="collect-profiles" Mar 18 21:15:21 crc kubenswrapper[4950]: I0318 21:15:21.222793 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="8356f317-bbd8-487f-9637-a85dd9025385" containerName="collect-profiles" Mar 18 21:15:21 crc kubenswrapper[4950]: I0318 21:15:21.223059 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="8356f317-bbd8-487f-9637-a85dd9025385" containerName="collect-profiles" Mar 18 21:15:21 crc kubenswrapper[4950]: I0318 21:15:21.224271 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hjm8k/must-gather-p5s84" Mar 18 21:15:21 crc kubenswrapper[4950]: I0318 21:15:21.226853 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-hjm8k"/"openshift-service-ca.crt" Mar 18 21:15:21 crc kubenswrapper[4950]: I0318 21:15:21.227323 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-hjm8k"/"kube-root-ca.crt" Mar 18 21:15:21 crc kubenswrapper[4950]: I0318 21:15:21.227781 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-hjm8k"/"default-dockercfg-fmw2p" Mar 18 21:15:21 crc kubenswrapper[4950]: I0318 21:15:21.242273 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-hjm8k/must-gather-p5s84"] Mar 18 21:15:21 crc kubenswrapper[4950]: I0318 21:15:21.373864 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ssbs\" (UniqueName: \"kubernetes.io/projected/cdc951ae-b4eb-46de-beac-53c6aa48073a-kube-api-access-5ssbs\") pod \"must-gather-p5s84\" (UID: \"cdc951ae-b4eb-46de-beac-53c6aa48073a\") " pod="openshift-must-gather-hjm8k/must-gather-p5s84" Mar 18 21:15:21 crc kubenswrapper[4950]: I0318 21:15:21.374335 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cdc951ae-b4eb-46de-beac-53c6aa48073a-must-gather-output\") pod \"must-gather-p5s84\" (UID: \"cdc951ae-b4eb-46de-beac-53c6aa48073a\") " pod="openshift-must-gather-hjm8k/must-gather-p5s84" Mar 18 21:15:21 crc kubenswrapper[4950]: I0318 21:15:21.475901 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cdc951ae-b4eb-46de-beac-53c6aa48073a-must-gather-output\") pod \"must-gather-p5s84\" (UID: \"cdc951ae-b4eb-46de-beac-53c6aa48073a\") " pod="openshift-must-gather-hjm8k/must-gather-p5s84" Mar 18 21:15:21 crc kubenswrapper[4950]: I0318 21:15:21.475984 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ssbs\" (UniqueName: \"kubernetes.io/projected/cdc951ae-b4eb-46de-beac-53c6aa48073a-kube-api-access-5ssbs\") pod \"must-gather-p5s84\" (UID: \"cdc951ae-b4eb-46de-beac-53c6aa48073a\") " pod="openshift-must-gather-hjm8k/must-gather-p5s84" Mar 18 21:15:21 crc kubenswrapper[4950]: I0318 21:15:21.476333 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cdc951ae-b4eb-46de-beac-53c6aa48073a-must-gather-output\") pod \"must-gather-p5s84\" (UID: \"cdc951ae-b4eb-46de-beac-53c6aa48073a\") " pod="openshift-must-gather-hjm8k/must-gather-p5s84" Mar 18 21:15:21 crc kubenswrapper[4950]: I0318 21:15:21.498128 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ssbs\" (UniqueName: \"kubernetes.io/projected/cdc951ae-b4eb-46de-beac-53c6aa48073a-kube-api-access-5ssbs\") pod \"must-gather-p5s84\" (UID: \"cdc951ae-b4eb-46de-beac-53c6aa48073a\") " pod="openshift-must-gather-hjm8k/must-gather-p5s84" Mar 18 21:15:21 crc kubenswrapper[4950]: I0318 21:15:21.540315 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hjm8k/must-gather-p5s84" Mar 18 21:15:21 crc kubenswrapper[4950]: W0318 21:15:21.989738 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdc951ae_b4eb_46de_beac_53c6aa48073a.slice/crio-a52b6836c05cf704344a4c5d233018a3755f186d6608cc61b93e6801fae5ef5f WatchSource:0}: Error finding container a52b6836c05cf704344a4c5d233018a3755f186d6608cc61b93e6801fae5ef5f: Status 404 returned error can't find the container with id a52b6836c05cf704344a4c5d233018a3755f186d6608cc61b93e6801fae5ef5f Mar 18 21:15:21 crc kubenswrapper[4950]: I0318 21:15:21.990089 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-hjm8k/must-gather-p5s84"] Mar 18 21:15:22 crc kubenswrapper[4950]: I0318 21:15:22.854302 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hjm8k/must-gather-p5s84" event={"ID":"cdc951ae-b4eb-46de-beac-53c6aa48073a","Type":"ContainerStarted","Data":"a52b6836c05cf704344a4c5d233018a3755f186d6608cc61b93e6801fae5ef5f"} Mar 18 21:15:24 crc kubenswrapper[4950]: I0318 21:15:24.480377 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:15:24 crc kubenswrapper[4950]: E0318 21:15:24.480913 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:15:30 crc kubenswrapper[4950]: I0318 21:15:30.937892 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hjm8k/must-gather-p5s84" event={"ID":"cdc951ae-b4eb-46de-beac-53c6aa48073a","Type":"ContainerStarted","Data":"6c6455eb3d0e44c0ade010eb8266c6e050ae06b477d64451e0ba8a239fd1eddb"} Mar 18 21:15:30 crc kubenswrapper[4950]: I0318 21:15:30.938556 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hjm8k/must-gather-p5s84" event={"ID":"cdc951ae-b4eb-46de-beac-53c6aa48073a","Type":"ContainerStarted","Data":"74520bc0cd6b5ec6357b4d557ad3193b248a8beafc461274c65c9cd5b6eb3571"} Mar 18 21:15:30 crc kubenswrapper[4950]: I0318 21:15:30.963935 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-hjm8k/must-gather-p5s84" podStartSLOduration=2.267235686 podStartE2EDuration="9.963912s" podCreationTimestamp="2026-03-18 21:15:21 +0000 UTC" firstStartedPulling="2026-03-18 21:15:21.992518146 +0000 UTC m=+4135.233360014" lastFinishedPulling="2026-03-18 21:15:29.68919445 +0000 UTC m=+4142.930036328" observedRunningTime="2026-03-18 21:15:30.953477646 +0000 UTC m=+4144.194319534" watchObservedRunningTime="2026-03-18 21:15:30.963912 +0000 UTC m=+4144.204753868" Mar 18 21:15:36 crc kubenswrapper[4950]: I0318 21:15:36.595951 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hjm8k/crc-debug-hzct9"] Mar 18 21:15:36 crc kubenswrapper[4950]: I0318 21:15:36.597677 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hjm8k/crc-debug-hzct9" Mar 18 21:15:36 crc kubenswrapper[4950]: I0318 21:15:36.609443 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdltw\" (UniqueName: \"kubernetes.io/projected/ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9-kube-api-access-hdltw\") pod \"crc-debug-hzct9\" (UID: \"ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9\") " pod="openshift-must-gather-hjm8k/crc-debug-hzct9" Mar 18 21:15:36 crc kubenswrapper[4950]: I0318 21:15:36.609511 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9-host\") pod \"crc-debug-hzct9\" (UID: \"ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9\") " pod="openshift-must-gather-hjm8k/crc-debug-hzct9" Mar 18 21:15:36 crc kubenswrapper[4950]: I0318 21:15:36.710946 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdltw\" (UniqueName: \"kubernetes.io/projected/ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9-kube-api-access-hdltw\") pod \"crc-debug-hzct9\" (UID: \"ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9\") " pod="openshift-must-gather-hjm8k/crc-debug-hzct9" Mar 18 21:15:36 crc kubenswrapper[4950]: I0318 21:15:36.711020 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9-host\") pod \"crc-debug-hzct9\" (UID: \"ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9\") " pod="openshift-must-gather-hjm8k/crc-debug-hzct9" Mar 18 21:15:36 crc kubenswrapper[4950]: I0318 21:15:36.711147 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9-host\") pod \"crc-debug-hzct9\" (UID: \"ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9\") " pod="openshift-must-gather-hjm8k/crc-debug-hzct9" Mar 18 21:15:36 crc kubenswrapper[4950]: I0318 21:15:36.740833 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdltw\" (UniqueName: \"kubernetes.io/projected/ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9-kube-api-access-hdltw\") pod \"crc-debug-hzct9\" (UID: \"ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9\") " pod="openshift-must-gather-hjm8k/crc-debug-hzct9" Mar 18 21:15:36 crc kubenswrapper[4950]: I0318 21:15:36.918706 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hjm8k/crc-debug-hzct9" Mar 18 21:15:36 crc kubenswrapper[4950]: W0318 21:15:36.963903 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecb0ce74_f47c_4c2e_aa8a_4ad77e7a07e9.slice/crio-011734844b96dd56dd028970b4f0a5fe74280cfa6262563ea53f2895a17a724c WatchSource:0}: Error finding container 011734844b96dd56dd028970b4f0a5fe74280cfa6262563ea53f2895a17a724c: Status 404 returned error can't find the container with id 011734844b96dd56dd028970b4f0a5fe74280cfa6262563ea53f2895a17a724c Mar 18 21:15:36 crc kubenswrapper[4950]: I0318 21:15:36.992066 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hjm8k/crc-debug-hzct9" event={"ID":"ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9","Type":"ContainerStarted","Data":"011734844b96dd56dd028970b4f0a5fe74280cfa6262563ea53f2895a17a724c"} Mar 18 21:15:38 crc kubenswrapper[4950]: I0318 21:15:38.480201 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:15:38 crc kubenswrapper[4950]: E0318 21:15:38.480783 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:15:50 crc kubenswrapper[4950]: I0318 21:15:50.479809 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:15:50 crc kubenswrapper[4950]: E0318 21:15:50.481043 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:15:51 crc kubenswrapper[4950]: I0318 21:15:51.140995 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hjm8k/crc-debug-hzct9" event={"ID":"ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9","Type":"ContainerStarted","Data":"c673e8b6123d7b1808035305c8600c3f0c7fa231a1fffcedbc2b9e1d0abe21af"} Mar 18 21:15:51 crc kubenswrapper[4950]: I0318 21:15:51.162539 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-hjm8k/crc-debug-hzct9" podStartSLOduration=1.7646128700000001 podStartE2EDuration="15.162519566s" podCreationTimestamp="2026-03-18 21:15:36 +0000 UTC" firstStartedPulling="2026-03-18 21:15:36.96596145 +0000 UTC m=+4150.206803318" lastFinishedPulling="2026-03-18 21:15:50.363868146 +0000 UTC m=+4163.604710014" observedRunningTime="2026-03-18 21:15:51.157495094 +0000 UTC m=+4164.398336962" watchObservedRunningTime="2026-03-18 21:15:51.162519566 +0000 UTC m=+4164.403361434" Mar 18 21:16:00 crc kubenswrapper[4950]: I0318 21:16:00.160232 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564476-jkc8h"] Mar 18 21:16:00 crc kubenswrapper[4950]: I0318 21:16:00.162195 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564476-jkc8h" Mar 18 21:16:00 crc kubenswrapper[4950]: I0318 21:16:00.167672 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 21:16:00 crc kubenswrapper[4950]: I0318 21:16:00.167848 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 21:16:00 crc kubenswrapper[4950]: I0318 21:16:00.170781 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 21:16:00 crc kubenswrapper[4950]: I0318 21:16:00.178158 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564476-jkc8h"] Mar 18 21:16:00 crc kubenswrapper[4950]: I0318 21:16:00.316841 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6x7z\" (UniqueName: \"kubernetes.io/projected/abf3efdf-f206-4990-bc5f-6a9b20d28438-kube-api-access-k6x7z\") pod \"auto-csr-approver-29564476-jkc8h\" (UID: \"abf3efdf-f206-4990-bc5f-6a9b20d28438\") " pod="openshift-infra/auto-csr-approver-29564476-jkc8h" Mar 18 21:16:00 crc kubenswrapper[4950]: I0318 21:16:00.418882 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6x7z\" (UniqueName: \"kubernetes.io/projected/abf3efdf-f206-4990-bc5f-6a9b20d28438-kube-api-access-k6x7z\") pod \"auto-csr-approver-29564476-jkc8h\" (UID: \"abf3efdf-f206-4990-bc5f-6a9b20d28438\") " pod="openshift-infra/auto-csr-approver-29564476-jkc8h" Mar 18 21:16:00 crc kubenswrapper[4950]: I0318 21:16:00.447440 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6x7z\" (UniqueName: \"kubernetes.io/projected/abf3efdf-f206-4990-bc5f-6a9b20d28438-kube-api-access-k6x7z\") pod \"auto-csr-approver-29564476-jkc8h\" (UID: \"abf3efdf-f206-4990-bc5f-6a9b20d28438\") " pod="openshift-infra/auto-csr-approver-29564476-jkc8h" Mar 18 21:16:00 crc kubenswrapper[4950]: I0318 21:16:00.490339 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564476-jkc8h" Mar 18 21:16:00 crc kubenswrapper[4950]: I0318 21:16:00.969972 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564476-jkc8h"] Mar 18 21:16:01 crc kubenswrapper[4950]: I0318 21:16:01.250453 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564476-jkc8h" event={"ID":"abf3efdf-f206-4990-bc5f-6a9b20d28438","Type":"ContainerStarted","Data":"4eb0d365c1b51821aefeecb69e2b75404d510bdf8f003eb710891b30d8f4bab6"} Mar 18 21:16:01 crc kubenswrapper[4950]: I0318 21:16:01.323182 4950 scope.go:117] "RemoveContainer" containerID="18d06b3d74085c796f46cf0af33ebf90c012776cc061dba2a1c1175506c6e48e" Mar 18 21:16:03 crc kubenswrapper[4950]: I0318 21:16:03.272701 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564476-jkc8h" event={"ID":"abf3efdf-f206-4990-bc5f-6a9b20d28438","Type":"ContainerStarted","Data":"109a167196d7db7588ce49778deb57663081ce9e34ed55fc93d330d2da292a5b"} Mar 18 21:16:03 crc kubenswrapper[4950]: I0318 21:16:03.292034 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29564476-jkc8h" podStartSLOduration=2.338436572 podStartE2EDuration="3.292016076s" podCreationTimestamp="2026-03-18 21:16:00 +0000 UTC" firstStartedPulling="2026-03-18 21:16:00.980266343 +0000 UTC m=+4174.221108211" lastFinishedPulling="2026-03-18 21:16:01.933845847 +0000 UTC m=+4175.174687715" observedRunningTime="2026-03-18 21:16:03.28834917 +0000 UTC m=+4176.529191038" watchObservedRunningTime="2026-03-18 21:16:03.292016076 +0000 UTC m=+4176.532857944" Mar 18 21:16:04 crc kubenswrapper[4950]: I0318 21:16:04.285808 4950 generic.go:334] "Generic (PLEG): container finished" podID="abf3efdf-f206-4990-bc5f-6a9b20d28438" containerID="109a167196d7db7588ce49778deb57663081ce9e34ed55fc93d330d2da292a5b" exitCode=0 Mar 18 21:16:04 crc kubenswrapper[4950]: I0318 21:16:04.285910 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564476-jkc8h" event={"ID":"abf3efdf-f206-4990-bc5f-6a9b20d28438","Type":"ContainerDied","Data":"109a167196d7db7588ce49778deb57663081ce9e34ed55fc93d330d2da292a5b"} Mar 18 21:16:04 crc kubenswrapper[4950]: I0318 21:16:04.479996 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:16:04 crc kubenswrapper[4950]: E0318 21:16:04.480588 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:16:06 crc kubenswrapper[4950]: I0318 21:16:05.750463 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564476-jkc8h" Mar 18 21:16:06 crc kubenswrapper[4950]: I0318 21:16:05.800897 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6x7z\" (UniqueName: \"kubernetes.io/projected/abf3efdf-f206-4990-bc5f-6a9b20d28438-kube-api-access-k6x7z\") pod \"abf3efdf-f206-4990-bc5f-6a9b20d28438\" (UID: \"abf3efdf-f206-4990-bc5f-6a9b20d28438\") " Mar 18 21:16:06 crc kubenswrapper[4950]: I0318 21:16:05.817709 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abf3efdf-f206-4990-bc5f-6a9b20d28438-kube-api-access-k6x7z" (OuterVolumeSpecName: "kube-api-access-k6x7z") pod "abf3efdf-f206-4990-bc5f-6a9b20d28438" (UID: "abf3efdf-f206-4990-bc5f-6a9b20d28438"). InnerVolumeSpecName "kube-api-access-k6x7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:16:06 crc kubenswrapper[4950]: I0318 21:16:05.903301 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6x7z\" (UniqueName: \"kubernetes.io/projected/abf3efdf-f206-4990-bc5f-6a9b20d28438-kube-api-access-k6x7z\") on node \"crc\" DevicePath \"\"" Mar 18 21:16:06 crc kubenswrapper[4950]: I0318 21:16:06.302795 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564476-jkc8h" event={"ID":"abf3efdf-f206-4990-bc5f-6a9b20d28438","Type":"ContainerDied","Data":"4eb0d365c1b51821aefeecb69e2b75404d510bdf8f003eb710891b30d8f4bab6"} Mar 18 21:16:06 crc kubenswrapper[4950]: I0318 21:16:06.302852 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4eb0d365c1b51821aefeecb69e2b75404d510bdf8f003eb710891b30d8f4bab6" Mar 18 21:16:06 crc kubenswrapper[4950]: I0318 21:16:06.302918 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564476-jkc8h" Mar 18 21:16:06 crc kubenswrapper[4950]: I0318 21:16:06.366077 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564470-mhw8z"] Mar 18 21:16:06 crc kubenswrapper[4950]: I0318 21:16:06.376156 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564470-mhw8z"] Mar 18 21:16:06 crc kubenswrapper[4950]: E0318 21:16:06.507799 4950 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podabf3efdf_f206_4990_bc5f_6a9b20d28438.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podabf3efdf_f206_4990_bc5f_6a9b20d28438.slice/crio-4eb0d365c1b51821aefeecb69e2b75404d510bdf8f003eb710891b30d8f4bab6\": RecentStats: unable to find data in memory cache]" Mar 18 21:16:07 crc kubenswrapper[4950]: I0318 21:16:07.493139 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7142cc1f-1577-45b7-bc79-a4dd4416455e" path="/var/lib/kubelet/pods/7142cc1f-1577-45b7-bc79-a4dd4416455e/volumes" Mar 18 21:16:18 crc kubenswrapper[4950]: I0318 21:16:18.578075 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b79zp"] Mar 18 21:16:18 crc kubenswrapper[4950]: E0318 21:16:18.579088 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abf3efdf-f206-4990-bc5f-6a9b20d28438" containerName="oc" Mar 18 21:16:18 crc kubenswrapper[4950]: I0318 21:16:18.579104 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="abf3efdf-f206-4990-bc5f-6a9b20d28438" containerName="oc" Mar 18 21:16:18 crc kubenswrapper[4950]: I0318 21:16:18.579350 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="abf3efdf-f206-4990-bc5f-6a9b20d28438" containerName="oc" Mar 18 21:16:18 crc kubenswrapper[4950]: I0318 21:16:18.582079 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b79zp" Mar 18 21:16:18 crc kubenswrapper[4950]: I0318 21:16:18.591767 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b79zp"] Mar 18 21:16:18 crc kubenswrapper[4950]: I0318 21:16:18.648655 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1a59a14-c501-4229-9508-aed7307123cf-catalog-content\") pod \"redhat-marketplace-b79zp\" (UID: \"d1a59a14-c501-4229-9508-aed7307123cf\") " pod="openshift-marketplace/redhat-marketplace-b79zp" Mar 18 21:16:18 crc kubenswrapper[4950]: I0318 21:16:18.648713 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1a59a14-c501-4229-9508-aed7307123cf-utilities\") pod \"redhat-marketplace-b79zp\" (UID: \"d1a59a14-c501-4229-9508-aed7307123cf\") " pod="openshift-marketplace/redhat-marketplace-b79zp" Mar 18 21:16:18 crc kubenswrapper[4950]: I0318 21:16:18.648773 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4ld9\" (UniqueName: \"kubernetes.io/projected/d1a59a14-c501-4229-9508-aed7307123cf-kube-api-access-p4ld9\") pod \"redhat-marketplace-b79zp\" (UID: \"d1a59a14-c501-4229-9508-aed7307123cf\") " pod="openshift-marketplace/redhat-marketplace-b79zp" Mar 18 21:16:18 crc kubenswrapper[4950]: I0318 21:16:18.750865 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1a59a14-c501-4229-9508-aed7307123cf-catalog-content\") pod \"redhat-marketplace-b79zp\" (UID: \"d1a59a14-c501-4229-9508-aed7307123cf\") " pod="openshift-marketplace/redhat-marketplace-b79zp" Mar 18 21:16:18 crc kubenswrapper[4950]: I0318 21:16:18.751370 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1a59a14-c501-4229-9508-aed7307123cf-catalog-content\") pod \"redhat-marketplace-b79zp\" (UID: \"d1a59a14-c501-4229-9508-aed7307123cf\") " pod="openshift-marketplace/redhat-marketplace-b79zp" Mar 18 21:16:18 crc kubenswrapper[4950]: I0318 21:16:18.751474 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1a59a14-c501-4229-9508-aed7307123cf-utilities\") pod \"redhat-marketplace-b79zp\" (UID: \"d1a59a14-c501-4229-9508-aed7307123cf\") " pod="openshift-marketplace/redhat-marketplace-b79zp" Mar 18 21:16:18 crc kubenswrapper[4950]: I0318 21:16:18.751721 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1a59a14-c501-4229-9508-aed7307123cf-utilities\") pod \"redhat-marketplace-b79zp\" (UID: \"d1a59a14-c501-4229-9508-aed7307123cf\") " pod="openshift-marketplace/redhat-marketplace-b79zp" Mar 18 21:16:18 crc kubenswrapper[4950]: I0318 21:16:18.752109 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4ld9\" (UniqueName: \"kubernetes.io/projected/d1a59a14-c501-4229-9508-aed7307123cf-kube-api-access-p4ld9\") pod \"redhat-marketplace-b79zp\" (UID: \"d1a59a14-c501-4229-9508-aed7307123cf\") " pod="openshift-marketplace/redhat-marketplace-b79zp" Mar 18 21:16:18 crc kubenswrapper[4950]: I0318 21:16:18.772738 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4ld9\" (UniqueName: \"kubernetes.io/projected/d1a59a14-c501-4229-9508-aed7307123cf-kube-api-access-p4ld9\") pod \"redhat-marketplace-b79zp\" (UID: \"d1a59a14-c501-4229-9508-aed7307123cf\") " pod="openshift-marketplace/redhat-marketplace-b79zp" Mar 18 21:16:18 crc kubenswrapper[4950]: I0318 21:16:18.902347 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b79zp" Mar 18 21:16:19 crc kubenswrapper[4950]: I0318 21:16:19.479868 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:16:19 crc kubenswrapper[4950]: E0318 21:16:19.480666 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:16:19 crc kubenswrapper[4950]: I0318 21:16:19.509165 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b79zp"] Mar 18 21:16:20 crc kubenswrapper[4950]: I0318 21:16:20.428150 4950 generic.go:334] "Generic (PLEG): container finished" podID="d1a59a14-c501-4229-9508-aed7307123cf" containerID="5b649012fbfaba94b3f85a42a8c26be202beee08046dc39bc4972fba10fc6155" exitCode=0 Mar 18 21:16:20 crc kubenswrapper[4950]: I0318 21:16:20.428260 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b79zp" event={"ID":"d1a59a14-c501-4229-9508-aed7307123cf","Type":"ContainerDied","Data":"5b649012fbfaba94b3f85a42a8c26be202beee08046dc39bc4972fba10fc6155"} Mar 18 21:16:20 crc kubenswrapper[4950]: I0318 21:16:20.428465 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b79zp" event={"ID":"d1a59a14-c501-4229-9508-aed7307123cf","Type":"ContainerStarted","Data":"3fe8ddccbedb412ca0bf72384296c27b19d21997bfaed94de927b2249720f990"} Mar 18 21:16:21 crc kubenswrapper[4950]: I0318 21:16:21.439901 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b79zp" event={"ID":"d1a59a14-c501-4229-9508-aed7307123cf","Type":"ContainerStarted","Data":"45cac7594668627aec1c56fd08cf600e7beed90ff4058c8431b3e86da2106fb3"} Mar 18 21:16:23 crc kubenswrapper[4950]: I0318 21:16:23.457461 4950 generic.go:334] "Generic (PLEG): container finished" podID="d1a59a14-c501-4229-9508-aed7307123cf" containerID="45cac7594668627aec1c56fd08cf600e7beed90ff4058c8431b3e86da2106fb3" exitCode=0 Mar 18 21:16:23 crc kubenswrapper[4950]: I0318 21:16:23.457584 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b79zp" event={"ID":"d1a59a14-c501-4229-9508-aed7307123cf","Type":"ContainerDied","Data":"45cac7594668627aec1c56fd08cf600e7beed90ff4058c8431b3e86da2106fb3"} Mar 18 21:16:24 crc kubenswrapper[4950]: I0318 21:16:24.467491 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b79zp" event={"ID":"d1a59a14-c501-4229-9508-aed7307123cf","Type":"ContainerStarted","Data":"edf49bcb079f2cd2e01fb04c40776aa03c4d4789bb5fee009462e4a94f357053"} Mar 18 21:16:24 crc kubenswrapper[4950]: I0318 21:16:24.484644 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b79zp" podStartSLOduration=2.921505397 podStartE2EDuration="6.484629098s" podCreationTimestamp="2026-03-18 21:16:18 +0000 UTC" firstStartedPulling="2026-03-18 21:16:20.430088896 +0000 UTC m=+4193.670930764" lastFinishedPulling="2026-03-18 21:16:23.993212597 +0000 UTC m=+4197.234054465" observedRunningTime="2026-03-18 21:16:24.481820414 +0000 UTC m=+4197.722662282" watchObservedRunningTime="2026-03-18 21:16:24.484629098 +0000 UTC m=+4197.725470956" Mar 18 21:16:28 crc kubenswrapper[4950]: I0318 21:16:28.904004 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b79zp" Mar 18 21:16:28 crc kubenswrapper[4950]: I0318 21:16:28.904961 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b79zp" Mar 18 21:16:29 crc kubenswrapper[4950]: I0318 21:16:29.963787 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-b79zp" podUID="d1a59a14-c501-4229-9508-aed7307123cf" containerName="registry-server" probeResult="failure" output=< Mar 18 21:16:29 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 21:16:29 crc kubenswrapper[4950]: > Mar 18 21:16:32 crc kubenswrapper[4950]: I0318 21:16:32.479752 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:16:32 crc kubenswrapper[4950]: E0318 21:16:32.480632 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:16:33 crc kubenswrapper[4950]: I0318 21:16:33.537252 4950 generic.go:334] "Generic (PLEG): container finished" podID="ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9" containerID="c673e8b6123d7b1808035305c8600c3f0c7fa231a1fffcedbc2b9e1d0abe21af" exitCode=0 Mar 18 21:16:33 crc kubenswrapper[4950]: I0318 21:16:33.537538 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hjm8k/crc-debug-hzct9" event={"ID":"ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9","Type":"ContainerDied","Data":"c673e8b6123d7b1808035305c8600c3f0c7fa231a1fffcedbc2b9e1d0abe21af"} Mar 18 21:16:34 crc kubenswrapper[4950]: I0318 21:16:34.712052 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hjm8k/crc-debug-hzct9" Mar 18 21:16:34 crc kubenswrapper[4950]: I0318 21:16:34.717326 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdltw\" (UniqueName: \"kubernetes.io/projected/ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9-kube-api-access-hdltw\") pod \"ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9\" (UID: \"ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9\") " Mar 18 21:16:34 crc kubenswrapper[4950]: I0318 21:16:34.717390 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9-host\") pod \"ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9\" (UID: \"ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9\") " Mar 18 21:16:34 crc kubenswrapper[4950]: I0318 21:16:34.718294 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9-host" (OuterVolumeSpecName: "host") pod "ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9" (UID: "ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 21:16:34 crc kubenswrapper[4950]: I0318 21:16:34.726118 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9-kube-api-access-hdltw" (OuterVolumeSpecName: "kube-api-access-hdltw") pod "ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9" (UID: "ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9"). InnerVolumeSpecName "kube-api-access-hdltw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:16:34 crc kubenswrapper[4950]: I0318 21:16:34.763606 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hjm8k/crc-debug-hzct9"] Mar 18 21:16:34 crc kubenswrapper[4950]: I0318 21:16:34.776020 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hjm8k/crc-debug-hzct9"] Mar 18 21:16:34 crc kubenswrapper[4950]: I0318 21:16:34.820260 4950 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9-host\") on node \"crc\" DevicePath \"\"" Mar 18 21:16:34 crc kubenswrapper[4950]: I0318 21:16:34.820294 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdltw\" (UniqueName: \"kubernetes.io/projected/ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9-kube-api-access-hdltw\") on node \"crc\" DevicePath \"\"" Mar 18 21:16:35 crc kubenswrapper[4950]: I0318 21:16:35.488390 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9" path="/var/lib/kubelet/pods/ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9/volumes" Mar 18 21:16:35 crc kubenswrapper[4950]: I0318 21:16:35.554926 4950 scope.go:117] "RemoveContainer" containerID="c673e8b6123d7b1808035305c8600c3f0c7fa231a1fffcedbc2b9e1d0abe21af" Mar 18 21:16:35 crc kubenswrapper[4950]: I0318 21:16:35.554939 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hjm8k/crc-debug-hzct9" Mar 18 21:16:36 crc kubenswrapper[4950]: I0318 21:16:36.060671 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hjm8k/crc-debug-pdcxl"] Mar 18 21:16:36 crc kubenswrapper[4950]: E0318 21:16:36.061340 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9" containerName="container-00" Mar 18 21:16:36 crc kubenswrapper[4950]: I0318 21:16:36.061368 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9" containerName="container-00" Mar 18 21:16:36 crc kubenswrapper[4950]: I0318 21:16:36.061603 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecb0ce74-f47c-4c2e-aa8a-4ad77e7a07e9" containerName="container-00" Mar 18 21:16:36 crc kubenswrapper[4950]: I0318 21:16:36.062276 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hjm8k/crc-debug-pdcxl" Mar 18 21:16:36 crc kubenswrapper[4950]: I0318 21:16:36.244203 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t69p5\" (UniqueName: \"kubernetes.io/projected/1f206a42-510c-4c4b-8e20-63c93350da03-kube-api-access-t69p5\") pod \"crc-debug-pdcxl\" (UID: \"1f206a42-510c-4c4b-8e20-63c93350da03\") " pod="openshift-must-gather-hjm8k/crc-debug-pdcxl" Mar 18 21:16:36 crc kubenswrapper[4950]: I0318 21:16:36.244581 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f206a42-510c-4c4b-8e20-63c93350da03-host\") pod \"crc-debug-pdcxl\" (UID: \"1f206a42-510c-4c4b-8e20-63c93350da03\") " pod="openshift-must-gather-hjm8k/crc-debug-pdcxl" Mar 18 21:16:36 crc kubenswrapper[4950]: I0318 21:16:36.346095 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f206a42-510c-4c4b-8e20-63c93350da03-host\") pod \"crc-debug-pdcxl\" (UID: \"1f206a42-510c-4c4b-8e20-63c93350da03\") " pod="openshift-must-gather-hjm8k/crc-debug-pdcxl" Mar 18 21:16:36 crc kubenswrapper[4950]: I0318 21:16:36.346148 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t69p5\" (UniqueName: \"kubernetes.io/projected/1f206a42-510c-4c4b-8e20-63c93350da03-kube-api-access-t69p5\") pod \"crc-debug-pdcxl\" (UID: \"1f206a42-510c-4c4b-8e20-63c93350da03\") " pod="openshift-must-gather-hjm8k/crc-debug-pdcxl" Mar 18 21:16:36 crc kubenswrapper[4950]: I0318 21:16:36.346199 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f206a42-510c-4c4b-8e20-63c93350da03-host\") pod \"crc-debug-pdcxl\" (UID: \"1f206a42-510c-4c4b-8e20-63c93350da03\") " pod="openshift-must-gather-hjm8k/crc-debug-pdcxl" Mar 18 21:16:36 crc kubenswrapper[4950]: I0318 21:16:36.382307 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t69p5\" (UniqueName: \"kubernetes.io/projected/1f206a42-510c-4c4b-8e20-63c93350da03-kube-api-access-t69p5\") pod \"crc-debug-pdcxl\" (UID: \"1f206a42-510c-4c4b-8e20-63c93350da03\") " pod="openshift-must-gather-hjm8k/crc-debug-pdcxl" Mar 18 21:16:36 crc kubenswrapper[4950]: I0318 21:16:36.678044 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hjm8k/crc-debug-pdcxl" Mar 18 21:16:37 crc kubenswrapper[4950]: I0318 21:16:37.575041 4950 generic.go:334] "Generic (PLEG): container finished" podID="1f206a42-510c-4c4b-8e20-63c93350da03" containerID="e87b391d5b2138ee5f641d92689130915d09867f939cbc80b4868942c0339ceb" exitCode=0 Mar 18 21:16:37 crc kubenswrapper[4950]: I0318 21:16:37.575548 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hjm8k/crc-debug-pdcxl" event={"ID":"1f206a42-510c-4c4b-8e20-63c93350da03","Type":"ContainerDied","Data":"e87b391d5b2138ee5f641d92689130915d09867f939cbc80b4868942c0339ceb"} Mar 18 21:16:37 crc kubenswrapper[4950]: I0318 21:16:37.575574 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hjm8k/crc-debug-pdcxl" event={"ID":"1f206a42-510c-4c4b-8e20-63c93350da03","Type":"ContainerStarted","Data":"2df8d20b1a77e3221e28e40f63ba0a4bc9e422934bcbcd09fcd5f43b3ed8961c"} Mar 18 21:16:37 crc kubenswrapper[4950]: I0318 21:16:37.938233 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hjm8k/crc-debug-pdcxl"] Mar 18 21:16:37 crc kubenswrapper[4950]: I0318 21:16:37.950475 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hjm8k/crc-debug-pdcxl"] Mar 18 21:16:38 crc kubenswrapper[4950]: I0318 21:16:38.672591 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hjm8k/crc-debug-pdcxl" Mar 18 21:16:38 crc kubenswrapper[4950]: I0318 21:16:38.794549 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f206a42-510c-4c4b-8e20-63c93350da03-host\") pod \"1f206a42-510c-4c4b-8e20-63c93350da03\" (UID: \"1f206a42-510c-4c4b-8e20-63c93350da03\") " Mar 18 21:16:38 crc kubenswrapper[4950]: I0318 21:16:38.794669 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t69p5\" (UniqueName: \"kubernetes.io/projected/1f206a42-510c-4c4b-8e20-63c93350da03-kube-api-access-t69p5\") pod \"1f206a42-510c-4c4b-8e20-63c93350da03\" (UID: \"1f206a42-510c-4c4b-8e20-63c93350da03\") " Mar 18 21:16:38 crc kubenswrapper[4950]: I0318 21:16:38.796016 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f206a42-510c-4c4b-8e20-63c93350da03-host" (OuterVolumeSpecName: "host") pod "1f206a42-510c-4c4b-8e20-63c93350da03" (UID: "1f206a42-510c-4c4b-8e20-63c93350da03"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 21:16:38 crc kubenswrapper[4950]: I0318 21:16:38.801206 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f206a42-510c-4c4b-8e20-63c93350da03-kube-api-access-t69p5" (OuterVolumeSpecName: "kube-api-access-t69p5") pod "1f206a42-510c-4c4b-8e20-63c93350da03" (UID: "1f206a42-510c-4c4b-8e20-63c93350da03"). InnerVolumeSpecName "kube-api-access-t69p5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:16:38 crc kubenswrapper[4950]: I0318 21:16:38.896893 4950 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f206a42-510c-4c4b-8e20-63c93350da03-host\") on node \"crc\" DevicePath \"\"" Mar 18 21:16:38 crc kubenswrapper[4950]: I0318 21:16:38.897115 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t69p5\" (UniqueName: \"kubernetes.io/projected/1f206a42-510c-4c4b-8e20-63c93350da03-kube-api-access-t69p5\") on node \"crc\" DevicePath \"\"" Mar 18 21:16:38 crc kubenswrapper[4950]: I0318 21:16:38.954770 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b79zp" Mar 18 21:16:39 crc kubenswrapper[4950]: I0318 21:16:39.018625 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b79zp" Mar 18 21:16:39 crc kubenswrapper[4950]: I0318 21:16:39.164730 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hjm8k/crc-debug-grqnh"] Mar 18 21:16:39 crc kubenswrapper[4950]: E0318 21:16:39.165132 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f206a42-510c-4c4b-8e20-63c93350da03" containerName="container-00" Mar 18 21:16:39 crc kubenswrapper[4950]: I0318 21:16:39.165148 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f206a42-510c-4c4b-8e20-63c93350da03" containerName="container-00" Mar 18 21:16:39 crc kubenswrapper[4950]: I0318 21:16:39.165316 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f206a42-510c-4c4b-8e20-63c93350da03" containerName="container-00" Mar 18 21:16:39 crc kubenswrapper[4950]: I0318 21:16:39.165952 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hjm8k/crc-debug-grqnh" Mar 18 21:16:39 crc kubenswrapper[4950]: I0318 21:16:39.199572 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b79zp"] Mar 18 21:16:39 crc kubenswrapper[4950]: I0318 21:16:39.309566 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ff8c2297-cd49-473c-bd31-7791c3ff463b-host\") pod \"crc-debug-grqnh\" (UID: \"ff8c2297-cd49-473c-bd31-7791c3ff463b\") " pod="openshift-must-gather-hjm8k/crc-debug-grqnh" Mar 18 21:16:39 crc kubenswrapper[4950]: I0318 21:16:39.309649 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t24xw\" (UniqueName: \"kubernetes.io/projected/ff8c2297-cd49-473c-bd31-7791c3ff463b-kube-api-access-t24xw\") pod \"crc-debug-grqnh\" (UID: \"ff8c2297-cd49-473c-bd31-7791c3ff463b\") " pod="openshift-must-gather-hjm8k/crc-debug-grqnh" Mar 18 21:16:39 crc kubenswrapper[4950]: I0318 21:16:39.412659 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ff8c2297-cd49-473c-bd31-7791c3ff463b-host\") pod \"crc-debug-grqnh\" (UID: \"ff8c2297-cd49-473c-bd31-7791c3ff463b\") " pod="openshift-must-gather-hjm8k/crc-debug-grqnh" Mar 18 21:16:39 crc kubenswrapper[4950]: I0318 21:16:39.412735 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t24xw\" (UniqueName: \"kubernetes.io/projected/ff8c2297-cd49-473c-bd31-7791c3ff463b-kube-api-access-t24xw\") pod \"crc-debug-grqnh\" (UID: \"ff8c2297-cd49-473c-bd31-7791c3ff463b\") " pod="openshift-must-gather-hjm8k/crc-debug-grqnh" Mar 18 21:16:39 crc kubenswrapper[4950]: I0318 21:16:39.413030 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ff8c2297-cd49-473c-bd31-7791c3ff463b-host\") pod \"crc-debug-grqnh\" (UID: \"ff8c2297-cd49-473c-bd31-7791c3ff463b\") " pod="openshift-must-gather-hjm8k/crc-debug-grqnh" Mar 18 21:16:39 crc kubenswrapper[4950]: I0318 21:16:39.431240 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t24xw\" (UniqueName: \"kubernetes.io/projected/ff8c2297-cd49-473c-bd31-7791c3ff463b-kube-api-access-t24xw\") pod \"crc-debug-grqnh\" (UID: \"ff8c2297-cd49-473c-bd31-7791c3ff463b\") " pod="openshift-must-gather-hjm8k/crc-debug-grqnh" Mar 18 21:16:39 crc kubenswrapper[4950]: I0318 21:16:39.489799 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f206a42-510c-4c4b-8e20-63c93350da03" path="/var/lib/kubelet/pods/1f206a42-510c-4c4b-8e20-63c93350da03/volumes" Mar 18 21:16:39 crc kubenswrapper[4950]: I0318 21:16:39.519877 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hjm8k/crc-debug-grqnh" Mar 18 21:16:39 crc kubenswrapper[4950]: W0318 21:16:39.551588 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff8c2297_cd49_473c_bd31_7791c3ff463b.slice/crio-f6ae6e02f3617eac85fa615c9ae8eb29fab13f2a492a7a928b6a34da9c433287 WatchSource:0}: Error finding container f6ae6e02f3617eac85fa615c9ae8eb29fab13f2a492a7a928b6a34da9c433287: Status 404 returned error can't find the container with id f6ae6e02f3617eac85fa615c9ae8eb29fab13f2a492a7a928b6a34da9c433287 Mar 18 21:16:39 crc kubenswrapper[4950]: I0318 21:16:39.593932 4950 scope.go:117] "RemoveContainer" containerID="e87b391d5b2138ee5f641d92689130915d09867f939cbc80b4868942c0339ceb" Mar 18 21:16:39 crc kubenswrapper[4950]: I0318 21:16:39.593992 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hjm8k/crc-debug-pdcxl" Mar 18 21:16:39 crc kubenswrapper[4950]: I0318 21:16:39.595190 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hjm8k/crc-debug-grqnh" event={"ID":"ff8c2297-cd49-473c-bd31-7791c3ff463b","Type":"ContainerStarted","Data":"f6ae6e02f3617eac85fa615c9ae8eb29fab13f2a492a7a928b6a34da9c433287"} Mar 18 21:16:40 crc kubenswrapper[4950]: I0318 21:16:40.608957 4950 generic.go:334] "Generic (PLEG): container finished" podID="ff8c2297-cd49-473c-bd31-7791c3ff463b" containerID="8d97037b880d666575986c5d59c545e32400e2b8cdf0e2f23086c8bf19e13039" exitCode=0 Mar 18 21:16:40 crc kubenswrapper[4950]: I0318 21:16:40.610688 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b79zp" podUID="d1a59a14-c501-4229-9508-aed7307123cf" containerName="registry-server" containerID="cri-o://edf49bcb079f2cd2e01fb04c40776aa03c4d4789bb5fee009462e4a94f357053" gracePeriod=2 Mar 18 21:16:40 crc kubenswrapper[4950]: I0318 21:16:40.609467 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hjm8k/crc-debug-grqnh" event={"ID":"ff8c2297-cd49-473c-bd31-7791c3ff463b","Type":"ContainerDied","Data":"8d97037b880d666575986c5d59c545e32400e2b8cdf0e2f23086c8bf19e13039"} Mar 18 21:16:40 crc kubenswrapper[4950]: I0318 21:16:40.651051 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hjm8k/crc-debug-grqnh"] Mar 18 21:16:40 crc kubenswrapper[4950]: I0318 21:16:40.662578 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hjm8k/crc-debug-grqnh"] Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.049737 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b79zp" Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.246066 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1a59a14-c501-4229-9508-aed7307123cf-utilities\") pod \"d1a59a14-c501-4229-9508-aed7307123cf\" (UID: \"d1a59a14-c501-4229-9508-aed7307123cf\") " Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.246933 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4ld9\" (UniqueName: \"kubernetes.io/projected/d1a59a14-c501-4229-9508-aed7307123cf-kube-api-access-p4ld9\") pod \"d1a59a14-c501-4229-9508-aed7307123cf\" (UID: \"d1a59a14-c501-4229-9508-aed7307123cf\") " Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.246976 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1a59a14-c501-4229-9508-aed7307123cf-catalog-content\") pod \"d1a59a14-c501-4229-9508-aed7307123cf\" (UID: \"d1a59a14-c501-4229-9508-aed7307123cf\") " Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.247543 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1a59a14-c501-4229-9508-aed7307123cf-utilities" (OuterVolumeSpecName: "utilities") pod "d1a59a14-c501-4229-9508-aed7307123cf" (UID: "d1a59a14-c501-4229-9508-aed7307123cf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.253005 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1a59a14-c501-4229-9508-aed7307123cf-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.253211 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1a59a14-c501-4229-9508-aed7307123cf-kube-api-access-p4ld9" (OuterVolumeSpecName: "kube-api-access-p4ld9") pod "d1a59a14-c501-4229-9508-aed7307123cf" (UID: "d1a59a14-c501-4229-9508-aed7307123cf"). InnerVolumeSpecName "kube-api-access-p4ld9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.280486 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1a59a14-c501-4229-9508-aed7307123cf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1a59a14-c501-4229-9508-aed7307123cf" (UID: "d1a59a14-c501-4229-9508-aed7307123cf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.355234 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4ld9\" (UniqueName: \"kubernetes.io/projected/d1a59a14-c501-4229-9508-aed7307123cf-kube-api-access-p4ld9\") on node \"crc\" DevicePath \"\"" Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.355279 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1a59a14-c501-4229-9508-aed7307123cf-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.619662 4950 generic.go:334] "Generic (PLEG): container finished" podID="d1a59a14-c501-4229-9508-aed7307123cf" containerID="edf49bcb079f2cd2e01fb04c40776aa03c4d4789bb5fee009462e4a94f357053" exitCode=0 Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.619738 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b79zp" Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.619784 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b79zp" event={"ID":"d1a59a14-c501-4229-9508-aed7307123cf","Type":"ContainerDied","Data":"edf49bcb079f2cd2e01fb04c40776aa03c4d4789bb5fee009462e4a94f357053"} Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.619852 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b79zp" event={"ID":"d1a59a14-c501-4229-9508-aed7307123cf","Type":"ContainerDied","Data":"3fe8ddccbedb412ca0bf72384296c27b19d21997bfaed94de927b2249720f990"} Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.619874 4950 scope.go:117] "RemoveContainer" containerID="edf49bcb079f2cd2e01fb04c40776aa03c4d4789bb5fee009462e4a94f357053" Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.645074 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b79zp"] Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.649639 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b79zp"] Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.697880 4950 scope.go:117] "RemoveContainer" containerID="45cac7594668627aec1c56fd08cf600e7beed90ff4058c8431b3e86da2106fb3" Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.701177 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hjm8k/crc-debug-grqnh" Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.715652 4950 scope.go:117] "RemoveContainer" containerID="5b649012fbfaba94b3f85a42a8c26be202beee08046dc39bc4972fba10fc6155" Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.764620 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t24xw\" (UniqueName: \"kubernetes.io/projected/ff8c2297-cd49-473c-bd31-7791c3ff463b-kube-api-access-t24xw\") pod \"ff8c2297-cd49-473c-bd31-7791c3ff463b\" (UID: \"ff8c2297-cd49-473c-bd31-7791c3ff463b\") " Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.764679 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ff8c2297-cd49-473c-bd31-7791c3ff463b-host\") pod \"ff8c2297-cd49-473c-bd31-7791c3ff463b\" (UID: \"ff8c2297-cd49-473c-bd31-7791c3ff463b\") " Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.764856 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ff8c2297-cd49-473c-bd31-7791c3ff463b-host" (OuterVolumeSpecName: "host") pod "ff8c2297-cd49-473c-bd31-7791c3ff463b" (UID: "ff8c2297-cd49-473c-bd31-7791c3ff463b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.765323 4950 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ff8c2297-cd49-473c-bd31-7791c3ff463b-host\") on node \"crc\" DevicePath \"\"" Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.775720 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff8c2297-cd49-473c-bd31-7791c3ff463b-kube-api-access-t24xw" (OuterVolumeSpecName: "kube-api-access-t24xw") pod "ff8c2297-cd49-473c-bd31-7791c3ff463b" (UID: "ff8c2297-cd49-473c-bd31-7791c3ff463b"). InnerVolumeSpecName "kube-api-access-t24xw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.789582 4950 scope.go:117] "RemoveContainer" containerID="edf49bcb079f2cd2e01fb04c40776aa03c4d4789bb5fee009462e4a94f357053" Mar 18 21:16:41 crc kubenswrapper[4950]: E0318 21:16:41.790400 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edf49bcb079f2cd2e01fb04c40776aa03c4d4789bb5fee009462e4a94f357053\": container with ID starting with edf49bcb079f2cd2e01fb04c40776aa03c4d4789bb5fee009462e4a94f357053 not found: ID does not exist" containerID="edf49bcb079f2cd2e01fb04c40776aa03c4d4789bb5fee009462e4a94f357053" Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.790621 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edf49bcb079f2cd2e01fb04c40776aa03c4d4789bb5fee009462e4a94f357053"} err="failed to get container status \"edf49bcb079f2cd2e01fb04c40776aa03c4d4789bb5fee009462e4a94f357053\": rpc error: code = NotFound desc = could not find container \"edf49bcb079f2cd2e01fb04c40776aa03c4d4789bb5fee009462e4a94f357053\": container with ID starting with edf49bcb079f2cd2e01fb04c40776aa03c4d4789bb5fee009462e4a94f357053 not found: ID does not exist" Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.790788 4950 scope.go:117] "RemoveContainer" containerID="45cac7594668627aec1c56fd08cf600e7beed90ff4058c8431b3e86da2106fb3" Mar 18 21:16:41 crc kubenswrapper[4950]: E0318 21:16:41.791290 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45cac7594668627aec1c56fd08cf600e7beed90ff4058c8431b3e86da2106fb3\": container with ID starting with 45cac7594668627aec1c56fd08cf600e7beed90ff4058c8431b3e86da2106fb3 not found: ID does not exist" containerID="45cac7594668627aec1c56fd08cf600e7beed90ff4058c8431b3e86da2106fb3" Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.791322 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45cac7594668627aec1c56fd08cf600e7beed90ff4058c8431b3e86da2106fb3"} err="failed to get container status \"45cac7594668627aec1c56fd08cf600e7beed90ff4058c8431b3e86da2106fb3\": rpc error: code = NotFound desc = could not find container \"45cac7594668627aec1c56fd08cf600e7beed90ff4058c8431b3e86da2106fb3\": container with ID starting with 45cac7594668627aec1c56fd08cf600e7beed90ff4058c8431b3e86da2106fb3 not found: ID does not exist" Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.791342 4950 scope.go:117] "RemoveContainer" containerID="5b649012fbfaba94b3f85a42a8c26be202beee08046dc39bc4972fba10fc6155" Mar 18 21:16:41 crc kubenswrapper[4950]: E0318 21:16:41.792569 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b649012fbfaba94b3f85a42a8c26be202beee08046dc39bc4972fba10fc6155\": container with ID starting with 5b649012fbfaba94b3f85a42a8c26be202beee08046dc39bc4972fba10fc6155 not found: ID does not exist" containerID="5b649012fbfaba94b3f85a42a8c26be202beee08046dc39bc4972fba10fc6155" Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.792611 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b649012fbfaba94b3f85a42a8c26be202beee08046dc39bc4972fba10fc6155"} err="failed to get container status \"5b649012fbfaba94b3f85a42a8c26be202beee08046dc39bc4972fba10fc6155\": rpc error: code = NotFound desc = could not find container \"5b649012fbfaba94b3f85a42a8c26be202beee08046dc39bc4972fba10fc6155\": container with ID starting with 5b649012fbfaba94b3f85a42a8c26be202beee08046dc39bc4972fba10fc6155 not found: ID does not exist" Mar 18 21:16:41 crc kubenswrapper[4950]: I0318 21:16:41.867199 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t24xw\" (UniqueName: \"kubernetes.io/projected/ff8c2297-cd49-473c-bd31-7791c3ff463b-kube-api-access-t24xw\") on node \"crc\" DevicePath \"\"" Mar 18 21:16:42 crc kubenswrapper[4950]: I0318 21:16:42.628668 4950 scope.go:117] "RemoveContainer" containerID="8d97037b880d666575986c5d59c545e32400e2b8cdf0e2f23086c8bf19e13039" Mar 18 21:16:42 crc kubenswrapper[4950]: I0318 21:16:42.628761 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hjm8k/crc-debug-grqnh" Mar 18 21:16:43 crc kubenswrapper[4950]: I0318 21:16:43.492320 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1a59a14-c501-4229-9508-aed7307123cf" path="/var/lib/kubelet/pods/d1a59a14-c501-4229-9508-aed7307123cf/volumes" Mar 18 21:16:43 crc kubenswrapper[4950]: I0318 21:16:43.493332 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff8c2297-cd49-473c-bd31-7791c3ff463b" path="/var/lib/kubelet/pods/ff8c2297-cd49-473c-bd31-7791c3ff463b/volumes" Mar 18 21:16:47 crc kubenswrapper[4950]: I0318 21:16:47.485274 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:16:47 crc kubenswrapper[4950]: E0318 21:16:47.486037 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:16:52 crc kubenswrapper[4950]: I0318 21:16:52.370174 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-67mf8"] Mar 18 21:16:52 crc kubenswrapper[4950]: E0318 21:16:52.371082 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1a59a14-c501-4229-9508-aed7307123cf" containerName="registry-server" Mar 18 21:16:52 crc kubenswrapper[4950]: I0318 21:16:52.371096 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1a59a14-c501-4229-9508-aed7307123cf" containerName="registry-server" Mar 18 21:16:52 crc kubenswrapper[4950]: E0318 21:16:52.371112 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff8c2297-cd49-473c-bd31-7791c3ff463b" containerName="container-00" Mar 18 21:16:52 crc kubenswrapper[4950]: I0318 21:16:52.371119 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff8c2297-cd49-473c-bd31-7791c3ff463b" containerName="container-00" Mar 18 21:16:52 crc kubenswrapper[4950]: E0318 21:16:52.371135 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1a59a14-c501-4229-9508-aed7307123cf" containerName="extract-content" Mar 18 21:16:52 crc kubenswrapper[4950]: I0318 21:16:52.371141 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1a59a14-c501-4229-9508-aed7307123cf" containerName="extract-content" Mar 18 21:16:52 crc kubenswrapper[4950]: E0318 21:16:52.371150 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1a59a14-c501-4229-9508-aed7307123cf" containerName="extract-utilities" Mar 18 21:16:52 crc kubenswrapper[4950]: I0318 21:16:52.371156 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1a59a14-c501-4229-9508-aed7307123cf" containerName="extract-utilities" Mar 18 21:16:52 crc kubenswrapper[4950]: I0318 21:16:52.371327 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1a59a14-c501-4229-9508-aed7307123cf" containerName="registry-server" Mar 18 21:16:52 crc kubenswrapper[4950]: I0318 21:16:52.371346 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff8c2297-cd49-473c-bd31-7791c3ff463b" containerName="container-00" Mar 18 21:16:52 crc kubenswrapper[4950]: I0318 21:16:52.372639 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-67mf8" Mar 18 21:16:52 crc kubenswrapper[4950]: I0318 21:16:52.382613 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-67mf8"] Mar 18 21:16:52 crc kubenswrapper[4950]: I0318 21:16:52.413976 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/046b4599-f9ba-48b1-b32d-7fdbce471f29-utilities\") pod \"community-operators-67mf8\" (UID: \"046b4599-f9ba-48b1-b32d-7fdbce471f29\") " pod="openshift-marketplace/community-operators-67mf8" Mar 18 21:16:52 crc kubenswrapper[4950]: I0318 21:16:52.414078 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/046b4599-f9ba-48b1-b32d-7fdbce471f29-catalog-content\") pod \"community-operators-67mf8\" (UID: \"046b4599-f9ba-48b1-b32d-7fdbce471f29\") " pod="openshift-marketplace/community-operators-67mf8" Mar 18 21:16:52 crc kubenswrapper[4950]: I0318 21:16:52.414355 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7h4n\" (UniqueName: \"kubernetes.io/projected/046b4599-f9ba-48b1-b32d-7fdbce471f29-kube-api-access-q7h4n\") pod \"community-operators-67mf8\" (UID: \"046b4599-f9ba-48b1-b32d-7fdbce471f29\") " pod="openshift-marketplace/community-operators-67mf8" Mar 18 21:16:52 crc kubenswrapper[4950]: I0318 21:16:52.516710 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/046b4599-f9ba-48b1-b32d-7fdbce471f29-utilities\") pod \"community-operators-67mf8\" (UID: \"046b4599-f9ba-48b1-b32d-7fdbce471f29\") " pod="openshift-marketplace/community-operators-67mf8" Mar 18 21:16:52 crc kubenswrapper[4950]: I0318 21:16:52.516760 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/046b4599-f9ba-48b1-b32d-7fdbce471f29-catalog-content\") pod \"community-operators-67mf8\" (UID: \"046b4599-f9ba-48b1-b32d-7fdbce471f29\") " pod="openshift-marketplace/community-operators-67mf8" Mar 18 21:16:52 crc kubenswrapper[4950]: I0318 21:16:52.516912 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7h4n\" (UniqueName: \"kubernetes.io/projected/046b4599-f9ba-48b1-b32d-7fdbce471f29-kube-api-access-q7h4n\") pod \"community-operators-67mf8\" (UID: \"046b4599-f9ba-48b1-b32d-7fdbce471f29\") " pod="openshift-marketplace/community-operators-67mf8" Mar 18 21:16:52 crc kubenswrapper[4950]: I0318 21:16:52.517199 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/046b4599-f9ba-48b1-b32d-7fdbce471f29-utilities\") pod \"community-operators-67mf8\" (UID: \"046b4599-f9ba-48b1-b32d-7fdbce471f29\") " pod="openshift-marketplace/community-operators-67mf8" Mar 18 21:16:52 crc kubenswrapper[4950]: I0318 21:16:52.517312 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/046b4599-f9ba-48b1-b32d-7fdbce471f29-catalog-content\") pod \"community-operators-67mf8\" (UID: \"046b4599-f9ba-48b1-b32d-7fdbce471f29\") " pod="openshift-marketplace/community-operators-67mf8" Mar 18 21:16:52 crc kubenswrapper[4950]: I0318 21:16:52.549644 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7h4n\" (UniqueName: \"kubernetes.io/projected/046b4599-f9ba-48b1-b32d-7fdbce471f29-kube-api-access-q7h4n\") pod \"community-operators-67mf8\" (UID: \"046b4599-f9ba-48b1-b32d-7fdbce471f29\") " pod="openshift-marketplace/community-operators-67mf8" Mar 18 21:16:52 crc kubenswrapper[4950]: I0318 21:16:52.727659 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-67mf8" Mar 18 21:16:53 crc kubenswrapper[4950]: I0318 21:16:53.137740 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-67mf8"] Mar 18 21:16:53 crc kubenswrapper[4950]: I0318 21:16:53.724239 4950 generic.go:334] "Generic (PLEG): container finished" podID="046b4599-f9ba-48b1-b32d-7fdbce471f29" containerID="22b79cc46edac2b579e9db87e353d8972d91d7363a74fbad4cc321e33eda1364" exitCode=0 Mar 18 21:16:53 crc kubenswrapper[4950]: I0318 21:16:53.724304 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-67mf8" event={"ID":"046b4599-f9ba-48b1-b32d-7fdbce471f29","Type":"ContainerDied","Data":"22b79cc46edac2b579e9db87e353d8972d91d7363a74fbad4cc321e33eda1364"} Mar 18 21:16:53 crc kubenswrapper[4950]: I0318 21:16:53.724607 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-67mf8" event={"ID":"046b4599-f9ba-48b1-b32d-7fdbce471f29","Type":"ContainerStarted","Data":"6b8b8135b54b648bc998feb5d094a32d6035b0190cc82f67cefe97e71122a668"} Mar 18 21:16:55 crc kubenswrapper[4950]: I0318 21:16:55.741134 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-67mf8" event={"ID":"046b4599-f9ba-48b1-b32d-7fdbce471f29","Type":"ContainerStarted","Data":"858f3144bd6e58c256d16c120f597aaf622caf3c7b33052bb599cc8a2927525c"} Mar 18 21:16:56 crc kubenswrapper[4950]: I0318 21:16:56.750808 4950 generic.go:334] "Generic (PLEG): container finished" podID="046b4599-f9ba-48b1-b32d-7fdbce471f29" containerID="858f3144bd6e58c256d16c120f597aaf622caf3c7b33052bb599cc8a2927525c" exitCode=0 Mar 18 21:16:56 crc kubenswrapper[4950]: I0318 21:16:56.750899 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-67mf8" event={"ID":"046b4599-f9ba-48b1-b32d-7fdbce471f29","Type":"ContainerDied","Data":"858f3144bd6e58c256d16c120f597aaf622caf3c7b33052bb599cc8a2927525c"} Mar 18 21:16:57 crc kubenswrapper[4950]: I0318 21:16:57.762751 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-67mf8" event={"ID":"046b4599-f9ba-48b1-b32d-7fdbce471f29","Type":"ContainerStarted","Data":"b5eb3bb10e408f48bdf31c62fde493ff6a090f2460160cd46d1252f3bfe2dc5d"} Mar 18 21:16:57 crc kubenswrapper[4950]: I0318 21:16:57.786876 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-67mf8" podStartSLOduration=2.359738806 podStartE2EDuration="5.786858369s" podCreationTimestamp="2026-03-18 21:16:52 +0000 UTC" firstStartedPulling="2026-03-18 21:16:53.72595263 +0000 UTC m=+4226.966794498" lastFinishedPulling="2026-03-18 21:16:57.153072193 +0000 UTC m=+4230.393914061" observedRunningTime="2026-03-18 21:16:57.782528906 +0000 UTC m=+4231.023370774" watchObservedRunningTime="2026-03-18 21:16:57.786858369 +0000 UTC m=+4231.027700237" Mar 18 21:16:58 crc kubenswrapper[4950]: I0318 21:16:58.480448 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:16:58 crc kubenswrapper[4950]: E0318 21:16:58.480810 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:17:01 crc kubenswrapper[4950]: I0318 21:17:01.415292 4950 scope.go:117] "RemoveContainer" containerID="0ca7d6412521119010b6f028cb103fb9b20d1e5515c87aa7a804389c40feee74" Mar 18 21:17:02 crc kubenswrapper[4950]: I0318 21:17:02.728706 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-67mf8" Mar 18 21:17:02 crc kubenswrapper[4950]: I0318 21:17:02.730385 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-67mf8" Mar 18 21:17:02 crc kubenswrapper[4950]: I0318 21:17:02.784579 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-67mf8" Mar 18 21:17:02 crc kubenswrapper[4950]: I0318 21:17:02.850503 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-67mf8" Mar 18 21:17:03 crc kubenswrapper[4950]: I0318 21:17:03.021601 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-67mf8"] Mar 18 21:17:04 crc kubenswrapper[4950]: I0318 21:17:04.822877 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-67mf8" podUID="046b4599-f9ba-48b1-b32d-7fdbce471f29" containerName="registry-server" containerID="cri-o://b5eb3bb10e408f48bdf31c62fde493ff6a090f2460160cd46d1252f3bfe2dc5d" gracePeriod=2 Mar 18 21:17:06 crc kubenswrapper[4950]: I0318 21:17:06.819583 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-67mf8" Mar 18 21:17:06 crc kubenswrapper[4950]: I0318 21:17:06.841093 4950 generic.go:334] "Generic (PLEG): container finished" podID="046b4599-f9ba-48b1-b32d-7fdbce471f29" containerID="b5eb3bb10e408f48bdf31c62fde493ff6a090f2460160cd46d1252f3bfe2dc5d" exitCode=0 Mar 18 21:17:06 crc kubenswrapper[4950]: I0318 21:17:06.841142 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-67mf8" event={"ID":"046b4599-f9ba-48b1-b32d-7fdbce471f29","Type":"ContainerDied","Data":"b5eb3bb10e408f48bdf31c62fde493ff6a090f2460160cd46d1252f3bfe2dc5d"} Mar 18 21:17:06 crc kubenswrapper[4950]: I0318 21:17:06.841174 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-67mf8" event={"ID":"046b4599-f9ba-48b1-b32d-7fdbce471f29","Type":"ContainerDied","Data":"6b8b8135b54b648bc998feb5d094a32d6035b0190cc82f67cefe97e71122a668"} Mar 18 21:17:06 crc kubenswrapper[4950]: I0318 21:17:06.841193 4950 scope.go:117] "RemoveContainer" containerID="b5eb3bb10e408f48bdf31c62fde493ff6a090f2460160cd46d1252f3bfe2dc5d" Mar 18 21:17:06 crc kubenswrapper[4950]: I0318 21:17:06.841332 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-67mf8" Mar 18 21:17:06 crc kubenswrapper[4950]: I0318 21:17:06.885790 4950 scope.go:117] "RemoveContainer" containerID="858f3144bd6e58c256d16c120f597aaf622caf3c7b33052bb599cc8a2927525c" Mar 18 21:17:06 crc kubenswrapper[4950]: I0318 21:17:06.905774 4950 scope.go:117] "RemoveContainer" containerID="22b79cc46edac2b579e9db87e353d8972d91d7363a74fbad4cc321e33eda1364" Mar 18 21:17:06 crc kubenswrapper[4950]: I0318 21:17:06.945162 4950 scope.go:117] "RemoveContainer" containerID="b5eb3bb10e408f48bdf31c62fde493ff6a090f2460160cd46d1252f3bfe2dc5d" Mar 18 21:17:06 crc kubenswrapper[4950]: E0318 21:17:06.950270 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5eb3bb10e408f48bdf31c62fde493ff6a090f2460160cd46d1252f3bfe2dc5d\": container with ID starting with b5eb3bb10e408f48bdf31c62fde493ff6a090f2460160cd46d1252f3bfe2dc5d not found: ID does not exist" containerID="b5eb3bb10e408f48bdf31c62fde493ff6a090f2460160cd46d1252f3bfe2dc5d" Mar 18 21:17:06 crc kubenswrapper[4950]: I0318 21:17:06.950324 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5eb3bb10e408f48bdf31c62fde493ff6a090f2460160cd46d1252f3bfe2dc5d"} err="failed to get container status \"b5eb3bb10e408f48bdf31c62fde493ff6a090f2460160cd46d1252f3bfe2dc5d\": rpc error: code = NotFound desc = could not find container \"b5eb3bb10e408f48bdf31c62fde493ff6a090f2460160cd46d1252f3bfe2dc5d\": container with ID starting with b5eb3bb10e408f48bdf31c62fde493ff6a090f2460160cd46d1252f3bfe2dc5d not found: ID does not exist" Mar 18 21:17:06 crc kubenswrapper[4950]: I0318 21:17:06.950350 4950 scope.go:117] "RemoveContainer" containerID="858f3144bd6e58c256d16c120f597aaf622caf3c7b33052bb599cc8a2927525c" Mar 18 21:17:06 crc kubenswrapper[4950]: E0318 21:17:06.950793 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"858f3144bd6e58c256d16c120f597aaf622caf3c7b33052bb599cc8a2927525c\": container with ID starting with 858f3144bd6e58c256d16c120f597aaf622caf3c7b33052bb599cc8a2927525c not found: ID does not exist" containerID="858f3144bd6e58c256d16c120f597aaf622caf3c7b33052bb599cc8a2927525c" Mar 18 21:17:06 crc kubenswrapper[4950]: I0318 21:17:06.950815 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"858f3144bd6e58c256d16c120f597aaf622caf3c7b33052bb599cc8a2927525c"} err="failed to get container status \"858f3144bd6e58c256d16c120f597aaf622caf3c7b33052bb599cc8a2927525c\": rpc error: code = NotFound desc = could not find container \"858f3144bd6e58c256d16c120f597aaf622caf3c7b33052bb599cc8a2927525c\": container with ID starting with 858f3144bd6e58c256d16c120f597aaf622caf3c7b33052bb599cc8a2927525c not found: ID does not exist" Mar 18 21:17:06 crc kubenswrapper[4950]: I0318 21:17:06.950827 4950 scope.go:117] "RemoveContainer" containerID="22b79cc46edac2b579e9db87e353d8972d91d7363a74fbad4cc321e33eda1364" Mar 18 21:17:06 crc kubenswrapper[4950]: E0318 21:17:06.951149 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22b79cc46edac2b579e9db87e353d8972d91d7363a74fbad4cc321e33eda1364\": container with ID starting with 22b79cc46edac2b579e9db87e353d8972d91d7363a74fbad4cc321e33eda1364 not found: ID does not exist" containerID="22b79cc46edac2b579e9db87e353d8972d91d7363a74fbad4cc321e33eda1364" Mar 18 21:17:06 crc kubenswrapper[4950]: I0318 21:17:06.951174 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22b79cc46edac2b579e9db87e353d8972d91d7363a74fbad4cc321e33eda1364"} err="failed to get container status \"22b79cc46edac2b579e9db87e353d8972d91d7363a74fbad4cc321e33eda1364\": rpc error: code = NotFound desc = could not find container \"22b79cc46edac2b579e9db87e353d8972d91d7363a74fbad4cc321e33eda1364\": container with ID starting with 22b79cc46edac2b579e9db87e353d8972d91d7363a74fbad4cc321e33eda1364 not found: ID does not exist" Mar 18 21:17:06 crc kubenswrapper[4950]: I0318 21:17:06.972319 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7h4n\" (UniqueName: \"kubernetes.io/projected/046b4599-f9ba-48b1-b32d-7fdbce471f29-kube-api-access-q7h4n\") pod \"046b4599-f9ba-48b1-b32d-7fdbce471f29\" (UID: \"046b4599-f9ba-48b1-b32d-7fdbce471f29\") " Mar 18 21:17:06 crc kubenswrapper[4950]: I0318 21:17:06.972464 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/046b4599-f9ba-48b1-b32d-7fdbce471f29-catalog-content\") pod \"046b4599-f9ba-48b1-b32d-7fdbce471f29\" (UID: \"046b4599-f9ba-48b1-b32d-7fdbce471f29\") " Mar 18 21:17:06 crc kubenswrapper[4950]: I0318 21:17:06.978711 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/046b4599-f9ba-48b1-b32d-7fdbce471f29-utilities\") pod \"046b4599-f9ba-48b1-b32d-7fdbce471f29\" (UID: \"046b4599-f9ba-48b1-b32d-7fdbce471f29\") " Mar 18 21:17:06 crc kubenswrapper[4950]: I0318 21:17:06.980255 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/046b4599-f9ba-48b1-b32d-7fdbce471f29-utilities" (OuterVolumeSpecName: "utilities") pod "046b4599-f9ba-48b1-b32d-7fdbce471f29" (UID: "046b4599-f9ba-48b1-b32d-7fdbce471f29"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:17:06 crc kubenswrapper[4950]: I0318 21:17:06.986654 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/046b4599-f9ba-48b1-b32d-7fdbce471f29-kube-api-access-q7h4n" (OuterVolumeSpecName: "kube-api-access-q7h4n") pod "046b4599-f9ba-48b1-b32d-7fdbce471f29" (UID: "046b4599-f9ba-48b1-b32d-7fdbce471f29"). InnerVolumeSpecName "kube-api-access-q7h4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:17:07 crc kubenswrapper[4950]: I0318 21:17:07.027445 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/046b4599-f9ba-48b1-b32d-7fdbce471f29-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "046b4599-f9ba-48b1-b32d-7fdbce471f29" (UID: "046b4599-f9ba-48b1-b32d-7fdbce471f29"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:17:07 crc kubenswrapper[4950]: I0318 21:17:07.082335 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/046b4599-f9ba-48b1-b32d-7fdbce471f29-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 21:17:07 crc kubenswrapper[4950]: I0318 21:17:07.082577 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/046b4599-f9ba-48b1-b32d-7fdbce471f29-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 21:17:07 crc kubenswrapper[4950]: I0318 21:17:07.082658 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7h4n\" (UniqueName: \"kubernetes.io/projected/046b4599-f9ba-48b1-b32d-7fdbce471f29-kube-api-access-q7h4n\") on node \"crc\" DevicePath \"\"" Mar 18 21:17:07 crc kubenswrapper[4950]: I0318 21:17:07.175067 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-67mf8"] Mar 18 21:17:07 crc kubenswrapper[4950]: I0318 21:17:07.185120 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-67mf8"] Mar 18 21:17:07 crc kubenswrapper[4950]: I0318 21:17:07.490899 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="046b4599-f9ba-48b1-b32d-7fdbce471f29" path="/var/lib/kubelet/pods/046b4599-f9ba-48b1-b32d-7fdbce471f29/volumes" Mar 18 21:17:09 crc kubenswrapper[4950]: I0318 21:17:09.481506 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:17:09 crc kubenswrapper[4950]: I0318 21:17:09.866819 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerStarted","Data":"6fb2941ff91c9d7625aec6265d6c6aecb9312056ca715d8dc848cdcf34657cae"} Mar 18 21:17:46 crc kubenswrapper[4950]: I0318 21:17:46.378134 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7d5775f6d4-lshhl_f111df34-12e9-441b-a362-e94471ec8eaf/barbican-api/0.log" Mar 18 21:17:46 crc kubenswrapper[4950]: I0318 21:17:46.547209 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7d5775f6d4-lshhl_f111df34-12e9-441b-a362-e94471ec8eaf/barbican-api-log/0.log" Mar 18 21:17:46 crc kubenswrapper[4950]: I0318 21:17:46.701163 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-546478654b-mwfqm_72fe734a-54a6-4391-a125-04ddd4c9c890/barbican-keystone-listener/0.log" Mar 18 21:17:46 crc kubenswrapper[4950]: I0318 21:17:46.706493 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-546478654b-mwfqm_72fe734a-54a6-4391-a125-04ddd4c9c890/barbican-keystone-listener-log/0.log" Mar 18 21:17:46 crc kubenswrapper[4950]: I0318 21:17:46.884609 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6c48df686c-lz5gl_f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9/barbican-worker/0.log" Mar 18 21:17:46 crc kubenswrapper[4950]: I0318 21:17:46.936452 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6c48df686c-lz5gl_f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9/barbican-worker-log/0.log" Mar 18 21:17:47 crc kubenswrapper[4950]: I0318 21:17:47.070057 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf_8230645b-e426-4c0e-a28b-565ba98d18a1/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:17:47 crc kubenswrapper[4950]: I0318 21:17:47.159387 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b5ef7308-b687-45b3-bcf1-1ef804092712/ceilometer-central-agent/0.log" Mar 18 21:17:47 crc kubenswrapper[4950]: I0318 21:17:47.220944 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b5ef7308-b687-45b3-bcf1-1ef804092712/ceilometer-notification-agent/0.log" Mar 18 21:17:47 crc kubenswrapper[4950]: I0318 21:17:47.271894 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b5ef7308-b687-45b3-bcf1-1ef804092712/proxy-httpd/0.log" Mar 18 21:17:47 crc kubenswrapper[4950]: I0318 21:17:47.399682 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b5ef7308-b687-45b3-bcf1-1ef804092712/sg-core/0.log" Mar 18 21:17:47 crc kubenswrapper[4950]: I0318 21:17:47.444542 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj_2e43754d-5118-4f9d-afe0-c70c6d4e7589/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:17:47 crc kubenswrapper[4950]: I0318 21:17:47.620359 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f_e53816d0-c4c0-412f-9568-af0a96c87145/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:17:47 crc kubenswrapper[4950]: I0318 21:17:47.701133 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_1bae85ae-0545-46ac-90fa-c8f967648ed4/cinder-api/0.log" Mar 18 21:17:47 crc kubenswrapper[4950]: I0318 21:17:47.824931 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_1bae85ae-0545-46ac-90fa-c8f967648ed4/cinder-api-log/0.log" Mar 18 21:17:47 crc kubenswrapper[4950]: I0318 21:17:47.978059 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_9afecfad-52c5-439c-a3bc-a56640e9a79a/probe/0.log" Mar 18 21:17:48 crc kubenswrapper[4950]: I0318 21:17:48.018514 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_9afecfad-52c5-439c-a3bc-a56640e9a79a/cinder-backup/0.log" Mar 18 21:17:48 crc kubenswrapper[4950]: I0318 21:17:48.171727 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_fbb2e2de-55ef-4bc1-a8df-1428e18c664e/cinder-scheduler/0.log" Mar 18 21:17:48 crc kubenswrapper[4950]: I0318 21:17:48.232654 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_fbb2e2de-55ef-4bc1-a8df-1428e18c664e/probe/0.log" Mar 18 21:17:48 crc kubenswrapper[4950]: I0318 21:17:48.373208 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_e22cc0cd-ebf4-4f03-91f3-1db5f2eac881/cinder-volume/0.log" Mar 18 21:17:48 crc kubenswrapper[4950]: I0318 21:17:48.458430 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_e22cc0cd-ebf4-4f03-91f3-1db5f2eac881/probe/0.log" Mar 18 21:17:48 crc kubenswrapper[4950]: I0318 21:17:48.522881 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q_f4dc08cb-4f1e-4430-b522-62f890be4336/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:17:48 crc kubenswrapper[4950]: I0318 21:17:48.735148 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb_3f18ebe5-e07d-40a7-911e-8d3ad683c78c/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:17:48 crc kubenswrapper[4950]: I0318 21:17:48.746632 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7f55f6cb75-46f5l_f9997bd5-b01a-490e-b4c3-b258b277d73c/init/0.log" Mar 18 21:17:48 crc kubenswrapper[4950]: I0318 21:17:48.938912 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7f55f6cb75-46f5l_f9997bd5-b01a-490e-b4c3-b258b277d73c/init/0.log" Mar 18 21:17:49 crc kubenswrapper[4950]: I0318 21:17:49.157002 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7f55f6cb75-46f5l_f9997bd5-b01a-490e-b4c3-b258b277d73c/dnsmasq-dns/0.log" Mar 18 21:17:49 crc kubenswrapper[4950]: I0318 21:17:49.188155 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_c8ec3994-fd11-4e42-80c7-01857df19a74/glance-httpd/0.log" Mar 18 21:17:49 crc kubenswrapper[4950]: I0318 21:17:49.227701 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_c8ec3994-fd11-4e42-80c7-01857df19a74/glance-log/0.log" Mar 18 21:17:49 crc kubenswrapper[4950]: I0318 21:17:49.346121 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd/glance-httpd/0.log" Mar 18 21:17:49 crc kubenswrapper[4950]: I0318 21:17:49.406431 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd/glance-log/0.log" Mar 18 21:17:49 crc kubenswrapper[4950]: I0318 21:17:49.621575 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7449579958-pdtkd_d648add7-37b3-44fa-a3e2-4ca57d274ca0/horizon/0.log" Mar 18 21:17:49 crc kubenswrapper[4950]: I0318 21:17:49.809979 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7449579958-pdtkd_d648add7-37b3-44fa-a3e2-4ca57d274ca0/horizon-log/0.log" Mar 18 21:17:49 crc kubenswrapper[4950]: I0318 21:17:49.816809 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-472gf_096e1bc2-b416-48d1-bc00-082fb7023df9/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:17:49 crc kubenswrapper[4950]: I0318 21:17:49.983909 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-w5fff_29a51c30-1655-4c6c-9e74-9b8797196361/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:17:50 crc kubenswrapper[4950]: I0318 21:17:50.278236 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-564b994b87-rbnt8_fb4d97b7-3492-4ccb-8489-98a9882ea782/keystone-api/0.log" Mar 18 21:17:50 crc kubenswrapper[4950]: I0318 21:17:50.335751 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29564461-jwfx5_e0d07dee-3b03-4b75-b150-85010f0bcf29/keystone-cron/0.log" Mar 18 21:17:50 crc kubenswrapper[4950]: I0318 21:17:50.427201 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e/kube-state-metrics/0.log" Mar 18 21:17:50 crc kubenswrapper[4950]: I0318 21:17:50.619161 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp_22d4a06f-ff3c-4fbf-9eae-8be967f1febd/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:17:50 crc kubenswrapper[4950]: I0318 21:17:50.782934 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_9f52ad06-7e1b-4736-9441-9b1749f6d325/manila-api/0.log" Mar 18 21:17:50 crc kubenswrapper[4950]: I0318 21:17:50.855288 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_9f52ad06-7e1b-4736-9441-9b1749f6d325/manila-api-log/0.log" Mar 18 21:17:51 crc kubenswrapper[4950]: I0318 21:17:51.045973 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_a9bcf400-c9cc-4556-a2e2-6b7214c68579/manila-scheduler/0.log" Mar 18 21:17:51 crc kubenswrapper[4950]: I0318 21:17:51.091737 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_a9bcf400-c9cc-4556-a2e2-6b7214c68579/probe/0.log" Mar 18 21:17:51 crc kubenswrapper[4950]: I0318 21:17:51.188693 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_139d12a8-de25-41a7-a92d-440cbeec98eb/manila-share/0.log" Mar 18 21:17:51 crc kubenswrapper[4950]: I0318 21:17:51.273725 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_139d12a8-de25-41a7-a92d-440cbeec98eb/probe/0.log" Mar 18 21:17:51 crc kubenswrapper[4950]: I0318 21:17:51.641565 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-d5c6d4b4c-6cdzp_2afba89a-006e-472f-8e49-0e8d69cd86c7/neutron-httpd/0.log" Mar 18 21:17:51 crc kubenswrapper[4950]: I0318 21:17:51.796262 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257_619538d3-bc36-4cfb-8eb7-9fad2bef5d4c/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:17:51 crc kubenswrapper[4950]: I0318 21:17:51.839766 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-d5c6d4b4c-6cdzp_2afba89a-006e-472f-8e49-0e8d69cd86c7/neutron-api/0.log" Mar 18 21:17:52 crc kubenswrapper[4950]: I0318 21:17:52.462298 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_603bbfd0-1fdf-44f7-bc34-ce9111ee204f/nova-api-log/0.log" Mar 18 21:17:52 crc kubenswrapper[4950]: I0318 21:17:52.524206 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_5fbcf2da-2f82-4065-85e5-9f701c550afb/nova-cell0-conductor-conductor/0.log" Mar 18 21:17:52 crc kubenswrapper[4950]: I0318 21:17:52.864819 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_41637a9b-2c22-4627-85df-b902f8d35ceb/nova-cell1-conductor-conductor/0.log" Mar 18 21:17:52 crc kubenswrapper[4950]: I0318 21:17:52.866621 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_49765550-f019-4bdd-8a9c-ffadf57e59c4/nova-cell1-novncproxy-novncproxy/0.log" Mar 18 21:17:52 crc kubenswrapper[4950]: I0318 21:17:52.869912 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_603bbfd0-1fdf-44f7-bc34-ce9111ee204f/nova-api-api/0.log" Mar 18 21:17:53 crc kubenswrapper[4950]: I0318 21:17:53.611538 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx_006a96ed-5832-49f8-a39c-d433a2b1bfd4/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:17:53 crc kubenswrapper[4950]: I0318 21:17:53.693764 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f05a94b0-defc-4c6c-af5c-3ce8346261a3/nova-metadata-log/0.log" Mar 18 21:17:54 crc kubenswrapper[4950]: I0318 21:17:54.142090 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_1e49132d-6e98-4270-8ba0-7500e8a84bfd/nova-scheduler-scheduler/0.log" Mar 18 21:17:54 crc kubenswrapper[4950]: I0318 21:17:54.144291 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_aedf1977-d498-4b4f-a92b-f84be888045d/mysql-bootstrap/0.log" Mar 18 21:17:54 crc kubenswrapper[4950]: I0318 21:17:54.317491 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_aedf1977-d498-4b4f-a92b-f84be888045d/mysql-bootstrap/0.log" Mar 18 21:17:54 crc kubenswrapper[4950]: I0318 21:17:54.321535 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f05a94b0-defc-4c6c-af5c-3ce8346261a3/nova-metadata-metadata/0.log" Mar 18 21:17:54 crc kubenswrapper[4950]: I0318 21:17:54.386308 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_aedf1977-d498-4b4f-a92b-f84be888045d/galera/0.log" Mar 18 21:17:55 crc kubenswrapper[4950]: I0318 21:17:55.138749 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3/mysql-bootstrap/0.log" Mar 18 21:17:55 crc kubenswrapper[4950]: I0318 21:17:55.378237 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3/mysql-bootstrap/0.log" Mar 18 21:17:55 crc kubenswrapper[4950]: I0318 21:17:55.391748 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3/galera/0.log" Mar 18 21:17:55 crc kubenswrapper[4950]: I0318 21:17:55.401517 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_e0941268-2503-4b8e-afeb-0a655b52f505/openstackclient/0.log" Mar 18 21:17:55 crc kubenswrapper[4950]: I0318 21:17:55.696095 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-9wnnb_3f70fe2d-f600-42da-8a3d-a4f814374ca7/openstack-network-exporter/0.log" Mar 18 21:17:55 crc kubenswrapper[4950]: I0318 21:17:55.814997 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lvf7g_21aec953-2e02-48d3-85ff-872d48c4d661/ovsdb-server-init/0.log" Mar 18 21:17:55 crc kubenswrapper[4950]: I0318 21:17:55.983527 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lvf7g_21aec953-2e02-48d3-85ff-872d48c4d661/ovsdb-server/0.log" Mar 18 21:17:56 crc kubenswrapper[4950]: I0318 21:17:56.088808 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lvf7g_21aec953-2e02-48d3-85ff-872d48c4d661/ovsdb-server-init/0.log" Mar 18 21:17:56 crc kubenswrapper[4950]: I0318 21:17:56.092321 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lvf7g_21aec953-2e02-48d3-85ff-872d48c4d661/ovs-vswitchd/0.log" Mar 18 21:17:56 crc kubenswrapper[4950]: I0318 21:17:56.217747 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-zntd5_b72c4ff1-1234-41b7-9faa-b5fd15749084/ovn-controller/0.log" Mar 18 21:17:56 crc kubenswrapper[4950]: I0318 21:17:56.417742 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-hnlg9_67bb928c-890d-480c-8ce7-ec7b299cb1ee/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:17:56 crc kubenswrapper[4950]: I0318 21:17:56.539227 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a3a71276-b0fa-481f-a91e-551d017c6462/openstack-network-exporter/0.log" Mar 18 21:17:56 crc kubenswrapper[4950]: I0318 21:17:56.686042 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_49c49971-1a60-453f-8283-922ea50cc323/openstack-network-exporter/0.log" Mar 18 21:17:56 crc kubenswrapper[4950]: I0318 21:17:56.776949 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a3a71276-b0fa-481f-a91e-551d017c6462/ovn-northd/0.log" Mar 18 21:17:56 crc kubenswrapper[4950]: I0318 21:17:56.818396 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_49c49971-1a60-453f-8283-922ea50cc323/ovsdbserver-nb/0.log" Mar 18 21:17:57 crc kubenswrapper[4950]: I0318 21:17:57.082467 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f37c3f8b-786f-42f6-8c5f-4db367edf6a5/ovsdbserver-sb/0.log" Mar 18 21:17:57 crc kubenswrapper[4950]: I0318 21:17:57.083300 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f37c3f8b-786f-42f6-8c5f-4db367edf6a5/openstack-network-exporter/0.log" Mar 18 21:17:57 crc kubenswrapper[4950]: I0318 21:17:57.607095 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5889fd87f8-6w269_96409976-b947-4ba5-af6e-eeb0ae687b0b/placement-api/0.log" Mar 18 21:17:57 crc kubenswrapper[4950]: I0318 21:17:57.720458 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_435fb3ca-c4ad-4b0a-a832-88f4a7203def/setup-container/0.log" Mar 18 21:17:57 crc kubenswrapper[4950]: I0318 21:17:57.807535 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5889fd87f8-6w269_96409976-b947-4ba5-af6e-eeb0ae687b0b/placement-log/0.log" Mar 18 21:17:57 crc kubenswrapper[4950]: I0318 21:17:57.838986 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_435fb3ca-c4ad-4b0a-a832-88f4a7203def/setup-container/0.log" Mar 18 21:17:57 crc kubenswrapper[4950]: I0318 21:17:57.986036 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_435fb3ca-c4ad-4b0a-a832-88f4a7203def/rabbitmq/0.log" Mar 18 21:17:58 crc kubenswrapper[4950]: I0318 21:17:58.028080 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_27be57d1-fc6d-4fe7-a6b3-c7276d5ce547/setup-container/0.log" Mar 18 21:17:58 crc kubenswrapper[4950]: I0318 21:17:58.296663 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_27be57d1-fc6d-4fe7-a6b3-c7276d5ce547/setup-container/0.log" Mar 18 21:17:58 crc kubenswrapper[4950]: I0318 21:17:58.373902 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4_afdcd712-fe52-40ce-a610-82b7dd39a59c/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:17:58 crc kubenswrapper[4950]: I0318 21:17:58.456506 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_27be57d1-fc6d-4fe7-a6b3-c7276d5ce547/rabbitmq/0.log" Mar 18 21:17:58 crc kubenswrapper[4950]: I0318 21:17:58.583473 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq_11355e20-3f52-4010-8c35-0318c88ad131/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:17:58 crc kubenswrapper[4950]: I0318 21:17:58.734128 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-gsrdb_5e58539e-bccd-459c-b1f7-2696ffe9a402/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:17:58 crc kubenswrapper[4950]: I0318 21:17:58.933870 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-zt89n_d304aae2-4025-4dea-980b-2f3b8c656a8e/ssh-known-hosts-edpm-deployment/0.log" Mar 18 21:17:59 crc kubenswrapper[4950]: I0318 21:17:59.121403 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_3c1bc0a9-e172-4749-b998-710b5a306618/tempest-tests-tempest-tests-runner/0.log" Mar 18 21:17:59 crc kubenswrapper[4950]: I0318 21:17:59.238389 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_c5edf1a8-44a7-4cd4-ae5f-f1a77338d35c/test-operator-logs-container/0.log" Mar 18 21:17:59 crc kubenswrapper[4950]: I0318 21:17:59.430361 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw_dd54300e-2b25-4bb8-b649-ac546072be73/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:18:00 crc kubenswrapper[4950]: I0318 21:18:00.165454 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564478-p9khw"] Mar 18 21:18:00 crc kubenswrapper[4950]: E0318 21:18:00.165880 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="046b4599-f9ba-48b1-b32d-7fdbce471f29" containerName="extract-content" Mar 18 21:18:00 crc kubenswrapper[4950]: I0318 21:18:00.165893 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="046b4599-f9ba-48b1-b32d-7fdbce471f29" containerName="extract-content" Mar 18 21:18:00 crc kubenswrapper[4950]: E0318 21:18:00.165934 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="046b4599-f9ba-48b1-b32d-7fdbce471f29" containerName="extract-utilities" Mar 18 21:18:00 crc kubenswrapper[4950]: I0318 21:18:00.165940 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="046b4599-f9ba-48b1-b32d-7fdbce471f29" containerName="extract-utilities" Mar 18 21:18:00 crc kubenswrapper[4950]: E0318 21:18:00.165954 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="046b4599-f9ba-48b1-b32d-7fdbce471f29" containerName="registry-server" Mar 18 21:18:00 crc kubenswrapper[4950]: I0318 21:18:00.165961 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="046b4599-f9ba-48b1-b32d-7fdbce471f29" containerName="registry-server" Mar 18 21:18:00 crc kubenswrapper[4950]: I0318 21:18:00.166126 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="046b4599-f9ba-48b1-b32d-7fdbce471f29" containerName="registry-server" Mar 18 21:18:00 crc kubenswrapper[4950]: I0318 21:18:00.166774 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564478-p9khw" Mar 18 21:18:00 crc kubenswrapper[4950]: I0318 21:18:00.170812 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 21:18:00 crc kubenswrapper[4950]: I0318 21:18:00.170992 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 21:18:00 crc kubenswrapper[4950]: I0318 21:18:00.171093 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 21:18:00 crc kubenswrapper[4950]: I0318 21:18:00.171797 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564478-p9khw"] Mar 18 21:18:00 crc kubenswrapper[4950]: I0318 21:18:00.173759 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgl8k\" (UniqueName: \"kubernetes.io/projected/2c0fcd06-ee4a-490b-bef9-6ea21dfa6987-kube-api-access-pgl8k\") pod \"auto-csr-approver-29564478-p9khw\" (UID: \"2c0fcd06-ee4a-490b-bef9-6ea21dfa6987\") " pod="openshift-infra/auto-csr-approver-29564478-p9khw" Mar 18 21:18:00 crc kubenswrapper[4950]: I0318 21:18:00.275865 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgl8k\" (UniqueName: \"kubernetes.io/projected/2c0fcd06-ee4a-490b-bef9-6ea21dfa6987-kube-api-access-pgl8k\") pod \"auto-csr-approver-29564478-p9khw\" (UID: \"2c0fcd06-ee4a-490b-bef9-6ea21dfa6987\") " pod="openshift-infra/auto-csr-approver-29564478-p9khw" Mar 18 21:18:00 crc kubenswrapper[4950]: I0318 21:18:00.310555 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgl8k\" (UniqueName: \"kubernetes.io/projected/2c0fcd06-ee4a-490b-bef9-6ea21dfa6987-kube-api-access-pgl8k\") pod \"auto-csr-approver-29564478-p9khw\" (UID: \"2c0fcd06-ee4a-490b-bef9-6ea21dfa6987\") " pod="openshift-infra/auto-csr-approver-29564478-p9khw" Mar 18 21:18:00 crc kubenswrapper[4950]: I0318 21:18:00.537822 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564478-p9khw" Mar 18 21:18:01 crc kubenswrapper[4950]: I0318 21:18:01.068535 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564478-p9khw"] Mar 18 21:18:01 crc kubenswrapper[4950]: I0318 21:18:01.077638 4950 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 21:18:01 crc kubenswrapper[4950]: I0318 21:18:01.360729 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564478-p9khw" event={"ID":"2c0fcd06-ee4a-490b-bef9-6ea21dfa6987","Type":"ContainerStarted","Data":"f92b096922511143a549ca1f8c5bd2610c4b03b47511c79af24221796ef75a13"} Mar 18 21:18:03 crc kubenswrapper[4950]: I0318 21:18:03.403235 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564478-p9khw" event={"ID":"2c0fcd06-ee4a-490b-bef9-6ea21dfa6987","Type":"ContainerStarted","Data":"e0b3baa00afead2ce35133363d84b3f066f738c0bb6e35a6e94da564aee28c9a"} Mar 18 21:18:03 crc kubenswrapper[4950]: I0318 21:18:03.428264 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29564478-p9khw" podStartSLOduration=2.171502767 podStartE2EDuration="3.428244975s" podCreationTimestamp="2026-03-18 21:18:00 +0000 UTC" firstStartedPulling="2026-03-18 21:18:01.077405357 +0000 UTC m=+4294.318247225" lastFinishedPulling="2026-03-18 21:18:02.334147565 +0000 UTC m=+4295.574989433" observedRunningTime="2026-03-18 21:18:03.417845882 +0000 UTC m=+4296.658687760" watchObservedRunningTime="2026-03-18 21:18:03.428244975 +0000 UTC m=+4296.669086843" Mar 18 21:18:05 crc kubenswrapper[4950]: I0318 21:18:05.445878 4950 generic.go:334] "Generic (PLEG): container finished" podID="2c0fcd06-ee4a-490b-bef9-6ea21dfa6987" containerID="e0b3baa00afead2ce35133363d84b3f066f738c0bb6e35a6e94da564aee28c9a" exitCode=0 Mar 18 21:18:05 crc kubenswrapper[4950]: I0318 21:18:05.446174 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564478-p9khw" event={"ID":"2c0fcd06-ee4a-490b-bef9-6ea21dfa6987","Type":"ContainerDied","Data":"e0b3baa00afead2ce35133363d84b3f066f738c0bb6e35a6e94da564aee28c9a"} Mar 18 21:18:06 crc kubenswrapper[4950]: I0318 21:18:06.863143 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564478-p9khw" Mar 18 21:18:06 crc kubenswrapper[4950]: I0318 21:18:06.895290 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgl8k\" (UniqueName: \"kubernetes.io/projected/2c0fcd06-ee4a-490b-bef9-6ea21dfa6987-kube-api-access-pgl8k\") pod \"2c0fcd06-ee4a-490b-bef9-6ea21dfa6987\" (UID: \"2c0fcd06-ee4a-490b-bef9-6ea21dfa6987\") " Mar 18 21:18:06 crc kubenswrapper[4950]: I0318 21:18:06.903643 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c0fcd06-ee4a-490b-bef9-6ea21dfa6987-kube-api-access-pgl8k" (OuterVolumeSpecName: "kube-api-access-pgl8k") pod "2c0fcd06-ee4a-490b-bef9-6ea21dfa6987" (UID: "2c0fcd06-ee4a-490b-bef9-6ea21dfa6987"). InnerVolumeSpecName "kube-api-access-pgl8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:18:07 crc kubenswrapper[4950]: I0318 21:18:07.001551 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgl8k\" (UniqueName: \"kubernetes.io/projected/2c0fcd06-ee4a-490b-bef9-6ea21dfa6987-kube-api-access-pgl8k\") on node \"crc\" DevicePath \"\"" Mar 18 21:18:07 crc kubenswrapper[4950]: I0318 21:18:07.495597 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564478-p9khw" Mar 18 21:18:07 crc kubenswrapper[4950]: I0318 21:18:07.510486 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564478-p9khw" event={"ID":"2c0fcd06-ee4a-490b-bef9-6ea21dfa6987","Type":"ContainerDied","Data":"f92b096922511143a549ca1f8c5bd2610c4b03b47511c79af24221796ef75a13"} Mar 18 21:18:07 crc kubenswrapper[4950]: I0318 21:18:07.510526 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f92b096922511143a549ca1f8c5bd2610c4b03b47511c79af24221796ef75a13" Mar 18 21:18:07 crc kubenswrapper[4950]: I0318 21:18:07.540201 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564472-fjh27"] Mar 18 21:18:07 crc kubenswrapper[4950]: I0318 21:18:07.548576 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564472-fjh27"] Mar 18 21:18:09 crc kubenswrapper[4950]: I0318 21:18:09.497514 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdd5815f-30c3-40b6-89fe-9fa8d7707928" path="/var/lib/kubelet/pods/cdd5815f-30c3-40b6-89fe-9fa8d7707928/volumes" Mar 18 21:18:12 crc kubenswrapper[4950]: I0318 21:18:12.077567 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_18eb6f19-11f0-4023-93ac-fc03ecdb9486/memcached/0.log" Mar 18 21:18:32 crc kubenswrapper[4950]: I0318 21:18:32.840373 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm_0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d/util/0.log" Mar 18 21:18:33 crc kubenswrapper[4950]: I0318 21:18:33.099245 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm_0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d/pull/0.log" Mar 18 21:18:33 crc kubenswrapper[4950]: I0318 21:18:33.105739 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm_0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d/pull/0.log" Mar 18 21:18:33 crc kubenswrapper[4950]: I0318 21:18:33.109794 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm_0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d/util/0.log" Mar 18 21:18:33 crc kubenswrapper[4950]: I0318 21:18:33.304246 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm_0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d/util/0.log" Mar 18 21:18:33 crc kubenswrapper[4950]: I0318 21:18:33.317827 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm_0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d/extract/0.log" Mar 18 21:18:33 crc kubenswrapper[4950]: I0318 21:18:33.382899 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm_0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d/pull/0.log" Mar 18 21:18:33 crc kubenswrapper[4950]: I0318 21:18:33.609056 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-59bc569d95-l2xgb_815a0db2-1d7b-4577-b1bf-cf3114f2a2a3/manager/0.log" Mar 18 21:18:33 crc kubenswrapper[4950]: I0318 21:18:33.950059 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-588d4d986b-h9vll_8081f4de-9137-49d2-a66d-bbf265033c5c/manager/0.log" Mar 18 21:18:34 crc kubenswrapper[4950]: I0318 21:18:34.249276 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-79df6bcc97-rgb5v_bc63ecc0-0c62-4607-8ce4-8d143f0c2ce6/manager/0.log" Mar 18 21:18:34 crc kubenswrapper[4950]: I0318 21:18:34.334694 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-67dd5f86f5-6jt4g_9046fce8-e07a-4cc6-9577-4ab38ecac0f3/manager/0.log" Mar 18 21:18:34 crc kubenswrapper[4950]: I0318 21:18:34.566097 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-8464cc45fb-5v45z_926aae26-bb2f-4cac-af12-abbd3585a90a/manager/0.log" Mar 18 21:18:34 crc kubenswrapper[4950]: I0318 21:18:34.949212 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-8d58dc466-xlx8b_19c17169-bafd-4a5d-b255-0c50248f3e1c/manager/0.log" Mar 18 21:18:35 crc kubenswrapper[4950]: I0318 21:18:35.031768 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6f787dddc9-kckr9_0277fe72-c2e8-4f07-ae64-bb1232a98518/manager/0.log" Mar 18 21:18:35 crc kubenswrapper[4950]: I0318 21:18:35.093782 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7b9c774f96-6cg5j_df0f2790-244c-4b4f-b3e9-d82e69bb9044/manager/0.log" Mar 18 21:18:35 crc kubenswrapper[4950]: I0318 21:18:35.257463 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-768b96df4c-dls6n_516ce0ee-83c3-40a3-bb00-2a88891a5791/manager/0.log" Mar 18 21:18:35 crc kubenswrapper[4950]: I0318 21:18:35.344630 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-55f864c847-ftljz_edc5286e-aa7d-4ee6-a8f7-3f5cad640484/manager/0.log" Mar 18 21:18:35 crc kubenswrapper[4950]: I0318 21:18:35.527651 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67ccfc9778-5h94g_80caf1bd-ddfd-498d-9b80-ea8cd37a3c72/manager/0.log" Mar 18 21:18:35 crc kubenswrapper[4950]: I0318 21:18:35.611639 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-767865f676-qxz4b_ec5a5aa1-65cc-4bb2-8644-fa0c7bb44da4/manager/0.log" Mar 18 21:18:35 crc kubenswrapper[4950]: I0318 21:18:35.877691 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5d488d59fb-bsv4c_608c505a-eb24-4cd9-aa6d-e07d4d8edc0c/manager/0.log" Mar 18 21:18:35 crc kubenswrapper[4950]: I0318 21:18:35.899576 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5b9f45d989-pvw7d_29d3abf6-9968-4603-9c38-d3c1809b7275/manager/0.log" Mar 18 21:18:36 crc kubenswrapper[4950]: I0318 21:18:36.118808 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-74c4796899rqqwt_5a502f8b-719d-4f52-812d-c559a8772461/manager/0.log" Mar 18 21:18:36 crc kubenswrapper[4950]: I0318 21:18:36.259610 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-b85c4d696-w5ws4_040697e2-a9b5-4ec0-be8d-f935227c4a0a/operator/0.log" Mar 18 21:18:36 crc kubenswrapper[4950]: I0318 21:18:36.609853 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-29kf5_e6006ce5-8d5e-40d2-b9cb-a1a7f501269f/registry-server/0.log" Mar 18 21:18:36 crc kubenswrapper[4950]: I0318 21:18:36.878553 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-884679f54-d6t88_f216b689-dfc9-4042-92f3-71014b3a2698/manager/0.log" Mar 18 21:18:37 crc kubenswrapper[4950]: I0318 21:18:37.023333 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5784578c99-zcjsh_6b04874c-d0d9-44e7-8047-b0cc2ba6bdfc/manager/0.log" Mar 18 21:18:37 crc kubenswrapper[4950]: I0318 21:18:37.232975 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-7vkgn_283bb5c8-8799-4b57-a3c0-3a8b28f3c2d6/operator/0.log" Mar 18 21:18:37 crc kubenswrapper[4950]: I0318 21:18:37.399169 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-c674c5965-tkbrk_598a0d70-d050-4d4c-8ef5-e274784e5c92/manager/0.log" Mar 18 21:18:37 crc kubenswrapper[4950]: I0318 21:18:37.556711 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-86bd8996f6-fxpwr_151d2334-a121-46f8-947b-3b17a21abcaf/manager/0.log" Mar 18 21:18:37 crc kubenswrapper[4950]: I0318 21:18:37.743183 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-d6b694c5-8k97x_b8e294f2-ac3e-4dbc-b2eb-cd4d2b947789/manager/0.log" Mar 18 21:18:37 crc kubenswrapper[4950]: I0318 21:18:37.779548 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5c5cb9c4d7-wnrmf_e4b560c4-a26f-4acf-bdc3-bb9cabbbd982/manager/0.log" Mar 18 21:18:37 crc kubenswrapper[4950]: I0318 21:18:37.822256 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6c4d75f7f9-v6qqx_e3196503-22ca-4f3a-ac09-f37171f08d37/manager/0.log" Mar 18 21:18:59 crc kubenswrapper[4950]: I0318 21:18:59.304963 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-sxcm9_8596ab2e-852c-4d94-a8ce-e02c823a76b1/control-plane-machine-set-operator/0.log" Mar 18 21:18:59 crc kubenswrapper[4950]: I0318 21:18:59.411603 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-kjb6z_bfd29e9b-fbda-41d0-8300-a0d9337e0e6d/kube-rbac-proxy/0.log" Mar 18 21:18:59 crc kubenswrapper[4950]: I0318 21:18:59.498904 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-kjb6z_bfd29e9b-fbda-41d0-8300-a0d9337e0e6d/machine-api-operator/0.log" Mar 18 21:19:01 crc kubenswrapper[4950]: I0318 21:19:01.612983 4950 scope.go:117] "RemoveContainer" containerID="8370b7e5b41bf9fbf5b4731d3ba51b27a9d591d7f40c6a149850c86d5f306c50" Mar 18 21:19:13 crc kubenswrapper[4950]: I0318 21:19:13.372912 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-9pk5n_5be9c32f-9b12-4d42-8dd1-3d08455965cc/cert-manager-controller/0.log" Mar 18 21:19:13 crc kubenswrapper[4950]: I0318 21:19:13.667581 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-njln9_f1979115-0bc1-4a2a-9973-72fc24aa570e/cert-manager-cainjector/0.log" Mar 18 21:19:13 crc kubenswrapper[4950]: I0318 21:19:13.805818 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-9gqqw_7eb3a8a5-cac1-4a6f-8dba-2cc19867e55c/cert-manager-webhook/0.log" Mar 18 21:19:28 crc kubenswrapper[4950]: I0318 21:19:28.387797 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-86f58fcf4-jnvkz_0db507f0-dd6f-405c-ae17-4a9e9553cc50/nmstate-console-plugin/0.log" Mar 18 21:19:28 crc kubenswrapper[4950]: I0318 21:19:28.549852 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-6x6mb_61d14d6f-1408-4309-90c9-2971b45d85dc/nmstate-handler/0.log" Mar 18 21:19:28 crc kubenswrapper[4950]: I0318 21:19:28.617480 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-gqrnm_e1b75269-989a-489f-843b-94004e8cdf62/kube-rbac-proxy/0.log" Mar 18 21:19:28 crc kubenswrapper[4950]: I0318 21:19:28.658021 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-gqrnm_e1b75269-989a-489f-843b-94004e8cdf62/nmstate-metrics/0.log" Mar 18 21:19:28 crc kubenswrapper[4950]: I0318 21:19:28.761533 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-796d4cfff4-phx52_de65b778-0bca-479c-826c-e7d935a294c4/nmstate-operator/0.log" Mar 18 21:19:28 crc kubenswrapper[4950]: I0318 21:19:28.850725 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f558f5558-64n89_452b41f3-a30b-4aec-a879-aff3872199d3/nmstate-webhook/0.log" Mar 18 21:19:33 crc kubenswrapper[4950]: I0318 21:19:33.836386 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 21:19:33 crc kubenswrapper[4950]: I0318 21:19:33.837030 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 21:19:58 crc kubenswrapper[4950]: I0318 21:19:58.684767 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-mhk7w_db65cfec-9483-4641-8794-5fbd50c4bc92/kube-rbac-proxy/0.log" Mar 18 21:19:58 crc kubenswrapper[4950]: I0318 21:19:58.794098 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-mhk7w_db65cfec-9483-4641-8794-5fbd50c4bc92/controller/0.log" Mar 18 21:19:58 crc kubenswrapper[4950]: I0318 21:19:58.908487 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-frr-files/0.log" Mar 18 21:19:59 crc kubenswrapper[4950]: I0318 21:19:59.099263 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-reloader/0.log" Mar 18 21:19:59 crc kubenswrapper[4950]: I0318 21:19:59.103872 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-reloader/0.log" Mar 18 21:19:59 crc kubenswrapper[4950]: I0318 21:19:59.129339 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-frr-files/0.log" Mar 18 21:19:59 crc kubenswrapper[4950]: I0318 21:19:59.204289 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-metrics/0.log" Mar 18 21:19:59 crc kubenswrapper[4950]: I0318 21:19:59.378022 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-frr-files/0.log" Mar 18 21:19:59 crc kubenswrapper[4950]: I0318 21:19:59.395616 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-reloader/0.log" Mar 18 21:19:59 crc kubenswrapper[4950]: I0318 21:19:59.440010 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-metrics/0.log" Mar 18 21:19:59 crc kubenswrapper[4950]: I0318 21:19:59.463050 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-metrics/0.log" Mar 18 21:19:59 crc kubenswrapper[4950]: I0318 21:19:59.617016 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-reloader/0.log" Mar 18 21:19:59 crc kubenswrapper[4950]: I0318 21:19:59.636136 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-metrics/0.log" Mar 18 21:19:59 crc kubenswrapper[4950]: I0318 21:19:59.668744 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-frr-files/0.log" Mar 18 21:19:59 crc kubenswrapper[4950]: I0318 21:19:59.712716 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/controller/0.log" Mar 18 21:20:00 crc kubenswrapper[4950]: I0318 21:20:00.153753 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564480-99n7f"] Mar 18 21:20:00 crc kubenswrapper[4950]: E0318 21:20:00.154194 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c0fcd06-ee4a-490b-bef9-6ea21dfa6987" containerName="oc" Mar 18 21:20:00 crc kubenswrapper[4950]: I0318 21:20:00.154215 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c0fcd06-ee4a-490b-bef9-6ea21dfa6987" containerName="oc" Mar 18 21:20:00 crc kubenswrapper[4950]: I0318 21:20:00.154476 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c0fcd06-ee4a-490b-bef9-6ea21dfa6987" containerName="oc" Mar 18 21:20:00 crc kubenswrapper[4950]: I0318 21:20:00.155202 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564480-99n7f" Mar 18 21:20:00 crc kubenswrapper[4950]: I0318 21:20:00.158199 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 21:20:00 crc kubenswrapper[4950]: I0318 21:20:00.158287 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 21:20:00 crc kubenswrapper[4950]: I0318 21:20:00.158494 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 21:20:00 crc kubenswrapper[4950]: I0318 21:20:00.246530 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564480-99n7f"] Mar 18 21:20:00 crc kubenswrapper[4950]: I0318 21:20:00.303474 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2mw4\" (UniqueName: \"kubernetes.io/projected/abf0abcc-8ac6-4243-82d0-f9d802143e9d-kube-api-access-j2mw4\") pod \"auto-csr-approver-29564480-99n7f\" (UID: \"abf0abcc-8ac6-4243-82d0-f9d802143e9d\") " pod="openshift-infra/auto-csr-approver-29564480-99n7f" Mar 18 21:20:00 crc kubenswrapper[4950]: I0318 21:20:00.404982 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2mw4\" (UniqueName: \"kubernetes.io/projected/abf0abcc-8ac6-4243-82d0-f9d802143e9d-kube-api-access-j2mw4\") pod \"auto-csr-approver-29564480-99n7f\" (UID: \"abf0abcc-8ac6-4243-82d0-f9d802143e9d\") " pod="openshift-infra/auto-csr-approver-29564480-99n7f" Mar 18 21:20:00 crc kubenswrapper[4950]: I0318 21:20:00.410642 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/kube-rbac-proxy-frr/0.log" Mar 18 21:20:00 crc kubenswrapper[4950]: I0318 21:20:00.430117 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/frr-metrics/0.log" Mar 18 21:20:00 crc kubenswrapper[4950]: I0318 21:20:00.442128 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2mw4\" (UniqueName: \"kubernetes.io/projected/abf0abcc-8ac6-4243-82d0-f9d802143e9d-kube-api-access-j2mw4\") pod \"auto-csr-approver-29564480-99n7f\" (UID: \"abf0abcc-8ac6-4243-82d0-f9d802143e9d\") " pod="openshift-infra/auto-csr-approver-29564480-99n7f" Mar 18 21:20:00 crc kubenswrapper[4950]: I0318 21:20:00.482549 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564480-99n7f" Mar 18 21:20:00 crc kubenswrapper[4950]: I0318 21:20:00.508905 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/kube-rbac-proxy/0.log" Mar 18 21:20:00 crc kubenswrapper[4950]: I0318 21:20:00.755983 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/reloader/0.log" Mar 18 21:20:00 crc kubenswrapper[4950]: I0318 21:20:00.909557 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-nqppx_b4c0afa4-f131-44df-b2f1-9908b3b578e6/frr-k8s-webhook-server/0.log" Mar 18 21:20:01 crc kubenswrapper[4950]: I0318 21:20:01.083224 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-648849f9d4-9zr5p_871672d1-30f1-4221-a434-2a8d60173f46/manager/0.log" Mar 18 21:20:01 crc kubenswrapper[4950]: I0318 21:20:01.155869 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564480-99n7f"] Mar 18 21:20:01 crc kubenswrapper[4950]: I0318 21:20:01.308533 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-bf6cfb5cf-tllp8_210bdf50-0f58-4342-af2c-e0df41ac3e91/webhook-server/0.log" Mar 18 21:20:01 crc kubenswrapper[4950]: I0318 21:20:01.415869 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6k4zn_6f8ab895-b55a-43e6-a6fd-71f4ac42a50a/kube-rbac-proxy/0.log" Mar 18 21:20:01 crc kubenswrapper[4950]: I0318 21:20:01.472845 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564480-99n7f" event={"ID":"abf0abcc-8ac6-4243-82d0-f9d802143e9d","Type":"ContainerStarted","Data":"74f7debf060f22d20140b482e2054eb3ef30f1d3116cfaaadd04e63e3479351a"} Mar 18 21:20:02 crc kubenswrapper[4950]: I0318 21:20:02.098068 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6k4zn_6f8ab895-b55a-43e6-a6fd-71f4ac42a50a/speaker/0.log" Mar 18 21:20:02 crc kubenswrapper[4950]: I0318 21:20:02.429388 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/frr/0.log" Mar 18 21:20:03 crc kubenswrapper[4950]: I0318 21:20:03.497231 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564480-99n7f" event={"ID":"abf0abcc-8ac6-4243-82d0-f9d802143e9d","Type":"ContainerStarted","Data":"6a89bf9cb437ebebe4c9dd52ea94bd2c063db1326deac67c8addde33c042a41c"} Mar 18 21:20:03 crc kubenswrapper[4950]: I0318 21:20:03.523860 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29564480-99n7f" podStartSLOduration=2.195086814 podStartE2EDuration="3.52384211s" podCreationTimestamp="2026-03-18 21:20:00 +0000 UTC" firstStartedPulling="2026-03-18 21:20:01.142899352 +0000 UTC m=+4414.383741220" lastFinishedPulling="2026-03-18 21:20:02.471654648 +0000 UTC m=+4415.712496516" observedRunningTime="2026-03-18 21:20:03.522643759 +0000 UTC m=+4416.763485627" watchObservedRunningTime="2026-03-18 21:20:03.52384211 +0000 UTC m=+4416.764683978" Mar 18 21:20:03 crc kubenswrapper[4950]: I0318 21:20:03.835820 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 21:20:03 crc kubenswrapper[4950]: I0318 21:20:03.835873 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 21:20:04 crc kubenswrapper[4950]: I0318 21:20:04.527260 4950 generic.go:334] "Generic (PLEG): container finished" podID="abf0abcc-8ac6-4243-82d0-f9d802143e9d" containerID="6a89bf9cb437ebebe4c9dd52ea94bd2c063db1326deac67c8addde33c042a41c" exitCode=0 Mar 18 21:20:04 crc kubenswrapper[4950]: I0318 21:20:04.527548 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564480-99n7f" event={"ID":"abf0abcc-8ac6-4243-82d0-f9d802143e9d","Type":"ContainerDied","Data":"6a89bf9cb437ebebe4c9dd52ea94bd2c063db1326deac67c8addde33c042a41c"} Mar 18 21:20:05 crc kubenswrapper[4950]: I0318 21:20:05.930312 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564480-99n7f" Mar 18 21:20:06 crc kubenswrapper[4950]: I0318 21:20:06.054105 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2mw4\" (UniqueName: \"kubernetes.io/projected/abf0abcc-8ac6-4243-82d0-f9d802143e9d-kube-api-access-j2mw4\") pod \"abf0abcc-8ac6-4243-82d0-f9d802143e9d\" (UID: \"abf0abcc-8ac6-4243-82d0-f9d802143e9d\") " Mar 18 21:20:06 crc kubenswrapper[4950]: I0318 21:20:06.067616 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abf0abcc-8ac6-4243-82d0-f9d802143e9d-kube-api-access-j2mw4" (OuterVolumeSpecName: "kube-api-access-j2mw4") pod "abf0abcc-8ac6-4243-82d0-f9d802143e9d" (UID: "abf0abcc-8ac6-4243-82d0-f9d802143e9d"). InnerVolumeSpecName "kube-api-access-j2mw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:20:06 crc kubenswrapper[4950]: I0318 21:20:06.156232 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2mw4\" (UniqueName: \"kubernetes.io/projected/abf0abcc-8ac6-4243-82d0-f9d802143e9d-kube-api-access-j2mw4\") on node \"crc\" DevicePath \"\"" Mar 18 21:20:06 crc kubenswrapper[4950]: I0318 21:20:06.328160 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nsf7r"] Mar 18 21:20:06 crc kubenswrapper[4950]: E0318 21:20:06.328612 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abf0abcc-8ac6-4243-82d0-f9d802143e9d" containerName="oc" Mar 18 21:20:06 crc kubenswrapper[4950]: I0318 21:20:06.328629 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="abf0abcc-8ac6-4243-82d0-f9d802143e9d" containerName="oc" Mar 18 21:20:06 crc kubenswrapper[4950]: I0318 21:20:06.328886 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="abf0abcc-8ac6-4243-82d0-f9d802143e9d" containerName="oc" Mar 18 21:20:06 crc kubenswrapper[4950]: I0318 21:20:06.330180 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nsf7r" Mar 18 21:20:06 crc kubenswrapper[4950]: I0318 21:20:06.346640 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nsf7r"] Mar 18 21:20:06 crc kubenswrapper[4950]: I0318 21:20:06.462064 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d463a961-08cd-4d60-8e0e-ade812e3f63f-catalog-content\") pod \"redhat-operators-nsf7r\" (UID: \"d463a961-08cd-4d60-8e0e-ade812e3f63f\") " pod="openshift-marketplace/redhat-operators-nsf7r" Mar 18 21:20:06 crc kubenswrapper[4950]: I0318 21:20:06.462443 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzwst\" (UniqueName: \"kubernetes.io/projected/d463a961-08cd-4d60-8e0e-ade812e3f63f-kube-api-access-jzwst\") pod \"redhat-operators-nsf7r\" (UID: \"d463a961-08cd-4d60-8e0e-ade812e3f63f\") " pod="openshift-marketplace/redhat-operators-nsf7r" Mar 18 21:20:06 crc kubenswrapper[4950]: I0318 21:20:06.462617 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d463a961-08cd-4d60-8e0e-ade812e3f63f-utilities\") pod \"redhat-operators-nsf7r\" (UID: \"d463a961-08cd-4d60-8e0e-ade812e3f63f\") " pod="openshift-marketplace/redhat-operators-nsf7r" Mar 18 21:20:06 crc kubenswrapper[4950]: I0318 21:20:06.544798 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564480-99n7f" event={"ID":"abf0abcc-8ac6-4243-82d0-f9d802143e9d","Type":"ContainerDied","Data":"74f7debf060f22d20140b482e2054eb3ef30f1d3116cfaaadd04e63e3479351a"} Mar 18 21:20:06 crc kubenswrapper[4950]: I0318 21:20:06.544834 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74f7debf060f22d20140b482e2054eb3ef30f1d3116cfaaadd04e63e3479351a" Mar 18 21:20:06 crc kubenswrapper[4950]: I0318 21:20:06.544857 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564480-99n7f" Mar 18 21:20:06 crc kubenswrapper[4950]: I0318 21:20:06.566473 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzwst\" (UniqueName: \"kubernetes.io/projected/d463a961-08cd-4d60-8e0e-ade812e3f63f-kube-api-access-jzwst\") pod \"redhat-operators-nsf7r\" (UID: \"d463a961-08cd-4d60-8e0e-ade812e3f63f\") " pod="openshift-marketplace/redhat-operators-nsf7r" Mar 18 21:20:06 crc kubenswrapper[4950]: I0318 21:20:06.566693 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d463a961-08cd-4d60-8e0e-ade812e3f63f-utilities\") pod \"redhat-operators-nsf7r\" (UID: \"d463a961-08cd-4d60-8e0e-ade812e3f63f\") " pod="openshift-marketplace/redhat-operators-nsf7r" Mar 18 21:20:06 crc kubenswrapper[4950]: I0318 21:20:06.566762 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d463a961-08cd-4d60-8e0e-ade812e3f63f-catalog-content\") pod \"redhat-operators-nsf7r\" (UID: \"d463a961-08cd-4d60-8e0e-ade812e3f63f\") " pod="openshift-marketplace/redhat-operators-nsf7r" Mar 18 21:20:06 crc kubenswrapper[4950]: I0318 21:20:06.567747 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d463a961-08cd-4d60-8e0e-ade812e3f63f-utilities\") pod \"redhat-operators-nsf7r\" (UID: \"d463a961-08cd-4d60-8e0e-ade812e3f63f\") " pod="openshift-marketplace/redhat-operators-nsf7r" Mar 18 21:20:06 crc kubenswrapper[4950]: I0318 21:20:06.569154 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d463a961-08cd-4d60-8e0e-ade812e3f63f-catalog-content\") pod \"redhat-operators-nsf7r\" (UID: \"d463a961-08cd-4d60-8e0e-ade812e3f63f\") " pod="openshift-marketplace/redhat-operators-nsf7r" Mar 18 21:20:06 crc kubenswrapper[4950]: I0318 21:20:06.587266 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzwst\" (UniqueName: \"kubernetes.io/projected/d463a961-08cd-4d60-8e0e-ade812e3f63f-kube-api-access-jzwst\") pod \"redhat-operators-nsf7r\" (UID: \"d463a961-08cd-4d60-8e0e-ade812e3f63f\") " pod="openshift-marketplace/redhat-operators-nsf7r" Mar 18 21:20:06 crc kubenswrapper[4950]: I0318 21:20:06.654351 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564474-tbsqb"] Mar 18 21:20:06 crc kubenswrapper[4950]: I0318 21:20:06.666107 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564474-tbsqb"] Mar 18 21:20:06 crc kubenswrapper[4950]: I0318 21:20:06.679258 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nsf7r" Mar 18 21:20:07 crc kubenswrapper[4950]: I0318 21:20:07.139736 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nsf7r"] Mar 18 21:20:07 crc kubenswrapper[4950]: I0318 21:20:07.492977 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d38b254-beb1-45fa-b20d-2b9c082c00b4" path="/var/lib/kubelet/pods/6d38b254-beb1-45fa-b20d-2b9c082c00b4/volumes" Mar 18 21:20:07 crc kubenswrapper[4950]: I0318 21:20:07.556543 4950 generic.go:334] "Generic (PLEG): container finished" podID="d463a961-08cd-4d60-8e0e-ade812e3f63f" containerID="02160664edb1414e67881e635f7eca40d092dcf3755c7d474ccd1aeb23680b50" exitCode=0 Mar 18 21:20:07 crc kubenswrapper[4950]: I0318 21:20:07.556589 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nsf7r" event={"ID":"d463a961-08cd-4d60-8e0e-ade812e3f63f","Type":"ContainerDied","Data":"02160664edb1414e67881e635f7eca40d092dcf3755c7d474ccd1aeb23680b50"} Mar 18 21:20:07 crc kubenswrapper[4950]: I0318 21:20:07.556615 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nsf7r" event={"ID":"d463a961-08cd-4d60-8e0e-ade812e3f63f","Type":"ContainerStarted","Data":"c94819be3229307d49595efbd2eced87073380ef3a283cb0e5b11ac86806fda9"} Mar 18 21:20:09 crc kubenswrapper[4950]: I0318 21:20:09.576124 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nsf7r" event={"ID":"d463a961-08cd-4d60-8e0e-ade812e3f63f","Type":"ContainerStarted","Data":"66ece35ac5af4f5c1caa681889b15421ba0a13905ff42f2dab42da1a7d96b841"} Mar 18 21:20:15 crc kubenswrapper[4950]: I0318 21:20:15.642592 4950 generic.go:334] "Generic (PLEG): container finished" podID="d463a961-08cd-4d60-8e0e-ade812e3f63f" containerID="66ece35ac5af4f5c1caa681889b15421ba0a13905ff42f2dab42da1a7d96b841" exitCode=0 Mar 18 21:20:15 crc kubenswrapper[4950]: I0318 21:20:15.642655 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nsf7r" event={"ID":"d463a961-08cd-4d60-8e0e-ade812e3f63f","Type":"ContainerDied","Data":"66ece35ac5af4f5c1caa681889b15421ba0a13905ff42f2dab42da1a7d96b841"} Mar 18 21:20:16 crc kubenswrapper[4950]: I0318 21:20:16.654255 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nsf7r" event={"ID":"d463a961-08cd-4d60-8e0e-ade812e3f63f","Type":"ContainerStarted","Data":"c165a5d4a88734dbcb76e7c622acea587cc435bafe008af7c669625f4c7cc4ce"} Mar 18 21:20:16 crc kubenswrapper[4950]: I0318 21:20:16.679471 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nsf7r" Mar 18 21:20:16 crc kubenswrapper[4950]: I0318 21:20:16.679527 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nsf7r" Mar 18 21:20:16 crc kubenswrapper[4950]: I0318 21:20:16.680806 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nsf7r" podStartSLOduration=2.201004266 podStartE2EDuration="10.680794844s" podCreationTimestamp="2026-03-18 21:20:06 +0000 UTC" firstStartedPulling="2026-03-18 21:20:07.55886471 +0000 UTC m=+4420.799706578" lastFinishedPulling="2026-03-18 21:20:16.038655298 +0000 UTC m=+4429.279497156" observedRunningTime="2026-03-18 21:20:16.678476853 +0000 UTC m=+4429.919318721" watchObservedRunningTime="2026-03-18 21:20:16.680794844 +0000 UTC m=+4429.921636712" Mar 18 21:20:17 crc kubenswrapper[4950]: I0318 21:20:17.727125 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nsf7r" podUID="d463a961-08cd-4d60-8e0e-ade812e3f63f" containerName="registry-server" probeResult="failure" output=< Mar 18 21:20:17 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 21:20:17 crc kubenswrapper[4950]: > Mar 18 21:20:18 crc kubenswrapper[4950]: I0318 21:20:18.357984 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k_534d37ef-a537-4ea3-832e-4609580d9f47/util/0.log" Mar 18 21:20:18 crc kubenswrapper[4950]: I0318 21:20:18.709762 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k_534d37ef-a537-4ea3-832e-4609580d9f47/util/0.log" Mar 18 21:20:18 crc kubenswrapper[4950]: I0318 21:20:18.791571 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k_534d37ef-a537-4ea3-832e-4609580d9f47/pull/0.log" Mar 18 21:20:18 crc kubenswrapper[4950]: I0318 21:20:18.821124 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k_534d37ef-a537-4ea3-832e-4609580d9f47/pull/0.log" Mar 18 21:20:19 crc kubenswrapper[4950]: I0318 21:20:19.119909 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k_534d37ef-a537-4ea3-832e-4609580d9f47/util/0.log" Mar 18 21:20:19 crc kubenswrapper[4950]: I0318 21:20:19.122310 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k_534d37ef-a537-4ea3-832e-4609580d9f47/pull/0.log" Mar 18 21:20:19 crc kubenswrapper[4950]: I0318 21:20:19.555462 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k_534d37ef-a537-4ea3-832e-4609580d9f47/extract/0.log" Mar 18 21:20:19 crc kubenswrapper[4950]: I0318 21:20:19.637853 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk_c955c23a-4fa9-4745-bb00-05a54ae897f4/util/0.log" Mar 18 21:20:19 crc kubenswrapper[4950]: I0318 21:20:19.847972 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk_c955c23a-4fa9-4745-bb00-05a54ae897f4/pull/0.log" Mar 18 21:20:19 crc kubenswrapper[4950]: I0318 21:20:19.868123 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk_c955c23a-4fa9-4745-bb00-05a54ae897f4/util/0.log" Mar 18 21:20:19 crc kubenswrapper[4950]: I0318 21:20:19.895179 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk_c955c23a-4fa9-4745-bb00-05a54ae897f4/pull/0.log" Mar 18 21:20:20 crc kubenswrapper[4950]: I0318 21:20:20.109482 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk_c955c23a-4fa9-4745-bb00-05a54ae897f4/pull/0.log" Mar 18 21:20:20 crc kubenswrapper[4950]: I0318 21:20:20.124882 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk_c955c23a-4fa9-4745-bb00-05a54ae897f4/extract/0.log" Mar 18 21:20:20 crc kubenswrapper[4950]: I0318 21:20:20.145203 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk_c955c23a-4fa9-4745-bb00-05a54ae897f4/util/0.log" Mar 18 21:20:20 crc kubenswrapper[4950]: I0318 21:20:20.316796 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xv6kx_df5f1fd5-ec77-45be-84de-d0d8a1568a73/extract-utilities/0.log" Mar 18 21:20:21 crc kubenswrapper[4950]: I0318 21:20:21.095040 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xv6kx_df5f1fd5-ec77-45be-84de-d0d8a1568a73/extract-utilities/0.log" Mar 18 21:20:21 crc kubenswrapper[4950]: I0318 21:20:21.149888 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xv6kx_df5f1fd5-ec77-45be-84de-d0d8a1568a73/extract-content/0.log" Mar 18 21:20:21 crc kubenswrapper[4950]: I0318 21:20:21.198486 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xv6kx_df5f1fd5-ec77-45be-84de-d0d8a1568a73/extract-content/0.log" Mar 18 21:20:21 crc kubenswrapper[4950]: I0318 21:20:21.448000 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xv6kx_df5f1fd5-ec77-45be-84de-d0d8a1568a73/extract-utilities/0.log" Mar 18 21:20:21 crc kubenswrapper[4950]: I0318 21:20:21.562086 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xv6kx_df5f1fd5-ec77-45be-84de-d0d8a1568a73/extract-content/0.log" Mar 18 21:20:21 crc kubenswrapper[4950]: I0318 21:20:21.841292 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zljc4_40ff6ac2-e78c-4501-9fdd-8d612fd6d29d/extract-utilities/0.log" Mar 18 21:20:21 crc kubenswrapper[4950]: I0318 21:20:21.994859 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xv6kx_df5f1fd5-ec77-45be-84de-d0d8a1568a73/registry-server/0.log" Mar 18 21:20:22 crc kubenswrapper[4950]: I0318 21:20:22.011898 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zljc4_40ff6ac2-e78c-4501-9fdd-8d612fd6d29d/extract-content/0.log" Mar 18 21:20:22 crc kubenswrapper[4950]: I0318 21:20:22.017396 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zljc4_40ff6ac2-e78c-4501-9fdd-8d612fd6d29d/extract-utilities/0.log" Mar 18 21:20:22 crc kubenswrapper[4950]: I0318 21:20:22.115345 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zljc4_40ff6ac2-e78c-4501-9fdd-8d612fd6d29d/extract-content/0.log" Mar 18 21:20:22 crc kubenswrapper[4950]: I0318 21:20:22.250750 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zljc4_40ff6ac2-e78c-4501-9fdd-8d612fd6d29d/extract-content/0.log" Mar 18 21:20:22 crc kubenswrapper[4950]: I0318 21:20:22.255224 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zljc4_40ff6ac2-e78c-4501-9fdd-8d612fd6d29d/extract-utilities/0.log" Mar 18 21:20:22 crc kubenswrapper[4950]: I0318 21:20:22.598676 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-c6tfb_51b0360c-046e-4360-b240-fdfd8d857a4b/marketplace-operator/0.log" Mar 18 21:20:22 crc kubenswrapper[4950]: I0318 21:20:22.910271 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-shgnx_19ec358c-0c71-4b43-b13f-688f2103d895/extract-utilities/0.log" Mar 18 21:20:23 crc kubenswrapper[4950]: I0318 21:20:23.014201 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zljc4_40ff6ac2-e78c-4501-9fdd-8d612fd6d29d/registry-server/0.log" Mar 18 21:20:23 crc kubenswrapper[4950]: I0318 21:20:23.161348 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-shgnx_19ec358c-0c71-4b43-b13f-688f2103d895/extract-content/0.log" Mar 18 21:20:23 crc kubenswrapper[4950]: I0318 21:20:23.197572 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-shgnx_19ec358c-0c71-4b43-b13f-688f2103d895/extract-content/0.log" Mar 18 21:20:23 crc kubenswrapper[4950]: I0318 21:20:23.217496 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-shgnx_19ec358c-0c71-4b43-b13f-688f2103d895/extract-utilities/0.log" Mar 18 21:20:23 crc kubenswrapper[4950]: I0318 21:20:23.351718 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-shgnx_19ec358c-0c71-4b43-b13f-688f2103d895/extract-utilities/0.log" Mar 18 21:20:23 crc kubenswrapper[4950]: I0318 21:20:23.361085 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-shgnx_19ec358c-0c71-4b43-b13f-688f2103d895/extract-content/0.log" Mar 18 21:20:23 crc kubenswrapper[4950]: I0318 21:20:23.522709 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-shgnx_19ec358c-0c71-4b43-b13f-688f2103d895/registry-server/0.log" Mar 18 21:20:23 crc kubenswrapper[4950]: I0318 21:20:23.530341 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-42xf9_d72524a5-b7b7-416f-b9ba-9527f98e787c/extract-utilities/0.log" Mar 18 21:20:23 crc kubenswrapper[4950]: I0318 21:20:23.768870 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-42xf9_d72524a5-b7b7-416f-b9ba-9527f98e787c/extract-content/0.log" Mar 18 21:20:23 crc kubenswrapper[4950]: I0318 21:20:23.847934 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-42xf9_d72524a5-b7b7-416f-b9ba-9527f98e787c/extract-utilities/0.log" Mar 18 21:20:23 crc kubenswrapper[4950]: I0318 21:20:23.881212 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-42xf9_d72524a5-b7b7-416f-b9ba-9527f98e787c/extract-content/0.log" Mar 18 21:20:24 crc kubenswrapper[4950]: I0318 21:20:24.028151 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-42xf9_d72524a5-b7b7-416f-b9ba-9527f98e787c/extract-content/0.log" Mar 18 21:20:24 crc kubenswrapper[4950]: I0318 21:20:24.046432 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-42xf9_d72524a5-b7b7-416f-b9ba-9527f98e787c/extract-utilities/0.log" Mar 18 21:20:24 crc kubenswrapper[4950]: I0318 21:20:24.185005 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nsf7r_d463a961-08cd-4d60-8e0e-ade812e3f63f/extract-utilities/0.log" Mar 18 21:20:24 crc kubenswrapper[4950]: I0318 21:20:24.277749 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-42xf9_d72524a5-b7b7-416f-b9ba-9527f98e787c/registry-server/0.log" Mar 18 21:20:24 crc kubenswrapper[4950]: I0318 21:20:24.423538 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nsf7r_d463a961-08cd-4d60-8e0e-ade812e3f63f/extract-utilities/0.log" Mar 18 21:20:24 crc kubenswrapper[4950]: I0318 21:20:24.460828 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nsf7r_d463a961-08cd-4d60-8e0e-ade812e3f63f/extract-content/0.log" Mar 18 21:20:24 crc kubenswrapper[4950]: I0318 21:20:24.461836 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nsf7r_d463a961-08cd-4d60-8e0e-ade812e3f63f/extract-content/0.log" Mar 18 21:20:24 crc kubenswrapper[4950]: I0318 21:20:24.651846 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nsf7r_d463a961-08cd-4d60-8e0e-ade812e3f63f/extract-utilities/0.log" Mar 18 21:20:24 crc kubenswrapper[4950]: I0318 21:20:24.677307 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nsf7r_d463a961-08cd-4d60-8e0e-ade812e3f63f/extract-content/0.log" Mar 18 21:20:24 crc kubenswrapper[4950]: I0318 21:20:24.735261 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-nsf7r_d463a961-08cd-4d60-8e0e-ade812e3f63f/registry-server/0.log" Mar 18 21:20:27 crc kubenswrapper[4950]: I0318 21:20:27.731442 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nsf7r" podUID="d463a961-08cd-4d60-8e0e-ade812e3f63f" containerName="registry-server" probeResult="failure" output=< Mar 18 21:20:27 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 21:20:27 crc kubenswrapper[4950]: > Mar 18 21:20:33 crc kubenswrapper[4950]: I0318 21:20:33.836801 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 21:20:33 crc kubenswrapper[4950]: I0318 21:20:33.837452 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 21:20:33 crc kubenswrapper[4950]: I0318 21:20:33.837509 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 21:20:33 crc kubenswrapper[4950]: I0318 21:20:33.838701 4950 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6fb2941ff91c9d7625aec6265d6c6aecb9312056ca715d8dc848cdcf34657cae"} pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 21:20:33 crc kubenswrapper[4950]: I0318 21:20:33.838770 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" containerID="cri-o://6fb2941ff91c9d7625aec6265d6c6aecb9312056ca715d8dc848cdcf34657cae" gracePeriod=600 Mar 18 21:20:34 crc kubenswrapper[4950]: I0318 21:20:34.812339 4950 generic.go:334] "Generic (PLEG): container finished" podID="4048b439-3266-46e7-9de0-22377efacc46" containerID="6fb2941ff91c9d7625aec6265d6c6aecb9312056ca715d8dc848cdcf34657cae" exitCode=0 Mar 18 21:20:34 crc kubenswrapper[4950]: I0318 21:20:34.812821 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerDied","Data":"6fb2941ff91c9d7625aec6265d6c6aecb9312056ca715d8dc848cdcf34657cae"} Mar 18 21:20:34 crc kubenswrapper[4950]: I0318 21:20:34.812850 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerStarted","Data":"cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d"} Mar 18 21:20:34 crc kubenswrapper[4950]: I0318 21:20:34.812865 4950 scope.go:117] "RemoveContainer" containerID="8c4179a5faeb3f6f0d4e766df2e8dd9edca78a12bdbfd8122715b8b9cd78eaa9" Mar 18 21:20:37 crc kubenswrapper[4950]: I0318 21:20:37.736714 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nsf7r" podUID="d463a961-08cd-4d60-8e0e-ade812e3f63f" containerName="registry-server" probeResult="failure" output=< Mar 18 21:20:37 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 21:20:37 crc kubenswrapper[4950]: > Mar 18 21:20:46 crc kubenswrapper[4950]: I0318 21:20:46.752491 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nsf7r" Mar 18 21:20:47 crc kubenswrapper[4950]: I0318 21:20:47.294615 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nsf7r" Mar 18 21:20:47 crc kubenswrapper[4950]: I0318 21:20:47.344817 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nsf7r"] Mar 18 21:20:47 crc kubenswrapper[4950]: I0318 21:20:47.916492 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nsf7r" podUID="d463a961-08cd-4d60-8e0e-ade812e3f63f" containerName="registry-server" containerID="cri-o://c165a5d4a88734dbcb76e7c622acea587cc435bafe008af7c669625f4c7cc4ce" gracePeriod=2 Mar 18 21:20:48 crc kubenswrapper[4950]: I0318 21:20:48.827367 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nsf7r" Mar 18 21:20:48 crc kubenswrapper[4950]: I0318 21:20:48.927430 4950 generic.go:334] "Generic (PLEG): container finished" podID="d463a961-08cd-4d60-8e0e-ade812e3f63f" containerID="c165a5d4a88734dbcb76e7c622acea587cc435bafe008af7c669625f4c7cc4ce" exitCode=0 Mar 18 21:20:48 crc kubenswrapper[4950]: I0318 21:20:48.927491 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nsf7r" event={"ID":"d463a961-08cd-4d60-8e0e-ade812e3f63f","Type":"ContainerDied","Data":"c165a5d4a88734dbcb76e7c622acea587cc435bafe008af7c669625f4c7cc4ce"} Mar 18 21:20:48 crc kubenswrapper[4950]: I0318 21:20:48.927549 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nsf7r" Mar 18 21:20:48 crc kubenswrapper[4950]: I0318 21:20:48.927572 4950 scope.go:117] "RemoveContainer" containerID="c165a5d4a88734dbcb76e7c622acea587cc435bafe008af7c669625f4c7cc4ce" Mar 18 21:20:48 crc kubenswrapper[4950]: I0318 21:20:48.927554 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nsf7r" event={"ID":"d463a961-08cd-4d60-8e0e-ade812e3f63f","Type":"ContainerDied","Data":"c94819be3229307d49595efbd2eced87073380ef3a283cb0e5b11ac86806fda9"} Mar 18 21:20:48 crc kubenswrapper[4950]: I0318 21:20:48.978397 4950 scope.go:117] "RemoveContainer" containerID="66ece35ac5af4f5c1caa681889b15421ba0a13905ff42f2dab42da1a7d96b841" Mar 18 21:20:49 crc kubenswrapper[4950]: I0318 21:20:49.003988 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzwst\" (UniqueName: \"kubernetes.io/projected/d463a961-08cd-4d60-8e0e-ade812e3f63f-kube-api-access-jzwst\") pod \"d463a961-08cd-4d60-8e0e-ade812e3f63f\" (UID: \"d463a961-08cd-4d60-8e0e-ade812e3f63f\") " Mar 18 21:20:49 crc kubenswrapper[4950]: I0318 21:20:49.004506 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d463a961-08cd-4d60-8e0e-ade812e3f63f-utilities\") pod \"d463a961-08cd-4d60-8e0e-ade812e3f63f\" (UID: \"d463a961-08cd-4d60-8e0e-ade812e3f63f\") " Mar 18 21:20:49 crc kubenswrapper[4950]: I0318 21:20:49.004598 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d463a961-08cd-4d60-8e0e-ade812e3f63f-catalog-content\") pod \"d463a961-08cd-4d60-8e0e-ade812e3f63f\" (UID: \"d463a961-08cd-4d60-8e0e-ade812e3f63f\") " Mar 18 21:20:49 crc kubenswrapper[4950]: I0318 21:20:49.005885 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d463a961-08cd-4d60-8e0e-ade812e3f63f-utilities" (OuterVolumeSpecName: "utilities") pod "d463a961-08cd-4d60-8e0e-ade812e3f63f" (UID: "d463a961-08cd-4d60-8e0e-ade812e3f63f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:20:49 crc kubenswrapper[4950]: I0318 21:20:49.014499 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d463a961-08cd-4d60-8e0e-ade812e3f63f-kube-api-access-jzwst" (OuterVolumeSpecName: "kube-api-access-jzwst") pod "d463a961-08cd-4d60-8e0e-ade812e3f63f" (UID: "d463a961-08cd-4d60-8e0e-ade812e3f63f"). InnerVolumeSpecName "kube-api-access-jzwst". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:20:49 crc kubenswrapper[4950]: I0318 21:20:49.016698 4950 scope.go:117] "RemoveContainer" containerID="02160664edb1414e67881e635f7eca40d092dcf3755c7d474ccd1aeb23680b50" Mar 18 21:20:49 crc kubenswrapper[4950]: I0318 21:20:49.099642 4950 scope.go:117] "RemoveContainer" containerID="c165a5d4a88734dbcb76e7c622acea587cc435bafe008af7c669625f4c7cc4ce" Mar 18 21:20:49 crc kubenswrapper[4950]: E0318 21:20:49.100815 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c165a5d4a88734dbcb76e7c622acea587cc435bafe008af7c669625f4c7cc4ce\": container with ID starting with c165a5d4a88734dbcb76e7c622acea587cc435bafe008af7c669625f4c7cc4ce not found: ID does not exist" containerID="c165a5d4a88734dbcb76e7c622acea587cc435bafe008af7c669625f4c7cc4ce" Mar 18 21:20:49 crc kubenswrapper[4950]: I0318 21:20:49.100842 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c165a5d4a88734dbcb76e7c622acea587cc435bafe008af7c669625f4c7cc4ce"} err="failed to get container status \"c165a5d4a88734dbcb76e7c622acea587cc435bafe008af7c669625f4c7cc4ce\": rpc error: code = NotFound desc = could not find container \"c165a5d4a88734dbcb76e7c622acea587cc435bafe008af7c669625f4c7cc4ce\": container with ID starting with c165a5d4a88734dbcb76e7c622acea587cc435bafe008af7c669625f4c7cc4ce not found: ID does not exist" Mar 18 21:20:49 crc kubenswrapper[4950]: I0318 21:20:49.100860 4950 scope.go:117] "RemoveContainer" containerID="66ece35ac5af4f5c1caa681889b15421ba0a13905ff42f2dab42da1a7d96b841" Mar 18 21:20:49 crc kubenswrapper[4950]: E0318 21:20:49.101132 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66ece35ac5af4f5c1caa681889b15421ba0a13905ff42f2dab42da1a7d96b841\": container with ID starting with 66ece35ac5af4f5c1caa681889b15421ba0a13905ff42f2dab42da1a7d96b841 not found: ID does not exist" containerID="66ece35ac5af4f5c1caa681889b15421ba0a13905ff42f2dab42da1a7d96b841" Mar 18 21:20:49 crc kubenswrapper[4950]: I0318 21:20:49.101154 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66ece35ac5af4f5c1caa681889b15421ba0a13905ff42f2dab42da1a7d96b841"} err="failed to get container status \"66ece35ac5af4f5c1caa681889b15421ba0a13905ff42f2dab42da1a7d96b841\": rpc error: code = NotFound desc = could not find container \"66ece35ac5af4f5c1caa681889b15421ba0a13905ff42f2dab42da1a7d96b841\": container with ID starting with 66ece35ac5af4f5c1caa681889b15421ba0a13905ff42f2dab42da1a7d96b841 not found: ID does not exist" Mar 18 21:20:49 crc kubenswrapper[4950]: I0318 21:20:49.101166 4950 scope.go:117] "RemoveContainer" containerID="02160664edb1414e67881e635f7eca40d092dcf3755c7d474ccd1aeb23680b50" Mar 18 21:20:49 crc kubenswrapper[4950]: E0318 21:20:49.101560 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02160664edb1414e67881e635f7eca40d092dcf3755c7d474ccd1aeb23680b50\": container with ID starting with 02160664edb1414e67881e635f7eca40d092dcf3755c7d474ccd1aeb23680b50 not found: ID does not exist" containerID="02160664edb1414e67881e635f7eca40d092dcf3755c7d474ccd1aeb23680b50" Mar 18 21:20:49 crc kubenswrapper[4950]: I0318 21:20:49.101604 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02160664edb1414e67881e635f7eca40d092dcf3755c7d474ccd1aeb23680b50"} err="failed to get container status \"02160664edb1414e67881e635f7eca40d092dcf3755c7d474ccd1aeb23680b50\": rpc error: code = NotFound desc = could not find container \"02160664edb1414e67881e635f7eca40d092dcf3755c7d474ccd1aeb23680b50\": container with ID starting with 02160664edb1414e67881e635f7eca40d092dcf3755c7d474ccd1aeb23680b50 not found: ID does not exist" Mar 18 21:20:49 crc kubenswrapper[4950]: I0318 21:20:49.106162 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d463a961-08cd-4d60-8e0e-ade812e3f63f-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 21:20:49 crc kubenswrapper[4950]: I0318 21:20:49.106185 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzwst\" (UniqueName: \"kubernetes.io/projected/d463a961-08cd-4d60-8e0e-ade812e3f63f-kube-api-access-jzwst\") on node \"crc\" DevicePath \"\"" Mar 18 21:20:49 crc kubenswrapper[4950]: I0318 21:20:49.188283 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d463a961-08cd-4d60-8e0e-ade812e3f63f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d463a961-08cd-4d60-8e0e-ade812e3f63f" (UID: "d463a961-08cd-4d60-8e0e-ade812e3f63f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:20:49 crc kubenswrapper[4950]: I0318 21:20:49.208030 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d463a961-08cd-4d60-8e0e-ade812e3f63f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 21:20:49 crc kubenswrapper[4950]: I0318 21:20:49.264093 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nsf7r"] Mar 18 21:20:49 crc kubenswrapper[4950]: I0318 21:20:49.276040 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nsf7r"] Mar 18 21:20:49 crc kubenswrapper[4950]: I0318 21:20:49.490339 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d463a961-08cd-4d60-8e0e-ade812e3f63f" path="/var/lib/kubelet/pods/d463a961-08cd-4d60-8e0e-ade812e3f63f/volumes" Mar 18 21:21:01 crc kubenswrapper[4950]: I0318 21:21:01.737265 4950 scope.go:117] "RemoveContainer" containerID="5e4f099e712c5548236f505a3a9a4c4486f422245e990ee927e547af07b8e21a" Mar 18 21:22:00 crc kubenswrapper[4950]: I0318 21:22:00.154858 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564482-jdb8f"] Mar 18 21:22:00 crc kubenswrapper[4950]: E0318 21:22:00.155737 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d463a961-08cd-4d60-8e0e-ade812e3f63f" containerName="registry-server" Mar 18 21:22:00 crc kubenswrapper[4950]: I0318 21:22:00.155751 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="d463a961-08cd-4d60-8e0e-ade812e3f63f" containerName="registry-server" Mar 18 21:22:00 crc kubenswrapper[4950]: E0318 21:22:00.155767 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d463a961-08cd-4d60-8e0e-ade812e3f63f" containerName="extract-utilities" Mar 18 21:22:00 crc kubenswrapper[4950]: I0318 21:22:00.155775 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="d463a961-08cd-4d60-8e0e-ade812e3f63f" containerName="extract-utilities" Mar 18 21:22:00 crc kubenswrapper[4950]: E0318 21:22:00.155787 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d463a961-08cd-4d60-8e0e-ade812e3f63f" containerName="extract-content" Mar 18 21:22:00 crc kubenswrapper[4950]: I0318 21:22:00.155793 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="d463a961-08cd-4d60-8e0e-ade812e3f63f" containerName="extract-content" Mar 18 21:22:00 crc kubenswrapper[4950]: I0318 21:22:00.155981 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="d463a961-08cd-4d60-8e0e-ade812e3f63f" containerName="registry-server" Mar 18 21:22:00 crc kubenswrapper[4950]: I0318 21:22:00.156591 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564482-jdb8f" Mar 18 21:22:00 crc kubenswrapper[4950]: I0318 21:22:00.160111 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 21:22:00 crc kubenswrapper[4950]: I0318 21:22:00.160125 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 21:22:00 crc kubenswrapper[4950]: I0318 21:22:00.160162 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 21:22:00 crc kubenswrapper[4950]: I0318 21:22:00.176709 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564482-jdb8f"] Mar 18 21:22:00 crc kubenswrapper[4950]: I0318 21:22:00.225450 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scbwl\" (UniqueName: \"kubernetes.io/projected/6490a2c9-0763-4230-9468-69175b459dfd-kube-api-access-scbwl\") pod \"auto-csr-approver-29564482-jdb8f\" (UID: \"6490a2c9-0763-4230-9468-69175b459dfd\") " pod="openshift-infra/auto-csr-approver-29564482-jdb8f" Mar 18 21:22:00 crc kubenswrapper[4950]: I0318 21:22:00.327439 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scbwl\" (UniqueName: \"kubernetes.io/projected/6490a2c9-0763-4230-9468-69175b459dfd-kube-api-access-scbwl\") pod \"auto-csr-approver-29564482-jdb8f\" (UID: \"6490a2c9-0763-4230-9468-69175b459dfd\") " pod="openshift-infra/auto-csr-approver-29564482-jdb8f" Mar 18 21:22:00 crc kubenswrapper[4950]: I0318 21:22:00.360311 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scbwl\" (UniqueName: \"kubernetes.io/projected/6490a2c9-0763-4230-9468-69175b459dfd-kube-api-access-scbwl\") pod \"auto-csr-approver-29564482-jdb8f\" (UID: \"6490a2c9-0763-4230-9468-69175b459dfd\") " pod="openshift-infra/auto-csr-approver-29564482-jdb8f" Mar 18 21:22:00 crc kubenswrapper[4950]: I0318 21:22:00.472612 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564482-jdb8f" Mar 18 21:22:01 crc kubenswrapper[4950]: I0318 21:22:01.043592 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564482-jdb8f"] Mar 18 21:22:01 crc kubenswrapper[4950]: I0318 21:22:01.588808 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564482-jdb8f" event={"ID":"6490a2c9-0763-4230-9468-69175b459dfd","Type":"ContainerStarted","Data":"c5c9646984396f72dc563c292d499db647de464477e61873ee3512ad5e5fef40"} Mar 18 21:22:03 crc kubenswrapper[4950]: I0318 21:22:03.609856 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564482-jdb8f" event={"ID":"6490a2c9-0763-4230-9468-69175b459dfd","Type":"ContainerStarted","Data":"2192741bafe79f3a7b3c28cc38bf5191e1c685bf0184e65be527946f98234353"} Mar 18 21:22:04 crc kubenswrapper[4950]: I0318 21:22:04.618563 4950 generic.go:334] "Generic (PLEG): container finished" podID="6490a2c9-0763-4230-9468-69175b459dfd" containerID="2192741bafe79f3a7b3c28cc38bf5191e1c685bf0184e65be527946f98234353" exitCode=0 Mar 18 21:22:04 crc kubenswrapper[4950]: I0318 21:22:04.618649 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564482-jdb8f" event={"ID":"6490a2c9-0763-4230-9468-69175b459dfd","Type":"ContainerDied","Data":"2192741bafe79f3a7b3c28cc38bf5191e1c685bf0184e65be527946f98234353"} Mar 18 21:22:06 crc kubenswrapper[4950]: I0318 21:22:06.481784 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564482-jdb8f" Mar 18 21:22:06 crc kubenswrapper[4950]: I0318 21:22:06.679352 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scbwl\" (UniqueName: \"kubernetes.io/projected/6490a2c9-0763-4230-9468-69175b459dfd-kube-api-access-scbwl\") pod \"6490a2c9-0763-4230-9468-69175b459dfd\" (UID: \"6490a2c9-0763-4230-9468-69175b459dfd\") " Mar 18 21:22:06 crc kubenswrapper[4950]: I0318 21:22:06.688670 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6490a2c9-0763-4230-9468-69175b459dfd-kube-api-access-scbwl" (OuterVolumeSpecName: "kube-api-access-scbwl") pod "6490a2c9-0763-4230-9468-69175b459dfd" (UID: "6490a2c9-0763-4230-9468-69175b459dfd"). InnerVolumeSpecName "kube-api-access-scbwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:22:06 crc kubenswrapper[4950]: I0318 21:22:06.705905 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564476-jkc8h"] Mar 18 21:22:06 crc kubenswrapper[4950]: I0318 21:22:06.717074 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564476-jkc8h"] Mar 18 21:22:06 crc kubenswrapper[4950]: I0318 21:22:06.781348 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scbwl\" (UniqueName: \"kubernetes.io/projected/6490a2c9-0763-4230-9468-69175b459dfd-kube-api-access-scbwl\") on node \"crc\" DevicePath \"\"" Mar 18 21:22:07 crc kubenswrapper[4950]: I0318 21:22:07.110538 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564482-jdb8f" event={"ID":"6490a2c9-0763-4230-9468-69175b459dfd","Type":"ContainerDied","Data":"c5c9646984396f72dc563c292d499db647de464477e61873ee3512ad5e5fef40"} Mar 18 21:22:07 crc kubenswrapper[4950]: I0318 21:22:07.110598 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5c9646984396f72dc563c292d499db647de464477e61873ee3512ad5e5fef40" Mar 18 21:22:07 crc kubenswrapper[4950]: I0318 21:22:07.110696 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564482-jdb8f" Mar 18 21:22:07 crc kubenswrapper[4950]: I0318 21:22:07.494897 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abf3efdf-f206-4990-bc5f-6a9b20d28438" path="/var/lib/kubelet/pods/abf3efdf-f206-4990-bc5f-6a9b20d28438/volumes" Mar 18 21:22:53 crc kubenswrapper[4950]: I0318 21:22:53.576390 4950 generic.go:334] "Generic (PLEG): container finished" podID="cdc951ae-b4eb-46de-beac-53c6aa48073a" containerID="74520bc0cd6b5ec6357b4d557ad3193b248a8beafc461274c65c9cd5b6eb3571" exitCode=0 Mar 18 21:22:53 crc kubenswrapper[4950]: I0318 21:22:53.576500 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hjm8k/must-gather-p5s84" event={"ID":"cdc951ae-b4eb-46de-beac-53c6aa48073a","Type":"ContainerDied","Data":"74520bc0cd6b5ec6357b4d557ad3193b248a8beafc461274c65c9cd5b6eb3571"} Mar 18 21:22:53 crc kubenswrapper[4950]: I0318 21:22:53.577615 4950 scope.go:117] "RemoveContainer" containerID="74520bc0cd6b5ec6357b4d557ad3193b248a8beafc461274c65c9cd5b6eb3571" Mar 18 21:22:54 crc kubenswrapper[4950]: I0318 21:22:54.066024 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hjm8k_must-gather-p5s84_cdc951ae-b4eb-46de-beac-53c6aa48073a/gather/0.log" Mar 18 21:23:01 crc kubenswrapper[4950]: I0318 21:23:01.861214 4950 scope.go:117] "RemoveContainer" containerID="109a167196d7db7588ce49778deb57663081ce9e34ed55fc93d330d2da292a5b" Mar 18 21:23:03 crc kubenswrapper[4950]: I0318 21:23:03.080778 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hjm8k/must-gather-p5s84"] Mar 18 21:23:03 crc kubenswrapper[4950]: I0318 21:23:03.081380 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-hjm8k/must-gather-p5s84" podUID="cdc951ae-b4eb-46de-beac-53c6aa48073a" containerName="copy" containerID="cri-o://6c6455eb3d0e44c0ade010eb8266c6e050ae06b477d64451e0ba8a239fd1eddb" gracePeriod=2 Mar 18 21:23:03 crc kubenswrapper[4950]: I0318 21:23:03.100096 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hjm8k/must-gather-p5s84"] Mar 18 21:23:03 crc kubenswrapper[4950]: I0318 21:23:03.519057 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hjm8k_must-gather-p5s84_cdc951ae-b4eb-46de-beac-53c6aa48073a/copy/0.log" Mar 18 21:23:03 crc kubenswrapper[4950]: I0318 21:23:03.519784 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hjm8k/must-gather-p5s84" Mar 18 21:23:03 crc kubenswrapper[4950]: I0318 21:23:03.591334 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ssbs\" (UniqueName: \"kubernetes.io/projected/cdc951ae-b4eb-46de-beac-53c6aa48073a-kube-api-access-5ssbs\") pod \"cdc951ae-b4eb-46de-beac-53c6aa48073a\" (UID: \"cdc951ae-b4eb-46de-beac-53c6aa48073a\") " Mar 18 21:23:03 crc kubenswrapper[4950]: I0318 21:23:03.591569 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cdc951ae-b4eb-46de-beac-53c6aa48073a-must-gather-output\") pod \"cdc951ae-b4eb-46de-beac-53c6aa48073a\" (UID: \"cdc951ae-b4eb-46de-beac-53c6aa48073a\") " Mar 18 21:23:03 crc kubenswrapper[4950]: I0318 21:23:03.600304 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdc951ae-b4eb-46de-beac-53c6aa48073a-kube-api-access-5ssbs" (OuterVolumeSpecName: "kube-api-access-5ssbs") pod "cdc951ae-b4eb-46de-beac-53c6aa48073a" (UID: "cdc951ae-b4eb-46de-beac-53c6aa48073a"). InnerVolumeSpecName "kube-api-access-5ssbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:23:03 crc kubenswrapper[4950]: I0318 21:23:03.682981 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hjm8k_must-gather-p5s84_cdc951ae-b4eb-46de-beac-53c6aa48073a/copy/0.log" Mar 18 21:23:03 crc kubenswrapper[4950]: I0318 21:23:03.683365 4950 generic.go:334] "Generic (PLEG): container finished" podID="cdc951ae-b4eb-46de-beac-53c6aa48073a" containerID="6c6455eb3d0e44c0ade010eb8266c6e050ae06b477d64451e0ba8a239fd1eddb" exitCode=143 Mar 18 21:23:03 crc kubenswrapper[4950]: I0318 21:23:03.683494 4950 scope.go:117] "RemoveContainer" containerID="6c6455eb3d0e44c0ade010eb8266c6e050ae06b477d64451e0ba8a239fd1eddb" Mar 18 21:23:03 crc kubenswrapper[4950]: I0318 21:23:03.683618 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hjm8k/must-gather-p5s84" Mar 18 21:23:03 crc kubenswrapper[4950]: I0318 21:23:03.694512 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ssbs\" (UniqueName: \"kubernetes.io/projected/cdc951ae-b4eb-46de-beac-53c6aa48073a-kube-api-access-5ssbs\") on node \"crc\" DevicePath \"\"" Mar 18 21:23:03 crc kubenswrapper[4950]: I0318 21:23:03.704756 4950 scope.go:117] "RemoveContainer" containerID="74520bc0cd6b5ec6357b4d557ad3193b248a8beafc461274c65c9cd5b6eb3571" Mar 18 21:23:03 crc kubenswrapper[4950]: I0318 21:23:03.757456 4950 scope.go:117] "RemoveContainer" containerID="6c6455eb3d0e44c0ade010eb8266c6e050ae06b477d64451e0ba8a239fd1eddb" Mar 18 21:23:03 crc kubenswrapper[4950]: E0318 21:23:03.757937 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c6455eb3d0e44c0ade010eb8266c6e050ae06b477d64451e0ba8a239fd1eddb\": container with ID starting with 6c6455eb3d0e44c0ade010eb8266c6e050ae06b477d64451e0ba8a239fd1eddb not found: ID does not exist" containerID="6c6455eb3d0e44c0ade010eb8266c6e050ae06b477d64451e0ba8a239fd1eddb" Mar 18 21:23:03 crc kubenswrapper[4950]: I0318 21:23:03.758007 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c6455eb3d0e44c0ade010eb8266c6e050ae06b477d64451e0ba8a239fd1eddb"} err="failed to get container status \"6c6455eb3d0e44c0ade010eb8266c6e050ae06b477d64451e0ba8a239fd1eddb\": rpc error: code = NotFound desc = could not find container \"6c6455eb3d0e44c0ade010eb8266c6e050ae06b477d64451e0ba8a239fd1eddb\": container with ID starting with 6c6455eb3d0e44c0ade010eb8266c6e050ae06b477d64451e0ba8a239fd1eddb not found: ID does not exist" Mar 18 21:23:03 crc kubenswrapper[4950]: I0318 21:23:03.758041 4950 scope.go:117] "RemoveContainer" containerID="74520bc0cd6b5ec6357b4d557ad3193b248a8beafc461274c65c9cd5b6eb3571" Mar 18 21:23:03 crc kubenswrapper[4950]: E0318 21:23:03.758318 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74520bc0cd6b5ec6357b4d557ad3193b248a8beafc461274c65c9cd5b6eb3571\": container with ID starting with 74520bc0cd6b5ec6357b4d557ad3193b248a8beafc461274c65c9cd5b6eb3571 not found: ID does not exist" containerID="74520bc0cd6b5ec6357b4d557ad3193b248a8beafc461274c65c9cd5b6eb3571" Mar 18 21:23:03 crc kubenswrapper[4950]: I0318 21:23:03.758345 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74520bc0cd6b5ec6357b4d557ad3193b248a8beafc461274c65c9cd5b6eb3571"} err="failed to get container status \"74520bc0cd6b5ec6357b4d557ad3193b248a8beafc461274c65c9cd5b6eb3571\": rpc error: code = NotFound desc = could not find container \"74520bc0cd6b5ec6357b4d557ad3193b248a8beafc461274c65c9cd5b6eb3571\": container with ID starting with 74520bc0cd6b5ec6357b4d557ad3193b248a8beafc461274c65c9cd5b6eb3571 not found: ID does not exist" Mar 18 21:23:03 crc kubenswrapper[4950]: I0318 21:23:03.784362 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdc951ae-b4eb-46de-beac-53c6aa48073a-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "cdc951ae-b4eb-46de-beac-53c6aa48073a" (UID: "cdc951ae-b4eb-46de-beac-53c6aa48073a"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:23:03 crc kubenswrapper[4950]: I0318 21:23:03.795941 4950 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/cdc951ae-b4eb-46de-beac-53c6aa48073a-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 18 21:23:03 crc kubenswrapper[4950]: I0318 21:23:03.836492 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 21:23:03 crc kubenswrapper[4950]: I0318 21:23:03.836538 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 21:23:05 crc kubenswrapper[4950]: I0318 21:23:05.500344 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdc951ae-b4eb-46de-beac-53c6aa48073a" path="/var/lib/kubelet/pods/cdc951ae-b4eb-46de-beac-53c6aa48073a/volumes" Mar 18 21:23:33 crc kubenswrapper[4950]: I0318 21:23:33.835964 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 21:23:33 crc kubenswrapper[4950]: I0318 21:23:33.836398 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 21:24:00 crc kubenswrapper[4950]: I0318 21:24:00.162321 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564484-6vdc2"] Mar 18 21:24:00 crc kubenswrapper[4950]: E0318 21:24:00.163687 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdc951ae-b4eb-46de-beac-53c6aa48073a" containerName="gather" Mar 18 21:24:00 crc kubenswrapper[4950]: I0318 21:24:00.163705 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdc951ae-b4eb-46de-beac-53c6aa48073a" containerName="gather" Mar 18 21:24:00 crc kubenswrapper[4950]: E0318 21:24:00.163733 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdc951ae-b4eb-46de-beac-53c6aa48073a" containerName="copy" Mar 18 21:24:00 crc kubenswrapper[4950]: I0318 21:24:00.163741 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdc951ae-b4eb-46de-beac-53c6aa48073a" containerName="copy" Mar 18 21:24:00 crc kubenswrapper[4950]: E0318 21:24:00.163763 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6490a2c9-0763-4230-9468-69175b459dfd" containerName="oc" Mar 18 21:24:00 crc kubenswrapper[4950]: I0318 21:24:00.163773 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="6490a2c9-0763-4230-9468-69175b459dfd" containerName="oc" Mar 18 21:24:00 crc kubenswrapper[4950]: I0318 21:24:00.163972 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdc951ae-b4eb-46de-beac-53c6aa48073a" containerName="copy" Mar 18 21:24:00 crc kubenswrapper[4950]: I0318 21:24:00.163986 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdc951ae-b4eb-46de-beac-53c6aa48073a" containerName="gather" Mar 18 21:24:00 crc kubenswrapper[4950]: I0318 21:24:00.163997 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="6490a2c9-0763-4230-9468-69175b459dfd" containerName="oc" Mar 18 21:24:00 crc kubenswrapper[4950]: I0318 21:24:00.164717 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564484-6vdc2" Mar 18 21:24:00 crc kubenswrapper[4950]: I0318 21:24:00.172488 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564484-6vdc2"] Mar 18 21:24:00 crc kubenswrapper[4950]: I0318 21:24:00.205779 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 21:24:00 crc kubenswrapper[4950]: I0318 21:24:00.205789 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 21:24:00 crc kubenswrapper[4950]: I0318 21:24:00.206526 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 21:24:00 crc kubenswrapper[4950]: I0318 21:24:00.309321 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsj79\" (UniqueName: \"kubernetes.io/projected/603b946e-91e5-49e1-b6e6-1c983bce0741-kube-api-access-hsj79\") pod \"auto-csr-approver-29564484-6vdc2\" (UID: \"603b946e-91e5-49e1-b6e6-1c983bce0741\") " pod="openshift-infra/auto-csr-approver-29564484-6vdc2" Mar 18 21:24:00 crc kubenswrapper[4950]: I0318 21:24:00.411952 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsj79\" (UniqueName: \"kubernetes.io/projected/603b946e-91e5-49e1-b6e6-1c983bce0741-kube-api-access-hsj79\") pod \"auto-csr-approver-29564484-6vdc2\" (UID: \"603b946e-91e5-49e1-b6e6-1c983bce0741\") " pod="openshift-infra/auto-csr-approver-29564484-6vdc2" Mar 18 21:24:00 crc kubenswrapper[4950]: I0318 21:24:00.432358 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsj79\" (UniqueName: \"kubernetes.io/projected/603b946e-91e5-49e1-b6e6-1c983bce0741-kube-api-access-hsj79\") pod \"auto-csr-approver-29564484-6vdc2\" (UID: \"603b946e-91e5-49e1-b6e6-1c983bce0741\") " pod="openshift-infra/auto-csr-approver-29564484-6vdc2" Mar 18 21:24:00 crc kubenswrapper[4950]: I0318 21:24:00.526272 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564484-6vdc2" Mar 18 21:24:01 crc kubenswrapper[4950]: I0318 21:24:01.031454 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564484-6vdc2"] Mar 18 21:24:01 crc kubenswrapper[4950]: W0318 21:24:01.039086 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod603b946e_91e5_49e1_b6e6_1c983bce0741.slice/crio-b3297ece0c13d8db1f8a45d2f793a045be578dca5f28026ef2d71a00c0d988cd WatchSource:0}: Error finding container b3297ece0c13d8db1f8a45d2f793a045be578dca5f28026ef2d71a00c0d988cd: Status 404 returned error can't find the container with id b3297ece0c13d8db1f8a45d2f793a045be578dca5f28026ef2d71a00c0d988cd Mar 18 21:24:01 crc kubenswrapper[4950]: I0318 21:24:01.042510 4950 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 21:24:01 crc kubenswrapper[4950]: I0318 21:24:01.231303 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564484-6vdc2" event={"ID":"603b946e-91e5-49e1-b6e6-1c983bce0741","Type":"ContainerStarted","Data":"b3297ece0c13d8db1f8a45d2f793a045be578dca5f28026ef2d71a00c0d988cd"} Mar 18 21:24:03 crc kubenswrapper[4950]: I0318 21:24:03.251591 4950 generic.go:334] "Generic (PLEG): container finished" podID="603b946e-91e5-49e1-b6e6-1c983bce0741" containerID="dad741bf8d907bf950db1e7da6e030c01bea436686f80d2875f1db450f84a185" exitCode=0 Mar 18 21:24:03 crc kubenswrapper[4950]: I0318 21:24:03.252005 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564484-6vdc2" event={"ID":"603b946e-91e5-49e1-b6e6-1c983bce0741","Type":"ContainerDied","Data":"dad741bf8d907bf950db1e7da6e030c01bea436686f80d2875f1db450f84a185"} Mar 18 21:24:03 crc kubenswrapper[4950]: I0318 21:24:03.839990 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 21:24:03 crc kubenswrapper[4950]: I0318 21:24:03.840038 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 21:24:03 crc kubenswrapper[4950]: I0318 21:24:03.840076 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 21:24:03 crc kubenswrapper[4950]: I0318 21:24:03.840788 4950 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d"} pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 21:24:03 crc kubenswrapper[4950]: I0318 21:24:03.840830 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" containerID="cri-o://cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" gracePeriod=600 Mar 18 21:24:03 crc kubenswrapper[4950]: E0318 21:24:03.990258 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:24:04 crc kubenswrapper[4950]: I0318 21:24:04.261084 4950 generic.go:334] "Generic (PLEG): container finished" podID="4048b439-3266-46e7-9de0-22377efacc46" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" exitCode=0 Mar 18 21:24:04 crc kubenswrapper[4950]: I0318 21:24:04.261105 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerDied","Data":"cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d"} Mar 18 21:24:04 crc kubenswrapper[4950]: I0318 21:24:04.262232 4950 scope.go:117] "RemoveContainer" containerID="6fb2941ff91c9d7625aec6265d6c6aecb9312056ca715d8dc848cdcf34657cae" Mar 18 21:24:04 crc kubenswrapper[4950]: I0318 21:24:04.262902 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:24:04 crc kubenswrapper[4950]: E0318 21:24:04.263205 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:24:04 crc kubenswrapper[4950]: I0318 21:24:04.647856 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564484-6vdc2" Mar 18 21:24:04 crc kubenswrapper[4950]: I0318 21:24:04.798325 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsj79\" (UniqueName: \"kubernetes.io/projected/603b946e-91e5-49e1-b6e6-1c983bce0741-kube-api-access-hsj79\") pod \"603b946e-91e5-49e1-b6e6-1c983bce0741\" (UID: \"603b946e-91e5-49e1-b6e6-1c983bce0741\") " Mar 18 21:24:04 crc kubenswrapper[4950]: I0318 21:24:04.805176 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/603b946e-91e5-49e1-b6e6-1c983bce0741-kube-api-access-hsj79" (OuterVolumeSpecName: "kube-api-access-hsj79") pod "603b946e-91e5-49e1-b6e6-1c983bce0741" (UID: "603b946e-91e5-49e1-b6e6-1c983bce0741"). InnerVolumeSpecName "kube-api-access-hsj79". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:24:04 crc kubenswrapper[4950]: I0318 21:24:04.900705 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsj79\" (UniqueName: \"kubernetes.io/projected/603b946e-91e5-49e1-b6e6-1c983bce0741-kube-api-access-hsj79\") on node \"crc\" DevicePath \"\"" Mar 18 21:24:04 crc kubenswrapper[4950]: I0318 21:24:04.949743 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k5h8f"] Mar 18 21:24:04 crc kubenswrapper[4950]: E0318 21:24:04.950202 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="603b946e-91e5-49e1-b6e6-1c983bce0741" containerName="oc" Mar 18 21:24:04 crc kubenswrapper[4950]: I0318 21:24:04.950220 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="603b946e-91e5-49e1-b6e6-1c983bce0741" containerName="oc" Mar 18 21:24:04 crc kubenswrapper[4950]: I0318 21:24:04.950437 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="603b946e-91e5-49e1-b6e6-1c983bce0741" containerName="oc" Mar 18 21:24:04 crc kubenswrapper[4950]: I0318 21:24:04.951659 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k5h8f" Mar 18 21:24:04 crc kubenswrapper[4950]: I0318 21:24:04.974633 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k5h8f"] Mar 18 21:24:05 crc kubenswrapper[4950]: I0318 21:24:05.104300 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d643a3a-937f-4b3a-a36e-1a18df74a949-utilities\") pod \"certified-operators-k5h8f\" (UID: \"1d643a3a-937f-4b3a-a36e-1a18df74a949\") " pod="openshift-marketplace/certified-operators-k5h8f" Mar 18 21:24:05 crc kubenswrapper[4950]: I0318 21:24:05.104710 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7lvg\" (UniqueName: \"kubernetes.io/projected/1d643a3a-937f-4b3a-a36e-1a18df74a949-kube-api-access-b7lvg\") pod \"certified-operators-k5h8f\" (UID: \"1d643a3a-937f-4b3a-a36e-1a18df74a949\") " pod="openshift-marketplace/certified-operators-k5h8f" Mar 18 21:24:05 crc kubenswrapper[4950]: I0318 21:24:05.104767 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d643a3a-937f-4b3a-a36e-1a18df74a949-catalog-content\") pod \"certified-operators-k5h8f\" (UID: \"1d643a3a-937f-4b3a-a36e-1a18df74a949\") " pod="openshift-marketplace/certified-operators-k5h8f" Mar 18 21:24:05 crc kubenswrapper[4950]: I0318 21:24:05.206421 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7lvg\" (UniqueName: \"kubernetes.io/projected/1d643a3a-937f-4b3a-a36e-1a18df74a949-kube-api-access-b7lvg\") pod \"certified-operators-k5h8f\" (UID: \"1d643a3a-937f-4b3a-a36e-1a18df74a949\") " pod="openshift-marketplace/certified-operators-k5h8f" Mar 18 21:24:05 crc kubenswrapper[4950]: I0318 21:24:05.206478 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d643a3a-937f-4b3a-a36e-1a18df74a949-catalog-content\") pod \"certified-operators-k5h8f\" (UID: \"1d643a3a-937f-4b3a-a36e-1a18df74a949\") " pod="openshift-marketplace/certified-operators-k5h8f" Mar 18 21:24:05 crc kubenswrapper[4950]: I0318 21:24:05.206528 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d643a3a-937f-4b3a-a36e-1a18df74a949-utilities\") pod \"certified-operators-k5h8f\" (UID: \"1d643a3a-937f-4b3a-a36e-1a18df74a949\") " pod="openshift-marketplace/certified-operators-k5h8f" Mar 18 21:24:05 crc kubenswrapper[4950]: I0318 21:24:05.206940 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d643a3a-937f-4b3a-a36e-1a18df74a949-catalog-content\") pod \"certified-operators-k5h8f\" (UID: \"1d643a3a-937f-4b3a-a36e-1a18df74a949\") " pod="openshift-marketplace/certified-operators-k5h8f" Mar 18 21:24:05 crc kubenswrapper[4950]: I0318 21:24:05.206999 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d643a3a-937f-4b3a-a36e-1a18df74a949-utilities\") pod \"certified-operators-k5h8f\" (UID: \"1d643a3a-937f-4b3a-a36e-1a18df74a949\") " pod="openshift-marketplace/certified-operators-k5h8f" Mar 18 21:24:05 crc kubenswrapper[4950]: I0318 21:24:05.223994 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7lvg\" (UniqueName: \"kubernetes.io/projected/1d643a3a-937f-4b3a-a36e-1a18df74a949-kube-api-access-b7lvg\") pod \"certified-operators-k5h8f\" (UID: \"1d643a3a-937f-4b3a-a36e-1a18df74a949\") " pod="openshift-marketplace/certified-operators-k5h8f" Mar 18 21:24:05 crc kubenswrapper[4950]: I0318 21:24:05.268493 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k5h8f" Mar 18 21:24:05 crc kubenswrapper[4950]: I0318 21:24:05.275101 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564484-6vdc2" event={"ID":"603b946e-91e5-49e1-b6e6-1c983bce0741","Type":"ContainerDied","Data":"b3297ece0c13d8db1f8a45d2f793a045be578dca5f28026ef2d71a00c0d988cd"} Mar 18 21:24:05 crc kubenswrapper[4950]: I0318 21:24:05.275163 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3297ece0c13d8db1f8a45d2f793a045be578dca5f28026ef2d71a00c0d988cd" Mar 18 21:24:05 crc kubenswrapper[4950]: I0318 21:24:05.275571 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564484-6vdc2" Mar 18 21:24:05 crc kubenswrapper[4950]: I0318 21:24:05.867685 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564478-p9khw"] Mar 18 21:24:05 crc kubenswrapper[4950]: I0318 21:24:05.875687 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564478-p9khw"] Mar 18 21:24:06 crc kubenswrapper[4950]: I0318 21:24:06.292608 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k5h8f"] Mar 18 21:24:06 crc kubenswrapper[4950]: W0318 21:24:06.298571 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d643a3a_937f_4b3a_a36e_1a18df74a949.slice/crio-bccbadc361ebf77b53439addf435e64a6bbb7ea9a70f15db7fee5089a12b2bdc WatchSource:0}: Error finding container bccbadc361ebf77b53439addf435e64a6bbb7ea9a70f15db7fee5089a12b2bdc: Status 404 returned error can't find the container with id bccbadc361ebf77b53439addf435e64a6bbb7ea9a70f15db7fee5089a12b2bdc Mar 18 21:24:07 crc kubenswrapper[4950]: I0318 21:24:07.295913 4950 generic.go:334] "Generic (PLEG): container finished" podID="1d643a3a-937f-4b3a-a36e-1a18df74a949" containerID="2d1f9a1535b272de1aa9c06ed3bcd01705f39fc3e78c8512f3d678d3f1021f24" exitCode=0 Mar 18 21:24:07 crc kubenswrapper[4950]: I0318 21:24:07.296003 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k5h8f" event={"ID":"1d643a3a-937f-4b3a-a36e-1a18df74a949","Type":"ContainerDied","Data":"2d1f9a1535b272de1aa9c06ed3bcd01705f39fc3e78c8512f3d678d3f1021f24"} Mar 18 21:24:07 crc kubenswrapper[4950]: I0318 21:24:07.297593 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k5h8f" event={"ID":"1d643a3a-937f-4b3a-a36e-1a18df74a949","Type":"ContainerStarted","Data":"bccbadc361ebf77b53439addf435e64a6bbb7ea9a70f15db7fee5089a12b2bdc"} Mar 18 21:24:07 crc kubenswrapper[4950]: I0318 21:24:07.490174 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c0fcd06-ee4a-490b-bef9-6ea21dfa6987" path="/var/lib/kubelet/pods/2c0fcd06-ee4a-490b-bef9-6ea21dfa6987/volumes" Mar 18 21:24:08 crc kubenswrapper[4950]: I0318 21:24:08.318349 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k5h8f" event={"ID":"1d643a3a-937f-4b3a-a36e-1a18df74a949","Type":"ContainerStarted","Data":"6a27acc137af1e61f5e5da655092db66ad34686c4806e162cc419f01951af7fd"} Mar 18 21:24:10 crc kubenswrapper[4950]: I0318 21:24:10.337127 4950 generic.go:334] "Generic (PLEG): container finished" podID="1d643a3a-937f-4b3a-a36e-1a18df74a949" containerID="6a27acc137af1e61f5e5da655092db66ad34686c4806e162cc419f01951af7fd" exitCode=0 Mar 18 21:24:10 crc kubenswrapper[4950]: I0318 21:24:10.337238 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k5h8f" event={"ID":"1d643a3a-937f-4b3a-a36e-1a18df74a949","Type":"ContainerDied","Data":"6a27acc137af1e61f5e5da655092db66ad34686c4806e162cc419f01951af7fd"} Mar 18 21:24:12 crc kubenswrapper[4950]: I0318 21:24:12.373887 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k5h8f" event={"ID":"1d643a3a-937f-4b3a-a36e-1a18df74a949","Type":"ContainerStarted","Data":"e0542b535a612ef259bd447f5f3d53d8bc154c89189ac64af32761150210e39a"} Mar 18 21:24:12 crc kubenswrapper[4950]: I0318 21:24:12.399817 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k5h8f" podStartSLOduration=4.9604054269999995 podStartE2EDuration="8.399794903s" podCreationTimestamp="2026-03-18 21:24:04 +0000 UTC" firstStartedPulling="2026-03-18 21:24:07.298753641 +0000 UTC m=+4660.539595509" lastFinishedPulling="2026-03-18 21:24:10.738143107 +0000 UTC m=+4663.978984985" observedRunningTime="2026-03-18 21:24:12.390057917 +0000 UTC m=+4665.630899785" watchObservedRunningTime="2026-03-18 21:24:12.399794903 +0000 UTC m=+4665.640636771" Mar 18 21:24:15 crc kubenswrapper[4950]: I0318 21:24:15.269201 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k5h8f" Mar 18 21:24:15 crc kubenswrapper[4950]: I0318 21:24:15.269844 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k5h8f" Mar 18 21:24:15 crc kubenswrapper[4950]: I0318 21:24:15.352306 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k5h8f" Mar 18 21:24:17 crc kubenswrapper[4950]: I0318 21:24:17.491526 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:24:17 crc kubenswrapper[4950]: E0318 21:24:17.492074 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:24:25 crc kubenswrapper[4950]: I0318 21:24:25.478299 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k5h8f" Mar 18 21:24:25 crc kubenswrapper[4950]: I0318 21:24:25.529910 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k5h8f"] Mar 18 21:24:25 crc kubenswrapper[4950]: I0318 21:24:25.530189 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-k5h8f" podUID="1d643a3a-937f-4b3a-a36e-1a18df74a949" containerName="registry-server" containerID="cri-o://e0542b535a612ef259bd447f5f3d53d8bc154c89189ac64af32761150210e39a" gracePeriod=2 Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.055566 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k5h8f" Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.070562 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7lvg\" (UniqueName: \"kubernetes.io/projected/1d643a3a-937f-4b3a-a36e-1a18df74a949-kube-api-access-b7lvg\") pod \"1d643a3a-937f-4b3a-a36e-1a18df74a949\" (UID: \"1d643a3a-937f-4b3a-a36e-1a18df74a949\") " Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.070620 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d643a3a-937f-4b3a-a36e-1a18df74a949-catalog-content\") pod \"1d643a3a-937f-4b3a-a36e-1a18df74a949\" (UID: \"1d643a3a-937f-4b3a-a36e-1a18df74a949\") " Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.070672 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d643a3a-937f-4b3a-a36e-1a18df74a949-utilities\") pod \"1d643a3a-937f-4b3a-a36e-1a18df74a949\" (UID: \"1d643a3a-937f-4b3a-a36e-1a18df74a949\") " Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.071833 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d643a3a-937f-4b3a-a36e-1a18df74a949-utilities" (OuterVolumeSpecName: "utilities") pod "1d643a3a-937f-4b3a-a36e-1a18df74a949" (UID: "1d643a3a-937f-4b3a-a36e-1a18df74a949"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.084750 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d643a3a-937f-4b3a-a36e-1a18df74a949-kube-api-access-b7lvg" (OuterVolumeSpecName: "kube-api-access-b7lvg") pod "1d643a3a-937f-4b3a-a36e-1a18df74a949" (UID: "1d643a3a-937f-4b3a-a36e-1a18df74a949"). InnerVolumeSpecName "kube-api-access-b7lvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.137140 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d643a3a-937f-4b3a-a36e-1a18df74a949-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d643a3a-937f-4b3a-a36e-1a18df74a949" (UID: "1d643a3a-937f-4b3a-a36e-1a18df74a949"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.172894 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d643a3a-937f-4b3a-a36e-1a18df74a949-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.172928 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7lvg\" (UniqueName: \"kubernetes.io/projected/1d643a3a-937f-4b3a-a36e-1a18df74a949-kube-api-access-b7lvg\") on node \"crc\" DevicePath \"\"" Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.172938 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d643a3a-937f-4b3a-a36e-1a18df74a949-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.519816 4950 generic.go:334] "Generic (PLEG): container finished" podID="1d643a3a-937f-4b3a-a36e-1a18df74a949" containerID="e0542b535a612ef259bd447f5f3d53d8bc154c89189ac64af32761150210e39a" exitCode=0 Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.519934 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k5h8f" Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.519935 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k5h8f" event={"ID":"1d643a3a-937f-4b3a-a36e-1a18df74a949","Type":"ContainerDied","Data":"e0542b535a612ef259bd447f5f3d53d8bc154c89189ac64af32761150210e39a"} Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.520026 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k5h8f" event={"ID":"1d643a3a-937f-4b3a-a36e-1a18df74a949","Type":"ContainerDied","Data":"bccbadc361ebf77b53439addf435e64a6bbb7ea9a70f15db7fee5089a12b2bdc"} Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.520066 4950 scope.go:117] "RemoveContainer" containerID="e0542b535a612ef259bd447f5f3d53d8bc154c89189ac64af32761150210e39a" Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.555312 4950 scope.go:117] "RemoveContainer" containerID="6a27acc137af1e61f5e5da655092db66ad34686c4806e162cc419f01951af7fd" Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.565956 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k5h8f"] Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.585248 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-k5h8f"] Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.616738 4950 scope.go:117] "RemoveContainer" containerID="2d1f9a1535b272de1aa9c06ed3bcd01705f39fc3e78c8512f3d678d3f1021f24" Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.655227 4950 scope.go:117] "RemoveContainer" containerID="e0542b535a612ef259bd447f5f3d53d8bc154c89189ac64af32761150210e39a" Mar 18 21:24:26 crc kubenswrapper[4950]: E0318 21:24:26.655771 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0542b535a612ef259bd447f5f3d53d8bc154c89189ac64af32761150210e39a\": container with ID starting with e0542b535a612ef259bd447f5f3d53d8bc154c89189ac64af32761150210e39a not found: ID does not exist" containerID="e0542b535a612ef259bd447f5f3d53d8bc154c89189ac64af32761150210e39a" Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.655842 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0542b535a612ef259bd447f5f3d53d8bc154c89189ac64af32761150210e39a"} err="failed to get container status \"e0542b535a612ef259bd447f5f3d53d8bc154c89189ac64af32761150210e39a\": rpc error: code = NotFound desc = could not find container \"e0542b535a612ef259bd447f5f3d53d8bc154c89189ac64af32761150210e39a\": container with ID starting with e0542b535a612ef259bd447f5f3d53d8bc154c89189ac64af32761150210e39a not found: ID does not exist" Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.655871 4950 scope.go:117] "RemoveContainer" containerID="6a27acc137af1e61f5e5da655092db66ad34686c4806e162cc419f01951af7fd" Mar 18 21:24:26 crc kubenswrapper[4950]: E0318 21:24:26.656222 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a27acc137af1e61f5e5da655092db66ad34686c4806e162cc419f01951af7fd\": container with ID starting with 6a27acc137af1e61f5e5da655092db66ad34686c4806e162cc419f01951af7fd not found: ID does not exist" containerID="6a27acc137af1e61f5e5da655092db66ad34686c4806e162cc419f01951af7fd" Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.656266 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a27acc137af1e61f5e5da655092db66ad34686c4806e162cc419f01951af7fd"} err="failed to get container status \"6a27acc137af1e61f5e5da655092db66ad34686c4806e162cc419f01951af7fd\": rpc error: code = NotFound desc = could not find container \"6a27acc137af1e61f5e5da655092db66ad34686c4806e162cc419f01951af7fd\": container with ID starting with 6a27acc137af1e61f5e5da655092db66ad34686c4806e162cc419f01951af7fd not found: ID does not exist" Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.656294 4950 scope.go:117] "RemoveContainer" containerID="2d1f9a1535b272de1aa9c06ed3bcd01705f39fc3e78c8512f3d678d3f1021f24" Mar 18 21:24:26 crc kubenswrapper[4950]: E0318 21:24:26.656780 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d1f9a1535b272de1aa9c06ed3bcd01705f39fc3e78c8512f3d678d3f1021f24\": container with ID starting with 2d1f9a1535b272de1aa9c06ed3bcd01705f39fc3e78c8512f3d678d3f1021f24 not found: ID does not exist" containerID="2d1f9a1535b272de1aa9c06ed3bcd01705f39fc3e78c8512f3d678d3f1021f24" Mar 18 21:24:26 crc kubenswrapper[4950]: I0318 21:24:26.656811 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d1f9a1535b272de1aa9c06ed3bcd01705f39fc3e78c8512f3d678d3f1021f24"} err="failed to get container status \"2d1f9a1535b272de1aa9c06ed3bcd01705f39fc3e78c8512f3d678d3f1021f24\": rpc error: code = NotFound desc = could not find container \"2d1f9a1535b272de1aa9c06ed3bcd01705f39fc3e78c8512f3d678d3f1021f24\": container with ID starting with 2d1f9a1535b272de1aa9c06ed3bcd01705f39fc3e78c8512f3d678d3f1021f24 not found: ID does not exist" Mar 18 21:24:27 crc kubenswrapper[4950]: I0318 21:24:27.494898 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d643a3a-937f-4b3a-a36e-1a18df74a949" path="/var/lib/kubelet/pods/1d643a3a-937f-4b3a-a36e-1a18df74a949/volumes" Mar 18 21:24:32 crc kubenswrapper[4950]: I0318 21:24:32.481093 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:24:32 crc kubenswrapper[4950]: E0318 21:24:32.483970 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:24:46 crc kubenswrapper[4950]: I0318 21:24:46.479738 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:24:46 crc kubenswrapper[4950]: E0318 21:24:46.480555 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:24:57 crc kubenswrapper[4950]: I0318 21:24:57.480624 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:24:57 crc kubenswrapper[4950]: E0318 21:24:57.483979 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:25:02 crc kubenswrapper[4950]: I0318 21:25:02.042693 4950 scope.go:117] "RemoveContainer" containerID="e0b3baa00afead2ce35133363d84b3f066f738c0bb6e35a6e94da564aee28c9a" Mar 18 21:25:10 crc kubenswrapper[4950]: I0318 21:25:10.480536 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:25:10 crc kubenswrapper[4950]: E0318 21:25:10.482694 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:25:21 crc kubenswrapper[4950]: I0318 21:25:21.480499 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:25:21 crc kubenswrapper[4950]: E0318 21:25:21.481579 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:25:35 crc kubenswrapper[4950]: I0318 21:25:35.479921 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:25:35 crc kubenswrapper[4950]: E0318 21:25:35.480502 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:25:47 crc kubenswrapper[4950]: I0318 21:25:47.733184 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xtwkj/must-gather-c9z75"] Mar 18 21:25:47 crc kubenswrapper[4950]: E0318 21:25:47.733956 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d643a3a-937f-4b3a-a36e-1a18df74a949" containerName="extract-content" Mar 18 21:25:47 crc kubenswrapper[4950]: I0318 21:25:47.733968 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d643a3a-937f-4b3a-a36e-1a18df74a949" containerName="extract-content" Mar 18 21:25:47 crc kubenswrapper[4950]: E0318 21:25:47.733985 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d643a3a-937f-4b3a-a36e-1a18df74a949" containerName="registry-server" Mar 18 21:25:47 crc kubenswrapper[4950]: I0318 21:25:47.733990 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d643a3a-937f-4b3a-a36e-1a18df74a949" containerName="registry-server" Mar 18 21:25:47 crc kubenswrapper[4950]: E0318 21:25:47.734006 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d643a3a-937f-4b3a-a36e-1a18df74a949" containerName="extract-utilities" Mar 18 21:25:47 crc kubenswrapper[4950]: I0318 21:25:47.734012 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d643a3a-937f-4b3a-a36e-1a18df74a949" containerName="extract-utilities" Mar 18 21:25:47 crc kubenswrapper[4950]: I0318 21:25:47.734162 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d643a3a-937f-4b3a-a36e-1a18df74a949" containerName="registry-server" Mar 18 21:25:47 crc kubenswrapper[4950]: I0318 21:25:47.737983 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xtwkj/must-gather-c9z75" Mar 18 21:25:47 crc kubenswrapper[4950]: I0318 21:25:47.742789 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-xtwkj"/"openshift-service-ca.crt" Mar 18 21:25:47 crc kubenswrapper[4950]: I0318 21:25:47.742911 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-xtwkj"/"kube-root-ca.crt" Mar 18 21:25:47 crc kubenswrapper[4950]: I0318 21:25:47.742798 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-xtwkj"/"default-dockercfg-r85dp" Mar 18 21:25:47 crc kubenswrapper[4950]: I0318 21:25:47.785115 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xtwkj/must-gather-c9z75"] Mar 18 21:25:47 crc kubenswrapper[4950]: I0318 21:25:47.880911 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfkgz\" (UniqueName: \"kubernetes.io/projected/b71b3828-70c5-4ec1-9b72-b761fc24e58a-kube-api-access-gfkgz\") pod \"must-gather-c9z75\" (UID: \"b71b3828-70c5-4ec1-9b72-b761fc24e58a\") " pod="openshift-must-gather-xtwkj/must-gather-c9z75" Mar 18 21:25:47 crc kubenswrapper[4950]: I0318 21:25:47.881129 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b71b3828-70c5-4ec1-9b72-b761fc24e58a-must-gather-output\") pod \"must-gather-c9z75\" (UID: \"b71b3828-70c5-4ec1-9b72-b761fc24e58a\") " pod="openshift-must-gather-xtwkj/must-gather-c9z75" Mar 18 21:25:47 crc kubenswrapper[4950]: I0318 21:25:47.982635 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b71b3828-70c5-4ec1-9b72-b761fc24e58a-must-gather-output\") pod \"must-gather-c9z75\" (UID: \"b71b3828-70c5-4ec1-9b72-b761fc24e58a\") " pod="openshift-must-gather-xtwkj/must-gather-c9z75" Mar 18 21:25:47 crc kubenswrapper[4950]: I0318 21:25:47.982750 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfkgz\" (UniqueName: \"kubernetes.io/projected/b71b3828-70c5-4ec1-9b72-b761fc24e58a-kube-api-access-gfkgz\") pod \"must-gather-c9z75\" (UID: \"b71b3828-70c5-4ec1-9b72-b761fc24e58a\") " pod="openshift-must-gather-xtwkj/must-gather-c9z75" Mar 18 21:25:47 crc kubenswrapper[4950]: I0318 21:25:47.983065 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b71b3828-70c5-4ec1-9b72-b761fc24e58a-must-gather-output\") pod \"must-gather-c9z75\" (UID: \"b71b3828-70c5-4ec1-9b72-b761fc24e58a\") " pod="openshift-must-gather-xtwkj/must-gather-c9z75" Mar 18 21:25:48 crc kubenswrapper[4950]: I0318 21:25:48.003493 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfkgz\" (UniqueName: \"kubernetes.io/projected/b71b3828-70c5-4ec1-9b72-b761fc24e58a-kube-api-access-gfkgz\") pod \"must-gather-c9z75\" (UID: \"b71b3828-70c5-4ec1-9b72-b761fc24e58a\") " pod="openshift-must-gather-xtwkj/must-gather-c9z75" Mar 18 21:25:48 crc kubenswrapper[4950]: I0318 21:25:48.055206 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xtwkj/must-gather-c9z75" Mar 18 21:25:48 crc kubenswrapper[4950]: I0318 21:25:48.512477 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xtwkj/must-gather-c9z75"] Mar 18 21:25:49 crc kubenswrapper[4950]: I0318 21:25:49.396199 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xtwkj/must-gather-c9z75" event={"ID":"b71b3828-70c5-4ec1-9b72-b761fc24e58a","Type":"ContainerStarted","Data":"24f78ca40020d2e14b0e130e8d00374b0bfe298ae379b47d22f5b37d6a4f3dd0"} Mar 18 21:25:49 crc kubenswrapper[4950]: I0318 21:25:49.396795 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xtwkj/must-gather-c9z75" event={"ID":"b71b3828-70c5-4ec1-9b72-b761fc24e58a","Type":"ContainerStarted","Data":"ca2236d9b2100846ae70c67f8ffe0adb64ea06dd53201aee8abc967e9afa051d"} Mar 18 21:25:49 crc kubenswrapper[4950]: I0318 21:25:49.479570 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:25:49 crc kubenswrapper[4950]: E0318 21:25:49.479964 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:25:50 crc kubenswrapper[4950]: I0318 21:25:50.406585 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xtwkj/must-gather-c9z75" event={"ID":"b71b3828-70c5-4ec1-9b72-b761fc24e58a","Type":"ContainerStarted","Data":"ba4fe7ecf4b2cb9c3ed96352b8e023b399cf69cf24921e59f0bb7e1b527d256d"} Mar 18 21:25:50 crc kubenswrapper[4950]: I0318 21:25:50.452799 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xtwkj/must-gather-c9z75" podStartSLOduration=3.452778368 podStartE2EDuration="3.452778368s" podCreationTimestamp="2026-03-18 21:25:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 21:25:50.450588941 +0000 UTC m=+4763.691430809" watchObservedRunningTime="2026-03-18 21:25:50.452778368 +0000 UTC m=+4763.693620236" Mar 18 21:25:53 crc kubenswrapper[4950]: I0318 21:25:53.754623 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xtwkj/crc-debug-74xfj"] Mar 18 21:25:53 crc kubenswrapper[4950]: I0318 21:25:53.756284 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xtwkj/crc-debug-74xfj" Mar 18 21:25:53 crc kubenswrapper[4950]: I0318 21:25:53.907526 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0162dc8c-7c36-41e4-b511-77931dc6de63-host\") pod \"crc-debug-74xfj\" (UID: \"0162dc8c-7c36-41e4-b511-77931dc6de63\") " pod="openshift-must-gather-xtwkj/crc-debug-74xfj" Mar 18 21:25:53 crc kubenswrapper[4950]: I0318 21:25:53.907606 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fvtm\" (UniqueName: \"kubernetes.io/projected/0162dc8c-7c36-41e4-b511-77931dc6de63-kube-api-access-7fvtm\") pod \"crc-debug-74xfj\" (UID: \"0162dc8c-7c36-41e4-b511-77931dc6de63\") " pod="openshift-must-gather-xtwkj/crc-debug-74xfj" Mar 18 21:25:54 crc kubenswrapper[4950]: I0318 21:25:54.009250 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0162dc8c-7c36-41e4-b511-77931dc6de63-host\") pod \"crc-debug-74xfj\" (UID: \"0162dc8c-7c36-41e4-b511-77931dc6de63\") " pod="openshift-must-gather-xtwkj/crc-debug-74xfj" Mar 18 21:25:54 crc kubenswrapper[4950]: I0318 21:25:54.009342 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fvtm\" (UniqueName: \"kubernetes.io/projected/0162dc8c-7c36-41e4-b511-77931dc6de63-kube-api-access-7fvtm\") pod \"crc-debug-74xfj\" (UID: \"0162dc8c-7c36-41e4-b511-77931dc6de63\") " pod="openshift-must-gather-xtwkj/crc-debug-74xfj" Mar 18 21:25:54 crc kubenswrapper[4950]: I0318 21:25:54.009367 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0162dc8c-7c36-41e4-b511-77931dc6de63-host\") pod \"crc-debug-74xfj\" (UID: \"0162dc8c-7c36-41e4-b511-77931dc6de63\") " pod="openshift-must-gather-xtwkj/crc-debug-74xfj" Mar 18 21:25:54 crc kubenswrapper[4950]: I0318 21:25:54.027473 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fvtm\" (UniqueName: \"kubernetes.io/projected/0162dc8c-7c36-41e4-b511-77931dc6de63-kube-api-access-7fvtm\") pod \"crc-debug-74xfj\" (UID: \"0162dc8c-7c36-41e4-b511-77931dc6de63\") " pod="openshift-must-gather-xtwkj/crc-debug-74xfj" Mar 18 21:25:54 crc kubenswrapper[4950]: I0318 21:25:54.071657 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xtwkj/crc-debug-74xfj" Mar 18 21:25:54 crc kubenswrapper[4950]: W0318 21:25:54.120675 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0162dc8c_7c36_41e4_b511_77931dc6de63.slice/crio-b1594eab07ae332eb6e44e3976ca50c4a8961f49b872f7cc9417df01c2d838e9 WatchSource:0}: Error finding container b1594eab07ae332eb6e44e3976ca50c4a8961f49b872f7cc9417df01c2d838e9: Status 404 returned error can't find the container with id b1594eab07ae332eb6e44e3976ca50c4a8961f49b872f7cc9417df01c2d838e9 Mar 18 21:25:54 crc kubenswrapper[4950]: I0318 21:25:54.438058 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xtwkj/crc-debug-74xfj" event={"ID":"0162dc8c-7c36-41e4-b511-77931dc6de63","Type":"ContainerStarted","Data":"4ce38ba885f04b637b38c80e06128e380789b6b003b7426db517ac2fce18ce75"} Mar 18 21:25:54 crc kubenswrapper[4950]: I0318 21:25:54.438499 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xtwkj/crc-debug-74xfj" event={"ID":"0162dc8c-7c36-41e4-b511-77931dc6de63","Type":"ContainerStarted","Data":"b1594eab07ae332eb6e44e3976ca50c4a8961f49b872f7cc9417df01c2d838e9"} Mar 18 21:25:54 crc kubenswrapper[4950]: I0318 21:25:54.464892 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xtwkj/crc-debug-74xfj" podStartSLOduration=1.46487172 podStartE2EDuration="1.46487172s" podCreationTimestamp="2026-03-18 21:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 21:25:54.451933211 +0000 UTC m=+4767.692775079" watchObservedRunningTime="2026-03-18 21:25:54.46487172 +0000 UTC m=+4767.705713598" Mar 18 21:26:00 crc kubenswrapper[4950]: I0318 21:26:00.141873 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564486-4qpb9"] Mar 18 21:26:00 crc kubenswrapper[4950]: I0318 21:26:00.144043 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564486-4qpb9" Mar 18 21:26:00 crc kubenswrapper[4950]: I0318 21:26:00.148427 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 21:26:00 crc kubenswrapper[4950]: I0318 21:26:00.148663 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 21:26:00 crc kubenswrapper[4950]: I0318 21:26:00.149250 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 21:26:00 crc kubenswrapper[4950]: I0318 21:26:00.153772 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564486-4qpb9"] Mar 18 21:26:00 crc kubenswrapper[4950]: I0318 21:26:00.228439 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zjmq\" (UniqueName: \"kubernetes.io/projected/be5f2e02-ab24-4e5f-8a15-eb438e75c46c-kube-api-access-5zjmq\") pod \"auto-csr-approver-29564486-4qpb9\" (UID: \"be5f2e02-ab24-4e5f-8a15-eb438e75c46c\") " pod="openshift-infra/auto-csr-approver-29564486-4qpb9" Mar 18 21:26:00 crc kubenswrapper[4950]: I0318 21:26:00.330376 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zjmq\" (UniqueName: \"kubernetes.io/projected/be5f2e02-ab24-4e5f-8a15-eb438e75c46c-kube-api-access-5zjmq\") pod \"auto-csr-approver-29564486-4qpb9\" (UID: \"be5f2e02-ab24-4e5f-8a15-eb438e75c46c\") " pod="openshift-infra/auto-csr-approver-29564486-4qpb9" Mar 18 21:26:00 crc kubenswrapper[4950]: I0318 21:26:00.438231 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zjmq\" (UniqueName: \"kubernetes.io/projected/be5f2e02-ab24-4e5f-8a15-eb438e75c46c-kube-api-access-5zjmq\") pod \"auto-csr-approver-29564486-4qpb9\" (UID: \"be5f2e02-ab24-4e5f-8a15-eb438e75c46c\") " pod="openshift-infra/auto-csr-approver-29564486-4qpb9" Mar 18 21:26:00 crc kubenswrapper[4950]: I0318 21:26:00.463952 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564486-4qpb9" Mar 18 21:26:01 crc kubenswrapper[4950]: I0318 21:26:01.018461 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564486-4qpb9"] Mar 18 21:26:01 crc kubenswrapper[4950]: I0318 21:26:01.496440 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564486-4qpb9" event={"ID":"be5f2e02-ab24-4e5f-8a15-eb438e75c46c","Type":"ContainerStarted","Data":"1cede7aac6745e178fe8376753b55ff8d3298db6f5e65f434e034830a5a78786"} Mar 18 21:26:03 crc kubenswrapper[4950]: I0318 21:26:03.480263 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:26:03 crc kubenswrapper[4950]: E0318 21:26:03.481852 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:26:03 crc kubenswrapper[4950]: I0318 21:26:03.514297 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564486-4qpb9" event={"ID":"be5f2e02-ab24-4e5f-8a15-eb438e75c46c","Type":"ContainerStarted","Data":"2272ce00289e64e90ff5e828d8a949259a4efa3bd66f8986c377527f8309eca8"} Mar 18 21:26:03 crc kubenswrapper[4950]: I0318 21:26:03.533249 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29564486-4qpb9" podStartSLOduration=2.700763274 podStartE2EDuration="3.53323228s" podCreationTimestamp="2026-03-18 21:26:00 +0000 UTC" firstStartedPulling="2026-03-18 21:26:01.029095065 +0000 UTC m=+4774.269936933" lastFinishedPulling="2026-03-18 21:26:01.861564071 +0000 UTC m=+4775.102405939" observedRunningTime="2026-03-18 21:26:03.524494551 +0000 UTC m=+4776.765336419" watchObservedRunningTime="2026-03-18 21:26:03.53323228 +0000 UTC m=+4776.774074148" Mar 18 21:26:04 crc kubenswrapper[4950]: I0318 21:26:04.523457 4950 generic.go:334] "Generic (PLEG): container finished" podID="be5f2e02-ab24-4e5f-8a15-eb438e75c46c" containerID="2272ce00289e64e90ff5e828d8a949259a4efa3bd66f8986c377527f8309eca8" exitCode=0 Mar 18 21:26:04 crc kubenswrapper[4950]: I0318 21:26:04.523508 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564486-4qpb9" event={"ID":"be5f2e02-ab24-4e5f-8a15-eb438e75c46c","Type":"ContainerDied","Data":"2272ce00289e64e90ff5e828d8a949259a4efa3bd66f8986c377527f8309eca8"} Mar 18 21:26:05 crc kubenswrapper[4950]: I0318 21:26:05.846027 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564486-4qpb9" Mar 18 21:26:06 crc kubenswrapper[4950]: I0318 21:26:06.008285 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zjmq\" (UniqueName: \"kubernetes.io/projected/be5f2e02-ab24-4e5f-8a15-eb438e75c46c-kube-api-access-5zjmq\") pod \"be5f2e02-ab24-4e5f-8a15-eb438e75c46c\" (UID: \"be5f2e02-ab24-4e5f-8a15-eb438e75c46c\") " Mar 18 21:26:06 crc kubenswrapper[4950]: I0318 21:26:06.017625 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be5f2e02-ab24-4e5f-8a15-eb438e75c46c-kube-api-access-5zjmq" (OuterVolumeSpecName: "kube-api-access-5zjmq") pod "be5f2e02-ab24-4e5f-8a15-eb438e75c46c" (UID: "be5f2e02-ab24-4e5f-8a15-eb438e75c46c"). InnerVolumeSpecName "kube-api-access-5zjmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:26:06 crc kubenswrapper[4950]: I0318 21:26:06.110917 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zjmq\" (UniqueName: \"kubernetes.io/projected/be5f2e02-ab24-4e5f-8a15-eb438e75c46c-kube-api-access-5zjmq\") on node \"crc\" DevicePath \"\"" Mar 18 21:26:06 crc kubenswrapper[4950]: I0318 21:26:06.542757 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564486-4qpb9" event={"ID":"be5f2e02-ab24-4e5f-8a15-eb438e75c46c","Type":"ContainerDied","Data":"1cede7aac6745e178fe8376753b55ff8d3298db6f5e65f434e034830a5a78786"} Mar 18 21:26:06 crc kubenswrapper[4950]: I0318 21:26:06.542803 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cede7aac6745e178fe8376753b55ff8d3298db6f5e65f434e034830a5a78786" Mar 18 21:26:06 crc kubenswrapper[4950]: I0318 21:26:06.542863 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564486-4qpb9" Mar 18 21:26:06 crc kubenswrapper[4950]: I0318 21:26:06.647369 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564480-99n7f"] Mar 18 21:26:06 crc kubenswrapper[4950]: I0318 21:26:06.657927 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564480-99n7f"] Mar 18 21:26:07 crc kubenswrapper[4950]: I0318 21:26:07.491343 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abf0abcc-8ac6-4243-82d0-f9d802143e9d" path="/var/lib/kubelet/pods/abf0abcc-8ac6-4243-82d0-f9d802143e9d/volumes" Mar 18 21:26:18 crc kubenswrapper[4950]: I0318 21:26:18.480774 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:26:18 crc kubenswrapper[4950]: E0318 21:26:18.481428 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:26:27 crc kubenswrapper[4950]: I0318 21:26:27.730074 4950 generic.go:334] "Generic (PLEG): container finished" podID="0162dc8c-7c36-41e4-b511-77931dc6de63" containerID="4ce38ba885f04b637b38c80e06128e380789b6b003b7426db517ac2fce18ce75" exitCode=0 Mar 18 21:26:27 crc kubenswrapper[4950]: I0318 21:26:27.730556 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xtwkj/crc-debug-74xfj" event={"ID":"0162dc8c-7c36-41e4-b511-77931dc6de63","Type":"ContainerDied","Data":"4ce38ba885f04b637b38c80e06128e380789b6b003b7426db517ac2fce18ce75"} Mar 18 21:26:28 crc kubenswrapper[4950]: I0318 21:26:28.842628 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xtwkj/crc-debug-74xfj" Mar 18 21:26:28 crc kubenswrapper[4950]: I0318 21:26:28.886051 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xtwkj/crc-debug-74xfj"] Mar 18 21:26:28 crc kubenswrapper[4950]: I0318 21:26:28.897656 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xtwkj/crc-debug-74xfj"] Mar 18 21:26:28 crc kubenswrapper[4950]: I0318 21:26:28.967324 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0162dc8c-7c36-41e4-b511-77931dc6de63-host\") pod \"0162dc8c-7c36-41e4-b511-77931dc6de63\" (UID: \"0162dc8c-7c36-41e4-b511-77931dc6de63\") " Mar 18 21:26:28 crc kubenswrapper[4950]: I0318 21:26:28.967556 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0162dc8c-7c36-41e4-b511-77931dc6de63-host" (OuterVolumeSpecName: "host") pod "0162dc8c-7c36-41e4-b511-77931dc6de63" (UID: "0162dc8c-7c36-41e4-b511-77931dc6de63"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 21:26:28 crc kubenswrapper[4950]: I0318 21:26:28.967615 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fvtm\" (UniqueName: \"kubernetes.io/projected/0162dc8c-7c36-41e4-b511-77931dc6de63-kube-api-access-7fvtm\") pod \"0162dc8c-7c36-41e4-b511-77931dc6de63\" (UID: \"0162dc8c-7c36-41e4-b511-77931dc6de63\") " Mar 18 21:26:28 crc kubenswrapper[4950]: I0318 21:26:28.967991 4950 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0162dc8c-7c36-41e4-b511-77931dc6de63-host\") on node \"crc\" DevicePath \"\"" Mar 18 21:26:28 crc kubenswrapper[4950]: I0318 21:26:28.975680 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0162dc8c-7c36-41e4-b511-77931dc6de63-kube-api-access-7fvtm" (OuterVolumeSpecName: "kube-api-access-7fvtm") pod "0162dc8c-7c36-41e4-b511-77931dc6de63" (UID: "0162dc8c-7c36-41e4-b511-77931dc6de63"). InnerVolumeSpecName "kube-api-access-7fvtm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:26:29 crc kubenswrapper[4950]: I0318 21:26:29.069744 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fvtm\" (UniqueName: \"kubernetes.io/projected/0162dc8c-7c36-41e4-b511-77931dc6de63-kube-api-access-7fvtm\") on node \"crc\" DevicePath \"\"" Mar 18 21:26:29 crc kubenswrapper[4950]: I0318 21:26:29.492994 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0162dc8c-7c36-41e4-b511-77931dc6de63" path="/var/lib/kubelet/pods/0162dc8c-7c36-41e4-b511-77931dc6de63/volumes" Mar 18 21:26:29 crc kubenswrapper[4950]: I0318 21:26:29.761322 4950 scope.go:117] "RemoveContainer" containerID="4ce38ba885f04b637b38c80e06128e380789b6b003b7426db517ac2fce18ce75" Mar 18 21:26:29 crc kubenswrapper[4950]: I0318 21:26:29.761343 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xtwkj/crc-debug-74xfj" Mar 18 21:26:30 crc kubenswrapper[4950]: I0318 21:26:30.084545 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xtwkj/crc-debug-jdvhl"] Mar 18 21:26:30 crc kubenswrapper[4950]: E0318 21:26:30.084917 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0162dc8c-7c36-41e4-b511-77931dc6de63" containerName="container-00" Mar 18 21:26:30 crc kubenswrapper[4950]: I0318 21:26:30.084929 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="0162dc8c-7c36-41e4-b511-77931dc6de63" containerName="container-00" Mar 18 21:26:30 crc kubenswrapper[4950]: E0318 21:26:30.084948 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be5f2e02-ab24-4e5f-8a15-eb438e75c46c" containerName="oc" Mar 18 21:26:30 crc kubenswrapper[4950]: I0318 21:26:30.084954 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="be5f2e02-ab24-4e5f-8a15-eb438e75c46c" containerName="oc" Mar 18 21:26:30 crc kubenswrapper[4950]: I0318 21:26:30.085172 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="be5f2e02-ab24-4e5f-8a15-eb438e75c46c" containerName="oc" Mar 18 21:26:30 crc kubenswrapper[4950]: I0318 21:26:30.085194 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="0162dc8c-7c36-41e4-b511-77931dc6de63" containerName="container-00" Mar 18 21:26:30 crc kubenswrapper[4950]: I0318 21:26:30.085909 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xtwkj/crc-debug-jdvhl" Mar 18 21:26:30 crc kubenswrapper[4950]: I0318 21:26:30.189924 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxc5r\" (UniqueName: \"kubernetes.io/projected/7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e-kube-api-access-mxc5r\") pod \"crc-debug-jdvhl\" (UID: \"7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e\") " pod="openshift-must-gather-xtwkj/crc-debug-jdvhl" Mar 18 21:26:30 crc kubenswrapper[4950]: I0318 21:26:30.189992 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e-host\") pod \"crc-debug-jdvhl\" (UID: \"7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e\") " pod="openshift-must-gather-xtwkj/crc-debug-jdvhl" Mar 18 21:26:30 crc kubenswrapper[4950]: I0318 21:26:30.291982 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxc5r\" (UniqueName: \"kubernetes.io/projected/7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e-kube-api-access-mxc5r\") pod \"crc-debug-jdvhl\" (UID: \"7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e\") " pod="openshift-must-gather-xtwkj/crc-debug-jdvhl" Mar 18 21:26:30 crc kubenswrapper[4950]: I0318 21:26:30.292032 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e-host\") pod \"crc-debug-jdvhl\" (UID: \"7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e\") " pod="openshift-must-gather-xtwkj/crc-debug-jdvhl" Mar 18 21:26:30 crc kubenswrapper[4950]: I0318 21:26:30.292194 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e-host\") pod \"crc-debug-jdvhl\" (UID: \"7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e\") " pod="openshift-must-gather-xtwkj/crc-debug-jdvhl" Mar 18 21:26:30 crc kubenswrapper[4950]: I0318 21:26:30.314134 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxc5r\" (UniqueName: \"kubernetes.io/projected/7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e-kube-api-access-mxc5r\") pod \"crc-debug-jdvhl\" (UID: \"7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e\") " pod="openshift-must-gather-xtwkj/crc-debug-jdvhl" Mar 18 21:26:30 crc kubenswrapper[4950]: I0318 21:26:30.405633 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xtwkj/crc-debug-jdvhl" Mar 18 21:26:30 crc kubenswrapper[4950]: I0318 21:26:30.770456 4950 generic.go:334] "Generic (PLEG): container finished" podID="7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e" containerID="8eee6fc77b0a34b9469ef886a01b7cfd0a60c57cd4f4aa48386289fa1d238f5d" exitCode=0 Mar 18 21:26:30 crc kubenswrapper[4950]: I0318 21:26:30.770543 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xtwkj/crc-debug-jdvhl" event={"ID":"7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e","Type":"ContainerDied","Data":"8eee6fc77b0a34b9469ef886a01b7cfd0a60c57cd4f4aa48386289fa1d238f5d"} Mar 18 21:26:30 crc kubenswrapper[4950]: I0318 21:26:30.770815 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xtwkj/crc-debug-jdvhl" event={"ID":"7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e","Type":"ContainerStarted","Data":"7cdcabdc77fc96518a3b656a8e3c4a5ac27f061f5742ef6da5999bdebab86ef2"} Mar 18 21:26:31 crc kubenswrapper[4950]: I0318 21:26:31.145879 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xtwkj/crc-debug-jdvhl"] Mar 18 21:26:31 crc kubenswrapper[4950]: I0318 21:26:31.150901 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xtwkj/crc-debug-jdvhl"] Mar 18 21:26:31 crc kubenswrapper[4950]: I0318 21:26:31.897291 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xtwkj/crc-debug-jdvhl" Mar 18 21:26:32 crc kubenswrapper[4950]: I0318 21:26:32.026746 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e-host\") pod \"7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e\" (UID: \"7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e\") " Mar 18 21:26:32 crc kubenswrapper[4950]: I0318 21:26:32.026825 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e-host" (OuterVolumeSpecName: "host") pod "7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e" (UID: "7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 21:26:32 crc kubenswrapper[4950]: I0318 21:26:32.027010 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxc5r\" (UniqueName: \"kubernetes.io/projected/7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e-kube-api-access-mxc5r\") pod \"7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e\" (UID: \"7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e\") " Mar 18 21:26:32 crc kubenswrapper[4950]: I0318 21:26:32.027395 4950 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e-host\") on node \"crc\" DevicePath \"\"" Mar 18 21:26:32 crc kubenswrapper[4950]: I0318 21:26:32.033620 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e-kube-api-access-mxc5r" (OuterVolumeSpecName: "kube-api-access-mxc5r") pod "7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e" (UID: "7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e"). InnerVolumeSpecName "kube-api-access-mxc5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:26:32 crc kubenswrapper[4950]: I0318 21:26:32.128862 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxc5r\" (UniqueName: \"kubernetes.io/projected/7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e-kube-api-access-mxc5r\") on node \"crc\" DevicePath \"\"" Mar 18 21:26:32 crc kubenswrapper[4950]: I0318 21:26:32.353808 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xtwkj/crc-debug-pv5st"] Mar 18 21:26:32 crc kubenswrapper[4950]: E0318 21:26:32.355109 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e" containerName="container-00" Mar 18 21:26:32 crc kubenswrapper[4950]: I0318 21:26:32.355129 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e" containerName="container-00" Mar 18 21:26:32 crc kubenswrapper[4950]: I0318 21:26:32.355488 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e" containerName="container-00" Mar 18 21:26:32 crc kubenswrapper[4950]: I0318 21:26:32.364017 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xtwkj/crc-debug-pv5st" Mar 18 21:26:32 crc kubenswrapper[4950]: I0318 21:26:32.537343 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdk8w\" (UniqueName: \"kubernetes.io/projected/a9828f0e-abbb-4488-8605-0c502535ee0d-kube-api-access-qdk8w\") pod \"crc-debug-pv5st\" (UID: \"a9828f0e-abbb-4488-8605-0c502535ee0d\") " pod="openshift-must-gather-xtwkj/crc-debug-pv5st" Mar 18 21:26:32 crc kubenswrapper[4950]: I0318 21:26:32.537458 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a9828f0e-abbb-4488-8605-0c502535ee0d-host\") pod \"crc-debug-pv5st\" (UID: \"a9828f0e-abbb-4488-8605-0c502535ee0d\") " pod="openshift-must-gather-xtwkj/crc-debug-pv5st" Mar 18 21:26:32 crc kubenswrapper[4950]: I0318 21:26:32.638994 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdk8w\" (UniqueName: \"kubernetes.io/projected/a9828f0e-abbb-4488-8605-0c502535ee0d-kube-api-access-qdk8w\") pod \"crc-debug-pv5st\" (UID: \"a9828f0e-abbb-4488-8605-0c502535ee0d\") " pod="openshift-must-gather-xtwkj/crc-debug-pv5st" Mar 18 21:26:32 crc kubenswrapper[4950]: I0318 21:26:32.639082 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a9828f0e-abbb-4488-8605-0c502535ee0d-host\") pod \"crc-debug-pv5st\" (UID: \"a9828f0e-abbb-4488-8605-0c502535ee0d\") " pod="openshift-must-gather-xtwkj/crc-debug-pv5st" Mar 18 21:26:32 crc kubenswrapper[4950]: I0318 21:26:32.639436 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a9828f0e-abbb-4488-8605-0c502535ee0d-host\") pod \"crc-debug-pv5st\" (UID: \"a9828f0e-abbb-4488-8605-0c502535ee0d\") " pod="openshift-must-gather-xtwkj/crc-debug-pv5st" Mar 18 21:26:32 crc kubenswrapper[4950]: I0318 21:26:32.665399 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdk8w\" (UniqueName: \"kubernetes.io/projected/a9828f0e-abbb-4488-8605-0c502535ee0d-kube-api-access-qdk8w\") pod \"crc-debug-pv5st\" (UID: \"a9828f0e-abbb-4488-8605-0c502535ee0d\") " pod="openshift-must-gather-xtwkj/crc-debug-pv5st" Mar 18 21:26:32 crc kubenswrapper[4950]: I0318 21:26:32.687325 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xtwkj/crc-debug-pv5st" Mar 18 21:26:32 crc kubenswrapper[4950]: W0318 21:26:32.727683 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9828f0e_abbb_4488_8605_0c502535ee0d.slice/crio-9abec6211018621f912ac3d89cf3d28e3bbc491592e9f76d76f71d3e1aa338a1 WatchSource:0}: Error finding container 9abec6211018621f912ac3d89cf3d28e3bbc491592e9f76d76f71d3e1aa338a1: Status 404 returned error can't find the container with id 9abec6211018621f912ac3d89cf3d28e3bbc491592e9f76d76f71d3e1aa338a1 Mar 18 21:26:32 crc kubenswrapper[4950]: I0318 21:26:32.795161 4950 scope.go:117] "RemoveContainer" containerID="8eee6fc77b0a34b9469ef886a01b7cfd0a60c57cd4f4aa48386289fa1d238f5d" Mar 18 21:26:32 crc kubenswrapper[4950]: I0318 21:26:32.795607 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xtwkj/crc-debug-jdvhl" Mar 18 21:26:32 crc kubenswrapper[4950]: I0318 21:26:32.811596 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xtwkj/crc-debug-pv5st" event={"ID":"a9828f0e-abbb-4488-8605-0c502535ee0d","Type":"ContainerStarted","Data":"9abec6211018621f912ac3d89cf3d28e3bbc491592e9f76d76f71d3e1aa338a1"} Mar 18 21:26:32 crc kubenswrapper[4950]: E0318 21:26:32.994381 4950 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7844e6f4_a7c9_4cb4_a5f3_dbd1182e4a8e.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7844e6f4_a7c9_4cb4_a5f3_dbd1182e4a8e.slice/crio-7cdcabdc77fc96518a3b656a8e3c4a5ac27f061f5742ef6da5999bdebab86ef2\": RecentStats: unable to find data in memory cache]" Mar 18 21:26:33 crc kubenswrapper[4950]: I0318 21:26:33.486094 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:26:33 crc kubenswrapper[4950]: E0318 21:26:33.486607 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:26:33 crc kubenswrapper[4950]: I0318 21:26:33.493995 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e" path="/var/lib/kubelet/pods/7844e6f4-a7c9-4cb4-a5f3-dbd1182e4a8e/volumes" Mar 18 21:26:33 crc kubenswrapper[4950]: I0318 21:26:33.825997 4950 generic.go:334] "Generic (PLEG): container finished" podID="a9828f0e-abbb-4488-8605-0c502535ee0d" containerID="a69445727515ad6e57274e43ec65eaccd4a837b7ed1223c84cf78e5bfc166a9f" exitCode=0 Mar 18 21:26:33 crc kubenswrapper[4950]: I0318 21:26:33.826091 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xtwkj/crc-debug-pv5st" event={"ID":"a9828f0e-abbb-4488-8605-0c502535ee0d","Type":"ContainerDied","Data":"a69445727515ad6e57274e43ec65eaccd4a837b7ed1223c84cf78e5bfc166a9f"} Mar 18 21:26:33 crc kubenswrapper[4950]: I0318 21:26:33.884113 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xtwkj/crc-debug-pv5st"] Mar 18 21:26:33 crc kubenswrapper[4950]: I0318 21:26:33.893079 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xtwkj/crc-debug-pv5st"] Mar 18 21:26:34 crc kubenswrapper[4950]: I0318 21:26:34.959989 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xtwkj/crc-debug-pv5st" Mar 18 21:26:35 crc kubenswrapper[4950]: I0318 21:26:35.083059 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a9828f0e-abbb-4488-8605-0c502535ee0d-host\") pod \"a9828f0e-abbb-4488-8605-0c502535ee0d\" (UID: \"a9828f0e-abbb-4488-8605-0c502535ee0d\") " Mar 18 21:26:35 crc kubenswrapper[4950]: I0318 21:26:35.083123 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdk8w\" (UniqueName: \"kubernetes.io/projected/a9828f0e-abbb-4488-8605-0c502535ee0d-kube-api-access-qdk8w\") pod \"a9828f0e-abbb-4488-8605-0c502535ee0d\" (UID: \"a9828f0e-abbb-4488-8605-0c502535ee0d\") " Mar 18 21:26:35 crc kubenswrapper[4950]: I0318 21:26:35.083200 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9828f0e-abbb-4488-8605-0c502535ee0d-host" (OuterVolumeSpecName: "host") pod "a9828f0e-abbb-4488-8605-0c502535ee0d" (UID: "a9828f0e-abbb-4488-8605-0c502535ee0d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 18 21:26:35 crc kubenswrapper[4950]: I0318 21:26:35.083877 4950 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a9828f0e-abbb-4488-8605-0c502535ee0d-host\") on node \"crc\" DevicePath \"\"" Mar 18 21:26:35 crc kubenswrapper[4950]: I0318 21:26:35.088369 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9828f0e-abbb-4488-8605-0c502535ee0d-kube-api-access-qdk8w" (OuterVolumeSpecName: "kube-api-access-qdk8w") pod "a9828f0e-abbb-4488-8605-0c502535ee0d" (UID: "a9828f0e-abbb-4488-8605-0c502535ee0d"). InnerVolumeSpecName "kube-api-access-qdk8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:26:35 crc kubenswrapper[4950]: I0318 21:26:35.186051 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdk8w\" (UniqueName: \"kubernetes.io/projected/a9828f0e-abbb-4488-8605-0c502535ee0d-kube-api-access-qdk8w\") on node \"crc\" DevicePath \"\"" Mar 18 21:26:35 crc kubenswrapper[4950]: I0318 21:26:35.490372 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9828f0e-abbb-4488-8605-0c502535ee0d" path="/var/lib/kubelet/pods/a9828f0e-abbb-4488-8605-0c502535ee0d/volumes" Mar 18 21:26:35 crc kubenswrapper[4950]: I0318 21:26:35.856278 4950 scope.go:117] "RemoveContainer" containerID="a69445727515ad6e57274e43ec65eaccd4a837b7ed1223c84cf78e5bfc166a9f" Mar 18 21:26:35 crc kubenswrapper[4950]: I0318 21:26:35.856333 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xtwkj/crc-debug-pv5st" Mar 18 21:26:45 crc kubenswrapper[4950]: I0318 21:26:45.480373 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:26:45 crc kubenswrapper[4950]: E0318 21:26:45.482696 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:26:56 crc kubenswrapper[4950]: I0318 21:26:56.480866 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:26:56 crc kubenswrapper[4950]: E0318 21:26:56.482707 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:27:02 crc kubenswrapper[4950]: I0318 21:27:02.176886 4950 scope.go:117] "RemoveContainer" containerID="6a89bf9cb437ebebe4c9dd52ea94bd2c063db1326deac67c8addde33c042a41c" Mar 18 21:27:03 crc kubenswrapper[4950]: I0318 21:27:03.420927 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sq2kk"] Mar 18 21:27:03 crc kubenswrapper[4950]: E0318 21:27:03.423725 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9828f0e-abbb-4488-8605-0c502535ee0d" containerName="container-00" Mar 18 21:27:03 crc kubenswrapper[4950]: I0318 21:27:03.423746 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9828f0e-abbb-4488-8605-0c502535ee0d" containerName="container-00" Mar 18 21:27:03 crc kubenswrapper[4950]: I0318 21:27:03.423924 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9828f0e-abbb-4488-8605-0c502535ee0d" containerName="container-00" Mar 18 21:27:03 crc kubenswrapper[4950]: I0318 21:27:03.426146 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sq2kk" Mar 18 21:27:03 crc kubenswrapper[4950]: I0318 21:27:03.435251 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sq2kk"] Mar 18 21:27:03 crc kubenswrapper[4950]: I0318 21:27:03.530633 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7830ab64-118c-4a86-a45e-683535ccbc0b-utilities\") pod \"redhat-marketplace-sq2kk\" (UID: \"7830ab64-118c-4a86-a45e-683535ccbc0b\") " pod="openshift-marketplace/redhat-marketplace-sq2kk" Mar 18 21:27:03 crc kubenswrapper[4950]: I0318 21:27:03.530710 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcf94\" (UniqueName: \"kubernetes.io/projected/7830ab64-118c-4a86-a45e-683535ccbc0b-kube-api-access-zcf94\") pod \"redhat-marketplace-sq2kk\" (UID: \"7830ab64-118c-4a86-a45e-683535ccbc0b\") " pod="openshift-marketplace/redhat-marketplace-sq2kk" Mar 18 21:27:03 crc kubenswrapper[4950]: I0318 21:27:03.530865 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7830ab64-118c-4a86-a45e-683535ccbc0b-catalog-content\") pod \"redhat-marketplace-sq2kk\" (UID: \"7830ab64-118c-4a86-a45e-683535ccbc0b\") " pod="openshift-marketplace/redhat-marketplace-sq2kk" Mar 18 21:27:03 crc kubenswrapper[4950]: I0318 21:27:03.632024 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7830ab64-118c-4a86-a45e-683535ccbc0b-utilities\") pod \"redhat-marketplace-sq2kk\" (UID: \"7830ab64-118c-4a86-a45e-683535ccbc0b\") " pod="openshift-marketplace/redhat-marketplace-sq2kk" Mar 18 21:27:03 crc kubenswrapper[4950]: I0318 21:27:03.632089 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcf94\" (UniqueName: \"kubernetes.io/projected/7830ab64-118c-4a86-a45e-683535ccbc0b-kube-api-access-zcf94\") pod \"redhat-marketplace-sq2kk\" (UID: \"7830ab64-118c-4a86-a45e-683535ccbc0b\") " pod="openshift-marketplace/redhat-marketplace-sq2kk" Mar 18 21:27:03 crc kubenswrapper[4950]: I0318 21:27:03.632224 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7830ab64-118c-4a86-a45e-683535ccbc0b-catalog-content\") pod \"redhat-marketplace-sq2kk\" (UID: \"7830ab64-118c-4a86-a45e-683535ccbc0b\") " pod="openshift-marketplace/redhat-marketplace-sq2kk" Mar 18 21:27:03 crc kubenswrapper[4950]: I0318 21:27:03.632740 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7830ab64-118c-4a86-a45e-683535ccbc0b-catalog-content\") pod \"redhat-marketplace-sq2kk\" (UID: \"7830ab64-118c-4a86-a45e-683535ccbc0b\") " pod="openshift-marketplace/redhat-marketplace-sq2kk" Mar 18 21:27:03 crc kubenswrapper[4950]: I0318 21:27:03.632799 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7830ab64-118c-4a86-a45e-683535ccbc0b-utilities\") pod \"redhat-marketplace-sq2kk\" (UID: \"7830ab64-118c-4a86-a45e-683535ccbc0b\") " pod="openshift-marketplace/redhat-marketplace-sq2kk" Mar 18 21:27:03 crc kubenswrapper[4950]: I0318 21:27:03.652315 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcf94\" (UniqueName: \"kubernetes.io/projected/7830ab64-118c-4a86-a45e-683535ccbc0b-kube-api-access-zcf94\") pod \"redhat-marketplace-sq2kk\" (UID: \"7830ab64-118c-4a86-a45e-683535ccbc0b\") " pod="openshift-marketplace/redhat-marketplace-sq2kk" Mar 18 21:27:03 crc kubenswrapper[4950]: I0318 21:27:03.747668 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sq2kk" Mar 18 21:27:04 crc kubenswrapper[4950]: I0318 21:27:04.316528 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sq2kk"] Mar 18 21:27:05 crc kubenswrapper[4950]: I0318 21:27:05.146722 4950 generic.go:334] "Generic (PLEG): container finished" podID="7830ab64-118c-4a86-a45e-683535ccbc0b" containerID="4903a3fc2c24e981ad69764d5b55b4de7f889a536cb0021ff8a4dad413d246aa" exitCode=0 Mar 18 21:27:05 crc kubenswrapper[4950]: I0318 21:27:05.146963 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sq2kk" event={"ID":"7830ab64-118c-4a86-a45e-683535ccbc0b","Type":"ContainerDied","Data":"4903a3fc2c24e981ad69764d5b55b4de7f889a536cb0021ff8a4dad413d246aa"} Mar 18 21:27:05 crc kubenswrapper[4950]: I0318 21:27:05.147052 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sq2kk" event={"ID":"7830ab64-118c-4a86-a45e-683535ccbc0b","Type":"ContainerStarted","Data":"64e9a13bf22b8a33378392d687fd0610cca5412c2fc83be3066daf985fab193b"} Mar 18 21:27:07 crc kubenswrapper[4950]: I0318 21:27:07.165224 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sq2kk" event={"ID":"7830ab64-118c-4a86-a45e-683535ccbc0b","Type":"ContainerStarted","Data":"c59661399fec4533392cd60f096cc7103dadf8a39e947bc003e872fe91f53cbb"} Mar 18 21:27:08 crc kubenswrapper[4950]: I0318 21:27:08.177733 4950 generic.go:334] "Generic (PLEG): container finished" podID="7830ab64-118c-4a86-a45e-683535ccbc0b" containerID="c59661399fec4533392cd60f096cc7103dadf8a39e947bc003e872fe91f53cbb" exitCode=0 Mar 18 21:27:08 crc kubenswrapper[4950]: I0318 21:27:08.177792 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sq2kk" event={"ID":"7830ab64-118c-4a86-a45e-683535ccbc0b","Type":"ContainerDied","Data":"c59661399fec4533392cd60f096cc7103dadf8a39e947bc003e872fe91f53cbb"} Mar 18 21:27:09 crc kubenswrapper[4950]: I0318 21:27:09.199155 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sq2kk" event={"ID":"7830ab64-118c-4a86-a45e-683535ccbc0b","Type":"ContainerStarted","Data":"a798b32a9bf58ec3ac96eed4eb40c79520008da495cf1c1486e0e32894a2bcd2"} Mar 18 21:27:09 crc kubenswrapper[4950]: I0318 21:27:09.224214 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sq2kk" podStartSLOduration=2.79079581 podStartE2EDuration="6.224186409s" podCreationTimestamp="2026-03-18 21:27:03 +0000 UTC" firstStartedPulling="2026-03-18 21:27:05.149090095 +0000 UTC m=+4838.389931983" lastFinishedPulling="2026-03-18 21:27:08.582480674 +0000 UTC m=+4841.823322582" observedRunningTime="2026-03-18 21:27:09.2147081 +0000 UTC m=+4842.455549968" watchObservedRunningTime="2026-03-18 21:27:09.224186409 +0000 UTC m=+4842.465028277" Mar 18 21:27:10 crc kubenswrapper[4950]: I0318 21:27:10.479765 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:27:10 crc kubenswrapper[4950]: E0318 21:27:10.480518 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:27:13 crc kubenswrapper[4950]: I0318 21:27:13.748389 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sq2kk" Mar 18 21:27:13 crc kubenswrapper[4950]: I0318 21:27:13.749344 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sq2kk" Mar 18 21:27:13 crc kubenswrapper[4950]: I0318 21:27:13.841400 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sq2kk" Mar 18 21:27:14 crc kubenswrapper[4950]: I0318 21:27:14.320995 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sq2kk" Mar 18 21:27:14 crc kubenswrapper[4950]: I0318 21:27:14.383131 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sq2kk"] Mar 18 21:27:16 crc kubenswrapper[4950]: I0318 21:27:16.272845 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sq2kk" podUID="7830ab64-118c-4a86-a45e-683535ccbc0b" containerName="registry-server" containerID="cri-o://a798b32a9bf58ec3ac96eed4eb40c79520008da495cf1c1486e0e32894a2bcd2" gracePeriod=2 Mar 18 21:27:16 crc kubenswrapper[4950]: I0318 21:27:16.781326 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sq2kk" Mar 18 21:27:16 crc kubenswrapper[4950]: I0318 21:27:16.967818 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7830ab64-118c-4a86-a45e-683535ccbc0b-utilities\") pod \"7830ab64-118c-4a86-a45e-683535ccbc0b\" (UID: \"7830ab64-118c-4a86-a45e-683535ccbc0b\") " Mar 18 21:27:16 crc kubenswrapper[4950]: I0318 21:27:16.967898 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcf94\" (UniqueName: \"kubernetes.io/projected/7830ab64-118c-4a86-a45e-683535ccbc0b-kube-api-access-zcf94\") pod \"7830ab64-118c-4a86-a45e-683535ccbc0b\" (UID: \"7830ab64-118c-4a86-a45e-683535ccbc0b\") " Mar 18 21:27:16 crc kubenswrapper[4950]: I0318 21:27:16.968061 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7830ab64-118c-4a86-a45e-683535ccbc0b-catalog-content\") pod \"7830ab64-118c-4a86-a45e-683535ccbc0b\" (UID: \"7830ab64-118c-4a86-a45e-683535ccbc0b\") " Mar 18 21:27:16 crc kubenswrapper[4950]: I0318 21:27:16.968716 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7830ab64-118c-4a86-a45e-683535ccbc0b-utilities" (OuterVolumeSpecName: "utilities") pod "7830ab64-118c-4a86-a45e-683535ccbc0b" (UID: "7830ab64-118c-4a86-a45e-683535ccbc0b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:27:16 crc kubenswrapper[4950]: I0318 21:27:16.975017 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7830ab64-118c-4a86-a45e-683535ccbc0b-kube-api-access-zcf94" (OuterVolumeSpecName: "kube-api-access-zcf94") pod "7830ab64-118c-4a86-a45e-683535ccbc0b" (UID: "7830ab64-118c-4a86-a45e-683535ccbc0b"). InnerVolumeSpecName "kube-api-access-zcf94". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:27:16 crc kubenswrapper[4950]: I0318 21:27:16.999058 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7830ab64-118c-4a86-a45e-683535ccbc0b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7830ab64-118c-4a86-a45e-683535ccbc0b" (UID: "7830ab64-118c-4a86-a45e-683535ccbc0b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:27:17 crc kubenswrapper[4950]: I0318 21:27:17.070890 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7830ab64-118c-4a86-a45e-683535ccbc0b-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 21:27:17 crc kubenswrapper[4950]: I0318 21:27:17.071486 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcf94\" (UniqueName: \"kubernetes.io/projected/7830ab64-118c-4a86-a45e-683535ccbc0b-kube-api-access-zcf94\") on node \"crc\" DevicePath \"\"" Mar 18 21:27:17 crc kubenswrapper[4950]: I0318 21:27:17.071499 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7830ab64-118c-4a86-a45e-683535ccbc0b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 21:27:17 crc kubenswrapper[4950]: I0318 21:27:17.284117 4950 generic.go:334] "Generic (PLEG): container finished" podID="7830ab64-118c-4a86-a45e-683535ccbc0b" containerID="a798b32a9bf58ec3ac96eed4eb40c79520008da495cf1c1486e0e32894a2bcd2" exitCode=0 Mar 18 21:27:17 crc kubenswrapper[4950]: I0318 21:27:17.284188 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sq2kk" event={"ID":"7830ab64-118c-4a86-a45e-683535ccbc0b","Type":"ContainerDied","Data":"a798b32a9bf58ec3ac96eed4eb40c79520008da495cf1c1486e0e32894a2bcd2"} Mar 18 21:27:17 crc kubenswrapper[4950]: I0318 21:27:17.284234 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sq2kk" Mar 18 21:27:17 crc kubenswrapper[4950]: I0318 21:27:17.284273 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sq2kk" event={"ID":"7830ab64-118c-4a86-a45e-683535ccbc0b","Type":"ContainerDied","Data":"64e9a13bf22b8a33378392d687fd0610cca5412c2fc83be3066daf985fab193b"} Mar 18 21:27:17 crc kubenswrapper[4950]: I0318 21:27:17.284300 4950 scope.go:117] "RemoveContainer" containerID="a798b32a9bf58ec3ac96eed4eb40c79520008da495cf1c1486e0e32894a2bcd2" Mar 18 21:27:17 crc kubenswrapper[4950]: I0318 21:27:17.310447 4950 scope.go:117] "RemoveContainer" containerID="c59661399fec4533392cd60f096cc7103dadf8a39e947bc003e872fe91f53cbb" Mar 18 21:27:17 crc kubenswrapper[4950]: I0318 21:27:17.337078 4950 scope.go:117] "RemoveContainer" containerID="4903a3fc2c24e981ad69764d5b55b4de7f889a536cb0021ff8a4dad413d246aa" Mar 18 21:27:17 crc kubenswrapper[4950]: I0318 21:27:17.353467 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sq2kk"] Mar 18 21:27:17 crc kubenswrapper[4950]: I0318 21:27:17.369906 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sq2kk"] Mar 18 21:27:17 crc kubenswrapper[4950]: I0318 21:27:17.383262 4950 scope.go:117] "RemoveContainer" containerID="a798b32a9bf58ec3ac96eed4eb40c79520008da495cf1c1486e0e32894a2bcd2" Mar 18 21:27:17 crc kubenswrapper[4950]: E0318 21:27:17.383741 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a798b32a9bf58ec3ac96eed4eb40c79520008da495cf1c1486e0e32894a2bcd2\": container with ID starting with a798b32a9bf58ec3ac96eed4eb40c79520008da495cf1c1486e0e32894a2bcd2 not found: ID does not exist" containerID="a798b32a9bf58ec3ac96eed4eb40c79520008da495cf1c1486e0e32894a2bcd2" Mar 18 21:27:17 crc kubenswrapper[4950]: I0318 21:27:17.383794 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a798b32a9bf58ec3ac96eed4eb40c79520008da495cf1c1486e0e32894a2bcd2"} err="failed to get container status \"a798b32a9bf58ec3ac96eed4eb40c79520008da495cf1c1486e0e32894a2bcd2\": rpc error: code = NotFound desc = could not find container \"a798b32a9bf58ec3ac96eed4eb40c79520008da495cf1c1486e0e32894a2bcd2\": container with ID starting with a798b32a9bf58ec3ac96eed4eb40c79520008da495cf1c1486e0e32894a2bcd2 not found: ID does not exist" Mar 18 21:27:17 crc kubenswrapper[4950]: I0318 21:27:17.383825 4950 scope.go:117] "RemoveContainer" containerID="c59661399fec4533392cd60f096cc7103dadf8a39e947bc003e872fe91f53cbb" Mar 18 21:27:17 crc kubenswrapper[4950]: E0318 21:27:17.384197 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c59661399fec4533392cd60f096cc7103dadf8a39e947bc003e872fe91f53cbb\": container with ID starting with c59661399fec4533392cd60f096cc7103dadf8a39e947bc003e872fe91f53cbb not found: ID does not exist" containerID="c59661399fec4533392cd60f096cc7103dadf8a39e947bc003e872fe91f53cbb" Mar 18 21:27:17 crc kubenswrapper[4950]: I0318 21:27:17.384224 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c59661399fec4533392cd60f096cc7103dadf8a39e947bc003e872fe91f53cbb"} err="failed to get container status \"c59661399fec4533392cd60f096cc7103dadf8a39e947bc003e872fe91f53cbb\": rpc error: code = NotFound desc = could not find container \"c59661399fec4533392cd60f096cc7103dadf8a39e947bc003e872fe91f53cbb\": container with ID starting with c59661399fec4533392cd60f096cc7103dadf8a39e947bc003e872fe91f53cbb not found: ID does not exist" Mar 18 21:27:17 crc kubenswrapper[4950]: I0318 21:27:17.384240 4950 scope.go:117] "RemoveContainer" containerID="4903a3fc2c24e981ad69764d5b55b4de7f889a536cb0021ff8a4dad413d246aa" Mar 18 21:27:17 crc kubenswrapper[4950]: E0318 21:27:17.384737 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4903a3fc2c24e981ad69764d5b55b4de7f889a536cb0021ff8a4dad413d246aa\": container with ID starting with 4903a3fc2c24e981ad69764d5b55b4de7f889a536cb0021ff8a4dad413d246aa not found: ID does not exist" containerID="4903a3fc2c24e981ad69764d5b55b4de7f889a536cb0021ff8a4dad413d246aa" Mar 18 21:27:17 crc kubenswrapper[4950]: I0318 21:27:17.384758 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4903a3fc2c24e981ad69764d5b55b4de7f889a536cb0021ff8a4dad413d246aa"} err="failed to get container status \"4903a3fc2c24e981ad69764d5b55b4de7f889a536cb0021ff8a4dad413d246aa\": rpc error: code = NotFound desc = could not find container \"4903a3fc2c24e981ad69764d5b55b4de7f889a536cb0021ff8a4dad413d246aa\": container with ID starting with 4903a3fc2c24e981ad69764d5b55b4de7f889a536cb0021ff8a4dad413d246aa not found: ID does not exist" Mar 18 21:27:17 crc kubenswrapper[4950]: I0318 21:27:17.492305 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7830ab64-118c-4a86-a45e-683535ccbc0b" path="/var/lib/kubelet/pods/7830ab64-118c-4a86-a45e-683535ccbc0b/volumes" Mar 18 21:27:25 crc kubenswrapper[4950]: I0318 21:27:25.481510 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:27:25 crc kubenswrapper[4950]: E0318 21:27:25.483044 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:27:36 crc kubenswrapper[4950]: I0318 21:27:36.481004 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:27:36 crc kubenswrapper[4950]: E0318 21:27:36.482166 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:27:47 crc kubenswrapper[4950]: I0318 21:27:47.305172 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7d5775f6d4-lshhl_f111df34-12e9-441b-a362-e94471ec8eaf/barbican-api/0.log" Mar 18 21:27:47 crc kubenswrapper[4950]: I0318 21:27:47.532462 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-546478654b-mwfqm_72fe734a-54a6-4391-a125-04ddd4c9c890/barbican-keystone-listener/0.log" Mar 18 21:27:47 crc kubenswrapper[4950]: I0318 21:27:47.583676 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7d5775f6d4-lshhl_f111df34-12e9-441b-a362-e94471ec8eaf/barbican-api-log/0.log" Mar 18 21:27:47 crc kubenswrapper[4950]: I0318 21:27:47.690025 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-546478654b-mwfqm_72fe734a-54a6-4391-a125-04ddd4c9c890/barbican-keystone-listener-log/0.log" Mar 18 21:27:47 crc kubenswrapper[4950]: I0318 21:27:47.810351 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6c48df686c-lz5gl_f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9/barbican-worker/0.log" Mar 18 21:27:47 crc kubenswrapper[4950]: I0318 21:27:47.847732 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6c48df686c-lz5gl_f7b1dfa0-7a31-413a-a5ef-a44b0141b4d9/barbican-worker-log/0.log" Mar 18 21:27:48 crc kubenswrapper[4950]: I0318 21:27:48.079214 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-fvghf_8230645b-e426-4c0e-a28b-565ba98d18a1/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:27:48 crc kubenswrapper[4950]: I0318 21:27:48.145766 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b5ef7308-b687-45b3-bcf1-1ef804092712/ceilometer-central-agent/0.log" Mar 18 21:27:48 crc kubenswrapper[4950]: I0318 21:27:48.279597 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b5ef7308-b687-45b3-bcf1-1ef804092712/ceilometer-notification-agent/0.log" Mar 18 21:27:48 crc kubenswrapper[4950]: I0318 21:27:48.320674 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b5ef7308-b687-45b3-bcf1-1ef804092712/proxy-httpd/0.log" Mar 18 21:27:48 crc kubenswrapper[4950]: I0318 21:27:48.327166 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b5ef7308-b687-45b3-bcf1-1ef804092712/sg-core/0.log" Mar 18 21:27:48 crc kubenswrapper[4950]: I0318 21:27:48.479718 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:27:48 crc kubenswrapper[4950]: E0318 21:27:48.480148 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:27:48 crc kubenswrapper[4950]: I0318 21:27:48.488798 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-68nlj_2e43754d-5118-4f9d-afe0-c70c6d4e7589/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:27:48 crc kubenswrapper[4950]: I0318 21:27:48.634077 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-jvb5f_e53816d0-c4c0-412f-9568-af0a96c87145/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:27:48 crc kubenswrapper[4950]: I0318 21:27:48.764643 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_1bae85ae-0545-46ac-90fa-c8f967648ed4/cinder-api/0.log" Mar 18 21:27:48 crc kubenswrapper[4950]: I0318 21:27:48.878672 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_1bae85ae-0545-46ac-90fa-c8f967648ed4/cinder-api-log/0.log" Mar 18 21:27:49 crc kubenswrapper[4950]: I0318 21:27:49.121746 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_9afecfad-52c5-439c-a3bc-a56640e9a79a/probe/0.log" Mar 18 21:27:49 crc kubenswrapper[4950]: I0318 21:27:49.187189 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_9afecfad-52c5-439c-a3bc-a56640e9a79a/cinder-backup/0.log" Mar 18 21:27:49 crc kubenswrapper[4950]: I0318 21:27:49.648997 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_fbb2e2de-55ef-4bc1-a8df-1428e18c664e/cinder-scheduler/0.log" Mar 18 21:27:49 crc kubenswrapper[4950]: I0318 21:27:49.751814 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_fbb2e2de-55ef-4bc1-a8df-1428e18c664e/probe/0.log" Mar 18 21:27:49 crc kubenswrapper[4950]: I0318 21:27:49.817097 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_e22cc0cd-ebf4-4f03-91f3-1db5f2eac881/cinder-volume/0.log" Mar 18 21:27:50 crc kubenswrapper[4950]: I0318 21:27:50.357360 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_e22cc0cd-ebf4-4f03-91f3-1db5f2eac881/probe/0.log" Mar 18 21:27:50 crc kubenswrapper[4950]: I0318 21:27:50.408163 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-vxq4q_f4dc08cb-4f1e-4430-b522-62f890be4336/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:27:50 crc kubenswrapper[4950]: I0318 21:27:50.672321 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-sgzjb_3f18ebe5-e07d-40a7-911e-8d3ad683c78c/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:27:50 crc kubenswrapper[4950]: I0318 21:27:50.791758 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7f55f6cb75-46f5l_f9997bd5-b01a-490e-b4c3-b258b277d73c/init/0.log" Mar 18 21:27:51 crc kubenswrapper[4950]: I0318 21:27:51.081484 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7f55f6cb75-46f5l_f9997bd5-b01a-490e-b4c3-b258b277d73c/init/0.log" Mar 18 21:27:51 crc kubenswrapper[4950]: I0318 21:27:51.174226 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_c8ec3994-fd11-4e42-80c7-01857df19a74/glance-httpd/0.log" Mar 18 21:27:51 crc kubenswrapper[4950]: I0318 21:27:51.237747 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7f55f6cb75-46f5l_f9997bd5-b01a-490e-b4c3-b258b277d73c/dnsmasq-dns/0.log" Mar 18 21:27:51 crc kubenswrapper[4950]: I0318 21:27:51.331333 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_c8ec3994-fd11-4e42-80c7-01857df19a74/glance-log/0.log" Mar 18 21:27:51 crc kubenswrapper[4950]: I0318 21:27:51.462556 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd/glance-httpd/0.log" Mar 18 21:27:51 crc kubenswrapper[4950]: I0318 21:27:51.853925 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_3ca7bb42-8e00-44cf-827d-b6dda7b3e1dd/glance-log/0.log" Mar 18 21:27:51 crc kubenswrapper[4950]: I0318 21:27:51.991832 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7449579958-pdtkd_d648add7-37b3-44fa-a3e2-4ca57d274ca0/horizon/0.log" Mar 18 21:27:52 crc kubenswrapper[4950]: I0318 21:27:52.163678 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7449579958-pdtkd_d648add7-37b3-44fa-a3e2-4ca57d274ca0/horizon-log/0.log" Mar 18 21:27:52 crc kubenswrapper[4950]: I0318 21:27:52.268387 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-472gf_096e1bc2-b416-48d1-bc00-082fb7023df9/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:27:52 crc kubenswrapper[4950]: I0318 21:27:52.361281 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-w5fff_29a51c30-1655-4c6c-9e74-9b8797196361/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:27:52 crc kubenswrapper[4950]: I0318 21:27:52.601726 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29564461-jwfx5_e0d07dee-3b03-4b75-b150-85010f0bcf29/keystone-cron/0.log" Mar 18 21:27:52 crc kubenswrapper[4950]: I0318 21:27:52.733728 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-564b994b87-rbnt8_fb4d97b7-3492-4ccb-8489-98a9882ea782/keystone-api/0.log" Mar 18 21:27:52 crc kubenswrapper[4950]: I0318 21:27:52.856756 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_23a67ca2-5df9-4946-a7d0-9e2dd7c0b91e/kube-state-metrics/0.log" Mar 18 21:27:53 crc kubenswrapper[4950]: I0318 21:27:53.042222 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-n4xjp_22d4a06f-ff3c-4fbf-9eae-8be967f1febd/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:27:53 crc kubenswrapper[4950]: I0318 21:27:53.202016 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_9f52ad06-7e1b-4736-9441-9b1749f6d325/manila-api-log/0.log" Mar 18 21:27:53 crc kubenswrapper[4950]: I0318 21:27:53.208714 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_9f52ad06-7e1b-4736-9441-9b1749f6d325/manila-api/0.log" Mar 18 21:27:53 crc kubenswrapper[4950]: I0318 21:27:53.321460 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_a9bcf400-c9cc-4556-a2e2-6b7214c68579/probe/0.log" Mar 18 21:27:53 crc kubenswrapper[4950]: I0318 21:27:53.400341 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_a9bcf400-c9cc-4556-a2e2-6b7214c68579/manila-scheduler/0.log" Mar 18 21:27:53 crc kubenswrapper[4950]: I0318 21:27:53.511767 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_139d12a8-de25-41a7-a92d-440cbeec98eb/probe/0.log" Mar 18 21:27:53 crc kubenswrapper[4950]: I0318 21:27:53.876788 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_139d12a8-de25-41a7-a92d-440cbeec98eb/manila-share/0.log" Mar 18 21:27:54 crc kubenswrapper[4950]: I0318 21:27:54.100163 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-d5c6d4b4c-6cdzp_2afba89a-006e-472f-8e49-0e8d69cd86c7/neutron-api/0.log" Mar 18 21:27:54 crc kubenswrapper[4950]: I0318 21:27:54.326397 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-d5c6d4b4c-6cdzp_2afba89a-006e-472f-8e49-0e8d69cd86c7/neutron-httpd/0.log" Mar 18 21:27:54 crc kubenswrapper[4950]: I0318 21:27:54.368125 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-ct257_619538d3-bc36-4cfb-8eb7-9fad2bef5d4c/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:27:54 crc kubenswrapper[4950]: I0318 21:27:54.933615 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_5fbcf2da-2f82-4065-85e5-9f701c550afb/nova-cell0-conductor-conductor/0.log" Mar 18 21:27:54 crc kubenswrapper[4950]: I0318 21:27:54.970776 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_603bbfd0-1fdf-44f7-bc34-ce9111ee204f/nova-api-log/0.log" Mar 18 21:27:55 crc kubenswrapper[4950]: I0318 21:27:55.507623 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_49765550-f019-4bdd-8a9c-ffadf57e59c4/nova-cell1-novncproxy-novncproxy/0.log" Mar 18 21:27:55 crc kubenswrapper[4950]: I0318 21:27:55.517717 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_41637a9b-2c22-4627-85df-b902f8d35ceb/nova-cell1-conductor-conductor/0.log" Mar 18 21:27:55 crc kubenswrapper[4950]: I0318 21:27:55.649328 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_603bbfd0-1fdf-44f7-bc34-ce9111ee204f/nova-api-api/0.log" Mar 18 21:27:55 crc kubenswrapper[4950]: I0318 21:27:55.731571 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-v4dhx_006a96ed-5832-49f8-a39c-d433a2b1bfd4/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:27:55 crc kubenswrapper[4950]: I0318 21:27:55.869487 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f05a94b0-defc-4c6c-af5c-3ce8346261a3/nova-metadata-log/0.log" Mar 18 21:27:56 crc kubenswrapper[4950]: I0318 21:27:56.282345 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_aedf1977-d498-4b4f-a92b-f84be888045d/mysql-bootstrap/0.log" Mar 18 21:27:56 crc kubenswrapper[4950]: I0318 21:27:56.395974 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f05a94b0-defc-4c6c-af5c-3ce8346261a3/nova-metadata-metadata/0.log" Mar 18 21:27:56 crc kubenswrapper[4950]: I0318 21:27:56.402378 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_1e49132d-6e98-4270-8ba0-7500e8a84bfd/nova-scheduler-scheduler/0.log" Mar 18 21:27:56 crc kubenswrapper[4950]: I0318 21:27:56.446579 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_aedf1977-d498-4b4f-a92b-f84be888045d/mysql-bootstrap/0.log" Mar 18 21:27:56 crc kubenswrapper[4950]: I0318 21:27:56.531722 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_aedf1977-d498-4b4f-a92b-f84be888045d/galera/0.log" Mar 18 21:27:56 crc kubenswrapper[4950]: I0318 21:27:56.707166 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3/mysql-bootstrap/0.log" Mar 18 21:27:56 crc kubenswrapper[4950]: I0318 21:27:56.895238 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3/mysql-bootstrap/0.log" Mar 18 21:27:56 crc kubenswrapper[4950]: I0318 21:27:56.992029 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e2930ceb-a000-4cb6-bb4e-0f7e8e15cca3/galera/0.log" Mar 18 21:27:57 crc kubenswrapper[4950]: I0318 21:27:57.031143 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_e0941268-2503-4b8e-afeb-0a655b52f505/openstackclient/0.log" Mar 18 21:27:57 crc kubenswrapper[4950]: I0318 21:27:57.279176 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-9wnnb_3f70fe2d-f600-42da-8a3d-a4f814374ca7/openstack-network-exporter/0.log" Mar 18 21:27:57 crc kubenswrapper[4950]: I0318 21:27:57.357446 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lvf7g_21aec953-2e02-48d3-85ff-872d48c4d661/ovsdb-server-init/0.log" Mar 18 21:27:57 crc kubenswrapper[4950]: I0318 21:27:57.522379 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lvf7g_21aec953-2e02-48d3-85ff-872d48c4d661/ovs-vswitchd/0.log" Mar 18 21:27:57 crc kubenswrapper[4950]: I0318 21:27:57.557687 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lvf7g_21aec953-2e02-48d3-85ff-872d48c4d661/ovsdb-server/0.log" Mar 18 21:27:57 crc kubenswrapper[4950]: I0318 21:27:57.566020 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lvf7g_21aec953-2e02-48d3-85ff-872d48c4d661/ovsdb-server-init/0.log" Mar 18 21:27:57 crc kubenswrapper[4950]: I0318 21:27:57.687000 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-zntd5_b72c4ff1-1234-41b7-9faa-b5fd15749084/ovn-controller/0.log" Mar 18 21:27:57 crc kubenswrapper[4950]: I0318 21:27:57.829241 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-hnlg9_67bb928c-890d-480c-8ce7-ec7b299cb1ee/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:27:58 crc kubenswrapper[4950]: I0318 21:27:58.006310 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a3a71276-b0fa-481f-a91e-551d017c6462/openstack-network-exporter/0.log" Mar 18 21:27:58 crc kubenswrapper[4950]: I0318 21:27:58.084396 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a3a71276-b0fa-481f-a91e-551d017c6462/ovn-northd/0.log" Mar 18 21:27:58 crc kubenswrapper[4950]: I0318 21:27:58.234296 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_49c49971-1a60-453f-8283-922ea50cc323/openstack-network-exporter/0.log" Mar 18 21:27:58 crc kubenswrapper[4950]: I0318 21:27:58.726245 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_49c49971-1a60-453f-8283-922ea50cc323/ovsdbserver-nb/0.log" Mar 18 21:27:58 crc kubenswrapper[4950]: I0318 21:27:58.733346 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f37c3f8b-786f-42f6-8c5f-4db367edf6a5/openstack-network-exporter/0.log" Mar 18 21:27:58 crc kubenswrapper[4950]: I0318 21:27:58.796954 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f37c3f8b-786f-42f6-8c5f-4db367edf6a5/ovsdbserver-sb/0.log" Mar 18 21:27:59 crc kubenswrapper[4950]: I0318 21:27:59.049459 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5889fd87f8-6w269_96409976-b947-4ba5-af6e-eeb0ae687b0b/placement-api/0.log" Mar 18 21:27:59 crc kubenswrapper[4950]: I0318 21:27:59.272174 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5889fd87f8-6w269_96409976-b947-4ba5-af6e-eeb0ae687b0b/placement-log/0.log" Mar 18 21:27:59 crc kubenswrapper[4950]: I0318 21:27:59.318814 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_435fb3ca-c4ad-4b0a-a832-88f4a7203def/setup-container/0.log" Mar 18 21:27:59 crc kubenswrapper[4950]: I0318 21:27:59.505111 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_435fb3ca-c4ad-4b0a-a832-88f4a7203def/setup-container/0.log" Mar 18 21:27:59 crc kubenswrapper[4950]: I0318 21:27:59.609793 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_27be57d1-fc6d-4fe7-a6b3-c7276d5ce547/setup-container/0.log" Mar 18 21:27:59 crc kubenswrapper[4950]: I0318 21:27:59.628804 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_435fb3ca-c4ad-4b0a-a832-88f4a7203def/rabbitmq/0.log" Mar 18 21:28:00 crc kubenswrapper[4950]: I0318 21:28:00.154426 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564488-t7mz2"] Mar 18 21:28:00 crc kubenswrapper[4950]: E0318 21:28:00.154813 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7830ab64-118c-4a86-a45e-683535ccbc0b" containerName="registry-server" Mar 18 21:28:00 crc kubenswrapper[4950]: I0318 21:28:00.154826 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="7830ab64-118c-4a86-a45e-683535ccbc0b" containerName="registry-server" Mar 18 21:28:00 crc kubenswrapper[4950]: E0318 21:28:00.154845 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7830ab64-118c-4a86-a45e-683535ccbc0b" containerName="extract-content" Mar 18 21:28:00 crc kubenswrapper[4950]: I0318 21:28:00.154851 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="7830ab64-118c-4a86-a45e-683535ccbc0b" containerName="extract-content" Mar 18 21:28:00 crc kubenswrapper[4950]: E0318 21:28:00.154873 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7830ab64-118c-4a86-a45e-683535ccbc0b" containerName="extract-utilities" Mar 18 21:28:00 crc kubenswrapper[4950]: I0318 21:28:00.154879 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="7830ab64-118c-4a86-a45e-683535ccbc0b" containerName="extract-utilities" Mar 18 21:28:00 crc kubenswrapper[4950]: I0318 21:28:00.155064 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="7830ab64-118c-4a86-a45e-683535ccbc0b" containerName="registry-server" Mar 18 21:28:00 crc kubenswrapper[4950]: I0318 21:28:00.155681 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564488-t7mz2" Mar 18 21:28:00 crc kubenswrapper[4950]: I0318 21:28:00.162014 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 21:28:00 crc kubenswrapper[4950]: I0318 21:28:00.162056 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 21:28:00 crc kubenswrapper[4950]: I0318 21:28:00.162154 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 21:28:00 crc kubenswrapper[4950]: I0318 21:28:00.162362 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564488-t7mz2"] Mar 18 21:28:00 crc kubenswrapper[4950]: I0318 21:28:00.238714 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_27be57d1-fc6d-4fe7-a6b3-c7276d5ce547/rabbitmq/0.log" Mar 18 21:28:00 crc kubenswrapper[4950]: I0318 21:28:00.268194 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdrd4\" (UniqueName: \"kubernetes.io/projected/053e72dd-7249-4665-848b-e1c98eedabb6-kube-api-access-sdrd4\") pod \"auto-csr-approver-29564488-t7mz2\" (UID: \"053e72dd-7249-4665-848b-e1c98eedabb6\") " pod="openshift-infra/auto-csr-approver-29564488-t7mz2" Mar 18 21:28:00 crc kubenswrapper[4950]: I0318 21:28:00.269718 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_27be57d1-fc6d-4fe7-a6b3-c7276d5ce547/setup-container/0.log" Mar 18 21:28:00 crc kubenswrapper[4950]: I0318 21:28:00.281251 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-kjbs4_afdcd712-fe52-40ce-a610-82b7dd39a59c/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:28:00 crc kubenswrapper[4950]: I0318 21:28:00.370520 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdrd4\" (UniqueName: \"kubernetes.io/projected/053e72dd-7249-4665-848b-e1c98eedabb6-kube-api-access-sdrd4\") pod \"auto-csr-approver-29564488-t7mz2\" (UID: \"053e72dd-7249-4665-848b-e1c98eedabb6\") " pod="openshift-infra/auto-csr-approver-29564488-t7mz2" Mar 18 21:28:00 crc kubenswrapper[4950]: I0318 21:28:00.396634 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdrd4\" (UniqueName: \"kubernetes.io/projected/053e72dd-7249-4665-848b-e1c98eedabb6-kube-api-access-sdrd4\") pod \"auto-csr-approver-29564488-t7mz2\" (UID: \"053e72dd-7249-4665-848b-e1c98eedabb6\") " pod="openshift-infra/auto-csr-approver-29564488-t7mz2" Mar 18 21:28:00 crc kubenswrapper[4950]: I0318 21:28:00.470924 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-8r6tq_11355e20-3f52-4010-8c35-0318c88ad131/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:28:00 crc kubenswrapper[4950]: I0318 21:28:00.479720 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:28:00 crc kubenswrapper[4950]: E0318 21:28:00.480180 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:28:00 crc kubenswrapper[4950]: I0318 21:28:00.490063 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564488-t7mz2" Mar 18 21:28:00 crc kubenswrapper[4950]: I0318 21:28:00.594797 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-gsrdb_5e58539e-bccd-459c-b1f7-2696ffe9a402/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:28:00 crc kubenswrapper[4950]: I0318 21:28:00.930613 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-zt89n_d304aae2-4025-4dea-980b-2f3b8c656a8e/ssh-known-hosts-edpm-deployment/0.log" Mar 18 21:28:01 crc kubenswrapper[4950]: I0318 21:28:01.028040 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_3c1bc0a9-e172-4749-b998-710b5a306618/tempest-tests-tempest-tests-runner/0.log" Mar 18 21:28:01 crc kubenswrapper[4950]: I0318 21:28:01.044670 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564488-t7mz2"] Mar 18 21:28:01 crc kubenswrapper[4950]: I0318 21:28:01.233640 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_c5edf1a8-44a7-4cd4-ae5f-f1a77338d35c/test-operator-logs-container/0.log" Mar 18 21:28:01 crc kubenswrapper[4950]: I0318 21:28:01.291808 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-j5hxw_dd54300e-2b25-4bb8-b649-ac546072be73/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 18 21:28:01 crc kubenswrapper[4950]: I0318 21:28:01.706310 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564488-t7mz2" event={"ID":"053e72dd-7249-4665-848b-e1c98eedabb6","Type":"ContainerStarted","Data":"1d9f87e21d046e9b364815c3e21b693cd3f3d5507e8fbd9fbc979932d97cc7af"} Mar 18 21:28:02 crc kubenswrapper[4950]: I0318 21:28:02.721300 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564488-t7mz2" event={"ID":"053e72dd-7249-4665-848b-e1c98eedabb6","Type":"ContainerStarted","Data":"2272e55f06f7d860bacc8aed3a0983a63354f4c1b71f8ba8c387cfa7397aae53"} Mar 18 21:28:02 crc kubenswrapper[4950]: I0318 21:28:02.738012 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29564488-t7mz2" podStartSLOduration=1.659695197 podStartE2EDuration="2.737997279s" podCreationTimestamp="2026-03-18 21:28:00 +0000 UTC" firstStartedPulling="2026-03-18 21:28:01.06027205 +0000 UTC m=+4894.301113918" lastFinishedPulling="2026-03-18 21:28:02.138574132 +0000 UTC m=+4895.379416000" observedRunningTime="2026-03-18 21:28:02.73271558 +0000 UTC m=+4895.973557438" watchObservedRunningTime="2026-03-18 21:28:02.737997279 +0000 UTC m=+4895.978839147" Mar 18 21:28:03 crc kubenswrapper[4950]: I0318 21:28:03.731610 4950 generic.go:334] "Generic (PLEG): container finished" podID="053e72dd-7249-4665-848b-e1c98eedabb6" containerID="2272e55f06f7d860bacc8aed3a0983a63354f4c1b71f8ba8c387cfa7397aae53" exitCode=0 Mar 18 21:28:03 crc kubenswrapper[4950]: I0318 21:28:03.731940 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564488-t7mz2" event={"ID":"053e72dd-7249-4665-848b-e1c98eedabb6","Type":"ContainerDied","Data":"2272e55f06f7d860bacc8aed3a0983a63354f4c1b71f8ba8c387cfa7397aae53"} Mar 18 21:28:05 crc kubenswrapper[4950]: I0318 21:28:05.528822 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564488-t7mz2" Mar 18 21:28:05 crc kubenswrapper[4950]: I0318 21:28:05.700907 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdrd4\" (UniqueName: \"kubernetes.io/projected/053e72dd-7249-4665-848b-e1c98eedabb6-kube-api-access-sdrd4\") pod \"053e72dd-7249-4665-848b-e1c98eedabb6\" (UID: \"053e72dd-7249-4665-848b-e1c98eedabb6\") " Mar 18 21:28:05 crc kubenswrapper[4950]: I0318 21:28:05.730125 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/053e72dd-7249-4665-848b-e1c98eedabb6-kube-api-access-sdrd4" (OuterVolumeSpecName: "kube-api-access-sdrd4") pod "053e72dd-7249-4665-848b-e1c98eedabb6" (UID: "053e72dd-7249-4665-848b-e1c98eedabb6"). InnerVolumeSpecName "kube-api-access-sdrd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:28:05 crc kubenswrapper[4950]: I0318 21:28:05.764222 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564488-t7mz2" event={"ID":"053e72dd-7249-4665-848b-e1c98eedabb6","Type":"ContainerDied","Data":"1d9f87e21d046e9b364815c3e21b693cd3f3d5507e8fbd9fbc979932d97cc7af"} Mar 18 21:28:05 crc kubenswrapper[4950]: I0318 21:28:05.764265 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d9f87e21d046e9b364815c3e21b693cd3f3d5507e8fbd9fbc979932d97cc7af" Mar 18 21:28:05 crc kubenswrapper[4950]: I0318 21:28:05.764319 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564488-t7mz2" Mar 18 21:28:05 crc kubenswrapper[4950]: I0318 21:28:05.803305 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdrd4\" (UniqueName: \"kubernetes.io/projected/053e72dd-7249-4665-848b-e1c98eedabb6-kube-api-access-sdrd4\") on node \"crc\" DevicePath \"\"" Mar 18 21:28:05 crc kubenswrapper[4950]: I0318 21:28:05.808297 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564482-jdb8f"] Mar 18 21:28:05 crc kubenswrapper[4950]: I0318 21:28:05.819788 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564482-jdb8f"] Mar 18 21:28:07 crc kubenswrapper[4950]: I0318 21:28:07.494300 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6490a2c9-0763-4230-9468-69175b459dfd" path="/var/lib/kubelet/pods/6490a2c9-0763-4230-9468-69175b459dfd/volumes" Mar 18 21:28:11 crc kubenswrapper[4950]: I0318 21:28:11.207606 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vnjbq"] Mar 18 21:28:11 crc kubenswrapper[4950]: E0318 21:28:11.208629 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="053e72dd-7249-4665-848b-e1c98eedabb6" containerName="oc" Mar 18 21:28:11 crc kubenswrapper[4950]: I0318 21:28:11.208643 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="053e72dd-7249-4665-848b-e1c98eedabb6" containerName="oc" Mar 18 21:28:11 crc kubenswrapper[4950]: I0318 21:28:11.208846 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="053e72dd-7249-4665-848b-e1c98eedabb6" containerName="oc" Mar 18 21:28:11 crc kubenswrapper[4950]: I0318 21:28:11.210238 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vnjbq" Mar 18 21:28:11 crc kubenswrapper[4950]: I0318 21:28:11.226501 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vnjbq"] Mar 18 21:28:11 crc kubenswrapper[4950]: I0318 21:28:11.315506 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239d3b4c-fa06-4588-9740-10787420ad7f-utilities\") pod \"community-operators-vnjbq\" (UID: \"239d3b4c-fa06-4588-9740-10787420ad7f\") " pod="openshift-marketplace/community-operators-vnjbq" Mar 18 21:28:11 crc kubenswrapper[4950]: I0318 21:28:11.315579 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239d3b4c-fa06-4588-9740-10787420ad7f-catalog-content\") pod \"community-operators-vnjbq\" (UID: \"239d3b4c-fa06-4588-9740-10787420ad7f\") " pod="openshift-marketplace/community-operators-vnjbq" Mar 18 21:28:11 crc kubenswrapper[4950]: I0318 21:28:11.315609 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpjmx\" (UniqueName: \"kubernetes.io/projected/239d3b4c-fa06-4588-9740-10787420ad7f-kube-api-access-qpjmx\") pod \"community-operators-vnjbq\" (UID: \"239d3b4c-fa06-4588-9740-10787420ad7f\") " pod="openshift-marketplace/community-operators-vnjbq" Mar 18 21:28:11 crc kubenswrapper[4950]: I0318 21:28:11.417006 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239d3b4c-fa06-4588-9740-10787420ad7f-catalog-content\") pod \"community-operators-vnjbq\" (UID: \"239d3b4c-fa06-4588-9740-10787420ad7f\") " pod="openshift-marketplace/community-operators-vnjbq" Mar 18 21:28:11 crc kubenswrapper[4950]: I0318 21:28:11.417056 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpjmx\" (UniqueName: \"kubernetes.io/projected/239d3b4c-fa06-4588-9740-10787420ad7f-kube-api-access-qpjmx\") pod \"community-operators-vnjbq\" (UID: \"239d3b4c-fa06-4588-9740-10787420ad7f\") " pod="openshift-marketplace/community-operators-vnjbq" Mar 18 21:28:11 crc kubenswrapper[4950]: I0318 21:28:11.417176 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239d3b4c-fa06-4588-9740-10787420ad7f-utilities\") pod \"community-operators-vnjbq\" (UID: \"239d3b4c-fa06-4588-9740-10787420ad7f\") " pod="openshift-marketplace/community-operators-vnjbq" Mar 18 21:28:11 crc kubenswrapper[4950]: I0318 21:28:11.417570 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239d3b4c-fa06-4588-9740-10787420ad7f-catalog-content\") pod \"community-operators-vnjbq\" (UID: \"239d3b4c-fa06-4588-9740-10787420ad7f\") " pod="openshift-marketplace/community-operators-vnjbq" Mar 18 21:28:11 crc kubenswrapper[4950]: I0318 21:28:11.417653 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239d3b4c-fa06-4588-9740-10787420ad7f-utilities\") pod \"community-operators-vnjbq\" (UID: \"239d3b4c-fa06-4588-9740-10787420ad7f\") " pod="openshift-marketplace/community-operators-vnjbq" Mar 18 21:28:11 crc kubenswrapper[4950]: I0318 21:28:11.643980 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpjmx\" (UniqueName: \"kubernetes.io/projected/239d3b4c-fa06-4588-9740-10787420ad7f-kube-api-access-qpjmx\") pod \"community-operators-vnjbq\" (UID: \"239d3b4c-fa06-4588-9740-10787420ad7f\") " pod="openshift-marketplace/community-operators-vnjbq" Mar 18 21:28:11 crc kubenswrapper[4950]: I0318 21:28:11.846160 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vnjbq" Mar 18 21:28:12 crc kubenswrapper[4950]: I0318 21:28:12.438465 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vnjbq"] Mar 18 21:28:12 crc kubenswrapper[4950]: I0318 21:28:12.488065 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:28:12 crc kubenswrapper[4950]: E0318 21:28:12.488350 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:28:12 crc kubenswrapper[4950]: I0318 21:28:12.848799 4950 generic.go:334] "Generic (PLEG): container finished" podID="239d3b4c-fa06-4588-9740-10787420ad7f" containerID="711962b3e835561141da5452bfb526c967d2fb50bf0a5faee3aea62bbcc42373" exitCode=0 Mar 18 21:28:12 crc kubenswrapper[4950]: I0318 21:28:12.849079 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vnjbq" event={"ID":"239d3b4c-fa06-4588-9740-10787420ad7f","Type":"ContainerDied","Data":"711962b3e835561141da5452bfb526c967d2fb50bf0a5faee3aea62bbcc42373"} Mar 18 21:28:12 crc kubenswrapper[4950]: I0318 21:28:12.849104 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vnjbq" event={"ID":"239d3b4c-fa06-4588-9740-10787420ad7f","Type":"ContainerStarted","Data":"21f9f98758c1a2082ab8d38f974d0e8cbee857c57d6586518646902907b95b59"} Mar 18 21:28:13 crc kubenswrapper[4950]: I0318 21:28:13.866901 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vnjbq" event={"ID":"239d3b4c-fa06-4588-9740-10787420ad7f","Type":"ContainerStarted","Data":"682a44f243c7616374f2fb39f87926355d36f1a86193337e861e8031844d41b2"} Mar 18 21:28:15 crc kubenswrapper[4950]: I0318 21:28:15.887756 4950 generic.go:334] "Generic (PLEG): container finished" podID="239d3b4c-fa06-4588-9740-10787420ad7f" containerID="682a44f243c7616374f2fb39f87926355d36f1a86193337e861e8031844d41b2" exitCode=0 Mar 18 21:28:15 crc kubenswrapper[4950]: I0318 21:28:15.888112 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vnjbq" event={"ID":"239d3b4c-fa06-4588-9740-10787420ad7f","Type":"ContainerDied","Data":"682a44f243c7616374f2fb39f87926355d36f1a86193337e861e8031844d41b2"} Mar 18 21:28:16 crc kubenswrapper[4950]: I0318 21:28:16.814197 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_18eb6f19-11f0-4023-93ac-fc03ecdb9486/memcached/0.log" Mar 18 21:28:16 crc kubenswrapper[4950]: I0318 21:28:16.897242 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vnjbq" event={"ID":"239d3b4c-fa06-4588-9740-10787420ad7f","Type":"ContainerStarted","Data":"796031e755fbd1660ebf5a4541ccdff942120ccd0ac6c9e7f5995df0faeae080"} Mar 18 21:28:16 crc kubenswrapper[4950]: I0318 21:28:16.914608 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vnjbq" podStartSLOduration=2.410414235 podStartE2EDuration="5.91459265s" podCreationTimestamp="2026-03-18 21:28:11 +0000 UTC" firstStartedPulling="2026-03-18 21:28:12.852322992 +0000 UTC m=+4906.093164860" lastFinishedPulling="2026-03-18 21:28:16.356501407 +0000 UTC m=+4909.597343275" observedRunningTime="2026-03-18 21:28:16.911130789 +0000 UTC m=+4910.151972657" watchObservedRunningTime="2026-03-18 21:28:16.91459265 +0000 UTC m=+4910.155434518" Mar 18 21:28:21 crc kubenswrapper[4950]: I0318 21:28:21.847753 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vnjbq" Mar 18 21:28:21 crc kubenswrapper[4950]: I0318 21:28:21.848388 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vnjbq" Mar 18 21:28:22 crc kubenswrapper[4950]: I0318 21:28:22.295889 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vnjbq" Mar 18 21:28:22 crc kubenswrapper[4950]: I0318 21:28:22.380149 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vnjbq" Mar 18 21:28:22 crc kubenswrapper[4950]: I0318 21:28:22.532995 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vnjbq"] Mar 18 21:28:23 crc kubenswrapper[4950]: I0318 21:28:23.957671 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vnjbq" podUID="239d3b4c-fa06-4588-9740-10787420ad7f" containerName="registry-server" containerID="cri-o://796031e755fbd1660ebf5a4541ccdff942120ccd0ac6c9e7f5995df0faeae080" gracePeriod=2 Mar 18 21:28:24 crc kubenswrapper[4950]: I0318 21:28:24.375155 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vnjbq" Mar 18 21:28:24 crc kubenswrapper[4950]: I0318 21:28:24.421125 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpjmx\" (UniqueName: \"kubernetes.io/projected/239d3b4c-fa06-4588-9740-10787420ad7f-kube-api-access-qpjmx\") pod \"239d3b4c-fa06-4588-9740-10787420ad7f\" (UID: \"239d3b4c-fa06-4588-9740-10787420ad7f\") " Mar 18 21:28:24 crc kubenswrapper[4950]: I0318 21:28:24.421611 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239d3b4c-fa06-4588-9740-10787420ad7f-utilities\") pod \"239d3b4c-fa06-4588-9740-10787420ad7f\" (UID: \"239d3b4c-fa06-4588-9740-10787420ad7f\") " Mar 18 21:28:24 crc kubenswrapper[4950]: I0318 21:28:24.421831 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239d3b4c-fa06-4588-9740-10787420ad7f-catalog-content\") pod \"239d3b4c-fa06-4588-9740-10787420ad7f\" (UID: \"239d3b4c-fa06-4588-9740-10787420ad7f\") " Mar 18 21:28:24 crc kubenswrapper[4950]: I0318 21:28:24.425317 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/239d3b4c-fa06-4588-9740-10787420ad7f-utilities" (OuterVolumeSpecName: "utilities") pod "239d3b4c-fa06-4588-9740-10787420ad7f" (UID: "239d3b4c-fa06-4588-9740-10787420ad7f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:28:24 crc kubenswrapper[4950]: I0318 21:28:24.430656 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/239d3b4c-fa06-4588-9740-10787420ad7f-kube-api-access-qpjmx" (OuterVolumeSpecName: "kube-api-access-qpjmx") pod "239d3b4c-fa06-4588-9740-10787420ad7f" (UID: "239d3b4c-fa06-4588-9740-10787420ad7f"). InnerVolumeSpecName "kube-api-access-qpjmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:28:24 crc kubenswrapper[4950]: I0318 21:28:24.479645 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/239d3b4c-fa06-4588-9740-10787420ad7f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "239d3b4c-fa06-4588-9740-10787420ad7f" (UID: "239d3b4c-fa06-4588-9740-10787420ad7f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:28:24 crc kubenswrapper[4950]: I0318 21:28:24.480194 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:28:24 crc kubenswrapper[4950]: E0318 21:28:24.480481 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:28:24 crc kubenswrapper[4950]: I0318 21:28:24.524139 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239d3b4c-fa06-4588-9740-10787420ad7f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 21:28:24 crc kubenswrapper[4950]: I0318 21:28:24.524174 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpjmx\" (UniqueName: \"kubernetes.io/projected/239d3b4c-fa06-4588-9740-10787420ad7f-kube-api-access-qpjmx\") on node \"crc\" DevicePath \"\"" Mar 18 21:28:24 crc kubenswrapper[4950]: I0318 21:28:24.524187 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239d3b4c-fa06-4588-9740-10787420ad7f-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 21:28:24 crc kubenswrapper[4950]: I0318 21:28:24.966005 4950 generic.go:334] "Generic (PLEG): container finished" podID="239d3b4c-fa06-4588-9740-10787420ad7f" containerID="796031e755fbd1660ebf5a4541ccdff942120ccd0ac6c9e7f5995df0faeae080" exitCode=0 Mar 18 21:28:24 crc kubenswrapper[4950]: I0318 21:28:24.966046 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vnjbq" event={"ID":"239d3b4c-fa06-4588-9740-10787420ad7f","Type":"ContainerDied","Data":"796031e755fbd1660ebf5a4541ccdff942120ccd0ac6c9e7f5995df0faeae080"} Mar 18 21:28:24 crc kubenswrapper[4950]: I0318 21:28:24.966070 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vnjbq" event={"ID":"239d3b4c-fa06-4588-9740-10787420ad7f","Type":"ContainerDied","Data":"21f9f98758c1a2082ab8d38f974d0e8cbee857c57d6586518646902907b95b59"} Mar 18 21:28:24 crc kubenswrapper[4950]: I0318 21:28:24.966086 4950 scope.go:117] "RemoveContainer" containerID="796031e755fbd1660ebf5a4541ccdff942120ccd0ac6c9e7f5995df0faeae080" Mar 18 21:28:24 crc kubenswrapper[4950]: I0318 21:28:24.967533 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vnjbq" Mar 18 21:28:24 crc kubenswrapper[4950]: I0318 21:28:24.987496 4950 scope.go:117] "RemoveContainer" containerID="682a44f243c7616374f2fb39f87926355d36f1a86193337e861e8031844d41b2" Mar 18 21:28:25 crc kubenswrapper[4950]: I0318 21:28:25.008367 4950 scope.go:117] "RemoveContainer" containerID="711962b3e835561141da5452bfb526c967d2fb50bf0a5faee3aea62bbcc42373" Mar 18 21:28:25 crc kubenswrapper[4950]: I0318 21:28:25.066653 4950 scope.go:117] "RemoveContainer" containerID="796031e755fbd1660ebf5a4541ccdff942120ccd0ac6c9e7f5995df0faeae080" Mar 18 21:28:25 crc kubenswrapper[4950]: E0318 21:28:25.067132 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"796031e755fbd1660ebf5a4541ccdff942120ccd0ac6c9e7f5995df0faeae080\": container with ID starting with 796031e755fbd1660ebf5a4541ccdff942120ccd0ac6c9e7f5995df0faeae080 not found: ID does not exist" containerID="796031e755fbd1660ebf5a4541ccdff942120ccd0ac6c9e7f5995df0faeae080" Mar 18 21:28:25 crc kubenswrapper[4950]: I0318 21:28:25.067166 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"796031e755fbd1660ebf5a4541ccdff942120ccd0ac6c9e7f5995df0faeae080"} err="failed to get container status \"796031e755fbd1660ebf5a4541ccdff942120ccd0ac6c9e7f5995df0faeae080\": rpc error: code = NotFound desc = could not find container \"796031e755fbd1660ebf5a4541ccdff942120ccd0ac6c9e7f5995df0faeae080\": container with ID starting with 796031e755fbd1660ebf5a4541ccdff942120ccd0ac6c9e7f5995df0faeae080 not found: ID does not exist" Mar 18 21:28:25 crc kubenswrapper[4950]: I0318 21:28:25.067191 4950 scope.go:117] "RemoveContainer" containerID="682a44f243c7616374f2fb39f87926355d36f1a86193337e861e8031844d41b2" Mar 18 21:28:25 crc kubenswrapper[4950]: E0318 21:28:25.067524 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"682a44f243c7616374f2fb39f87926355d36f1a86193337e861e8031844d41b2\": container with ID starting with 682a44f243c7616374f2fb39f87926355d36f1a86193337e861e8031844d41b2 not found: ID does not exist" containerID="682a44f243c7616374f2fb39f87926355d36f1a86193337e861e8031844d41b2" Mar 18 21:28:25 crc kubenswrapper[4950]: I0318 21:28:25.067549 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"682a44f243c7616374f2fb39f87926355d36f1a86193337e861e8031844d41b2"} err="failed to get container status \"682a44f243c7616374f2fb39f87926355d36f1a86193337e861e8031844d41b2\": rpc error: code = NotFound desc = could not find container \"682a44f243c7616374f2fb39f87926355d36f1a86193337e861e8031844d41b2\": container with ID starting with 682a44f243c7616374f2fb39f87926355d36f1a86193337e861e8031844d41b2 not found: ID does not exist" Mar 18 21:28:25 crc kubenswrapper[4950]: I0318 21:28:25.067567 4950 scope.go:117] "RemoveContainer" containerID="711962b3e835561141da5452bfb526c967d2fb50bf0a5faee3aea62bbcc42373" Mar 18 21:28:25 crc kubenswrapper[4950]: E0318 21:28:25.068108 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"711962b3e835561141da5452bfb526c967d2fb50bf0a5faee3aea62bbcc42373\": container with ID starting with 711962b3e835561141da5452bfb526c967d2fb50bf0a5faee3aea62bbcc42373 not found: ID does not exist" containerID="711962b3e835561141da5452bfb526c967d2fb50bf0a5faee3aea62bbcc42373" Mar 18 21:28:25 crc kubenswrapper[4950]: I0318 21:28:25.068166 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"711962b3e835561141da5452bfb526c967d2fb50bf0a5faee3aea62bbcc42373"} err="failed to get container status \"711962b3e835561141da5452bfb526c967d2fb50bf0a5faee3aea62bbcc42373\": rpc error: code = NotFound desc = could not find container \"711962b3e835561141da5452bfb526c967d2fb50bf0a5faee3aea62bbcc42373\": container with ID starting with 711962b3e835561141da5452bfb526c967d2fb50bf0a5faee3aea62bbcc42373 not found: ID does not exist" Mar 18 21:28:25 crc kubenswrapper[4950]: I0318 21:28:25.074894 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vnjbq"] Mar 18 21:28:25 crc kubenswrapper[4950]: I0318 21:28:25.083562 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vnjbq"] Mar 18 21:28:25 crc kubenswrapper[4950]: I0318 21:28:25.490893 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="239d3b4c-fa06-4588-9740-10787420ad7f" path="/var/lib/kubelet/pods/239d3b4c-fa06-4588-9740-10787420ad7f/volumes" Mar 18 21:28:34 crc kubenswrapper[4950]: I0318 21:28:34.307666 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm_0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d/util/0.log" Mar 18 21:28:34 crc kubenswrapper[4950]: I0318 21:28:34.550379 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm_0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d/util/0.log" Mar 18 21:28:34 crc kubenswrapper[4950]: I0318 21:28:34.586587 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm_0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d/pull/0.log" Mar 18 21:28:34 crc kubenswrapper[4950]: I0318 21:28:34.603983 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm_0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d/pull/0.log" Mar 18 21:28:34 crc kubenswrapper[4950]: I0318 21:28:34.795788 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm_0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d/pull/0.log" Mar 18 21:28:34 crc kubenswrapper[4950]: I0318 21:28:34.803147 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm_0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d/util/0.log" Mar 18 21:28:34 crc kubenswrapper[4950]: I0318 21:28:34.882282 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7c80869988bfa7821a7e3d4d9e7801b12993e99d05df1815488a38514c8dvtm_0b6bceb7-5b53-4098-9b3b-d0c4ffa97a7d/extract/0.log" Mar 18 21:28:35 crc kubenswrapper[4950]: I0318 21:28:35.085114 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-59bc569d95-l2xgb_815a0db2-1d7b-4577-b1bf-cf3114f2a2a3/manager/0.log" Mar 18 21:28:35 crc kubenswrapper[4950]: I0318 21:28:35.363324 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-588d4d986b-h9vll_8081f4de-9137-49d2-a66d-bbf265033c5c/manager/0.log" Mar 18 21:28:35 crc kubenswrapper[4950]: I0318 21:28:35.579442 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-79df6bcc97-rgb5v_bc63ecc0-0c62-4607-8ce4-8d143f0c2ce6/manager/0.log" Mar 18 21:28:35 crc kubenswrapper[4950]: I0318 21:28:35.653554 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-67dd5f86f5-6jt4g_9046fce8-e07a-4cc6-9577-4ab38ecac0f3/manager/0.log" Mar 18 21:28:35 crc kubenswrapper[4950]: I0318 21:28:35.988050 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-8464cc45fb-5v45z_926aae26-bb2f-4cac-af12-abbd3585a90a/manager/0.log" Mar 18 21:28:36 crc kubenswrapper[4950]: I0318 21:28:36.305947 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6f787dddc9-kckr9_0277fe72-c2e8-4f07-ae64-bb1232a98518/manager/0.log" Mar 18 21:28:36 crc kubenswrapper[4950]: I0318 21:28:36.479663 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:28:36 crc kubenswrapper[4950]: E0318 21:28:36.479953 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:28:36 crc kubenswrapper[4950]: I0318 21:28:36.689750 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7b9c774f96-6cg5j_df0f2790-244c-4b4f-b3e9-d82e69bb9044/manager/0.log" Mar 18 21:28:36 crc kubenswrapper[4950]: I0318 21:28:36.711976 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-768b96df4c-dls6n_516ce0ee-83c3-40a3-bb00-2a88891a5791/manager/0.log" Mar 18 21:28:36 crc kubenswrapper[4950]: I0318 21:28:36.763612 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-8d58dc466-xlx8b_19c17169-bafd-4a5d-b255-0c50248f3e1c/manager/0.log" Mar 18 21:28:36 crc kubenswrapper[4950]: I0318 21:28:36.966478 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67ccfc9778-5h94g_80caf1bd-ddfd-498d-9b80-ea8cd37a3c72/manager/0.log" Mar 18 21:28:36 crc kubenswrapper[4950]: I0318 21:28:36.981848 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-55f864c847-ftljz_edc5286e-aa7d-4ee6-a8f7-3f5cad640484/manager/0.log" Mar 18 21:28:37 crc kubenswrapper[4950]: I0318 21:28:37.206847 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-767865f676-qxz4b_ec5a5aa1-65cc-4bb2-8644-fa0c7bb44da4/manager/0.log" Mar 18 21:28:37 crc kubenswrapper[4950]: I0318 21:28:37.266917 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5d488d59fb-bsv4c_608c505a-eb24-4cd9-aa6d-e07d4d8edc0c/manager/0.log" Mar 18 21:28:37 crc kubenswrapper[4950]: I0318 21:28:37.451820 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5b9f45d989-pvw7d_29d3abf6-9968-4603-9c38-d3c1809b7275/manager/0.log" Mar 18 21:28:37 crc kubenswrapper[4950]: I0318 21:28:37.462773 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-74c4796899rqqwt_5a502f8b-719d-4f52-812d-c559a8772461/manager/0.log" Mar 18 21:28:37 crc kubenswrapper[4950]: I0318 21:28:37.737057 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-b85c4d696-w5ws4_040697e2-a9b5-4ec0-be8d-f935227c4a0a/operator/0.log" Mar 18 21:28:38 crc kubenswrapper[4950]: I0318 21:28:38.087483 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-884679f54-d6t88_f216b689-dfc9-4042-92f3-71014b3a2698/manager/0.log" Mar 18 21:28:38 crc kubenswrapper[4950]: I0318 21:28:38.163791 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-29kf5_e6006ce5-8d5e-40d2-b9cb-a1a7f501269f/registry-server/0.log" Mar 18 21:28:38 crc kubenswrapper[4950]: I0318 21:28:38.435056 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5784578c99-zcjsh_6b04874c-d0d9-44e7-8047-b0cc2ba6bdfc/manager/0.log" Mar 18 21:28:38 crc kubenswrapper[4950]: I0318 21:28:38.735495 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-7vkgn_283bb5c8-8799-4b57-a3c0-3a8b28f3c2d6/operator/0.log" Mar 18 21:28:38 crc kubenswrapper[4950]: I0318 21:28:38.933393 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-c674c5965-tkbrk_598a0d70-d050-4d4c-8ef5-e274784e5c92/manager/0.log" Mar 18 21:28:39 crc kubenswrapper[4950]: I0318 21:28:39.100190 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-d6b694c5-8k97x_b8e294f2-ac3e-4dbc-b2eb-cd4d2b947789/manager/0.log" Mar 18 21:28:39 crc kubenswrapper[4950]: I0318 21:28:39.212998 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-86bd8996f6-fxpwr_151d2334-a121-46f8-947b-3b17a21abcaf/manager/0.log" Mar 18 21:28:39 crc kubenswrapper[4950]: I0318 21:28:39.315092 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5c5cb9c4d7-wnrmf_e4b560c4-a26f-4acf-bdc3-bb9cabbbd982/manager/0.log" Mar 18 21:28:39 crc kubenswrapper[4950]: I0318 21:28:39.409992 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6c4d75f7f9-v6qqx_e3196503-22ca-4f3a-ac09-f37171f08d37/manager/0.log" Mar 18 21:28:51 crc kubenswrapper[4950]: I0318 21:28:51.479733 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:28:51 crc kubenswrapper[4950]: E0318 21:28:51.480491 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:29:02 crc kubenswrapper[4950]: I0318 21:29:02.062748 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-sxcm9_8596ab2e-852c-4d94-a8ce-e02c823a76b1/control-plane-machine-set-operator/0.log" Mar 18 21:29:02 crc kubenswrapper[4950]: I0318 21:29:02.172263 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-kjb6z_bfd29e9b-fbda-41d0-8300-a0d9337e0e6d/kube-rbac-proxy/0.log" Mar 18 21:29:02 crc kubenswrapper[4950]: I0318 21:29:02.228187 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-kjb6z_bfd29e9b-fbda-41d0-8300-a0d9337e0e6d/machine-api-operator/0.log" Mar 18 21:29:02 crc kubenswrapper[4950]: I0318 21:29:02.357245 4950 scope.go:117] "RemoveContainer" containerID="2192741bafe79f3a7b3c28cc38bf5191e1c685bf0184e65be527946f98234353" Mar 18 21:29:04 crc kubenswrapper[4950]: I0318 21:29:04.480614 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:29:05 crc kubenswrapper[4950]: I0318 21:29:05.306602 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerStarted","Data":"8a3de554c0ec05e3ab570b15a3ddb9444d28e1d8a8dd09556e19d1262de8dcfc"} Mar 18 21:29:16 crc kubenswrapper[4950]: I0318 21:29:16.935711 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-9pk5n_5be9c32f-9b12-4d42-8dd1-3d08455965cc/cert-manager-controller/0.log" Mar 18 21:29:17 crc kubenswrapper[4950]: I0318 21:29:17.163364 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-njln9_f1979115-0bc1-4a2a-9973-72fc24aa570e/cert-manager-cainjector/0.log" Mar 18 21:29:17 crc kubenswrapper[4950]: I0318 21:29:17.180018 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-9gqqw_7eb3a8a5-cac1-4a6f-8dba-2cc19867e55c/cert-manager-webhook/0.log" Mar 18 21:29:30 crc kubenswrapper[4950]: I0318 21:29:30.127180 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-86f58fcf4-jnvkz_0db507f0-dd6f-405c-ae17-4a9e9553cc50/nmstate-console-plugin/0.log" Mar 18 21:29:30 crc kubenswrapper[4950]: I0318 21:29:30.261933 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-6x6mb_61d14d6f-1408-4309-90c9-2971b45d85dc/nmstate-handler/0.log" Mar 18 21:29:30 crc kubenswrapper[4950]: I0318 21:29:30.337559 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-gqrnm_e1b75269-989a-489f-843b-94004e8cdf62/kube-rbac-proxy/0.log" Mar 18 21:29:30 crc kubenswrapper[4950]: I0318 21:29:30.405405 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-gqrnm_e1b75269-989a-489f-843b-94004e8cdf62/nmstate-metrics/0.log" Mar 18 21:29:30 crc kubenswrapper[4950]: I0318 21:29:30.541287 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-796d4cfff4-phx52_de65b778-0bca-479c-826c-e7d935a294c4/nmstate-operator/0.log" Mar 18 21:29:30 crc kubenswrapper[4950]: I0318 21:29:30.639762 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f558f5558-64n89_452b41f3-a30b-4aec-a879-aff3872199d3/nmstate-webhook/0.log" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.016517 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-mhk7w_db65cfec-9483-4641-8794-5fbd50c4bc92/controller/0.log" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.085684 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-mhk7w_db65cfec-9483-4641-8794-5fbd50c4bc92/kube-rbac-proxy/0.log" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.146247 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564490-w6t7b"] Mar 18 21:30:00 crc kubenswrapper[4950]: E0318 21:30:00.146711 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="239d3b4c-fa06-4588-9740-10787420ad7f" containerName="registry-server" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.146730 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="239d3b4c-fa06-4588-9740-10787420ad7f" containerName="registry-server" Mar 18 21:30:00 crc kubenswrapper[4950]: E0318 21:30:00.146758 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="239d3b4c-fa06-4588-9740-10787420ad7f" containerName="extract-utilities" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.146765 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="239d3b4c-fa06-4588-9740-10787420ad7f" containerName="extract-utilities" Mar 18 21:30:00 crc kubenswrapper[4950]: E0318 21:30:00.146785 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="239d3b4c-fa06-4588-9740-10787420ad7f" containerName="extract-content" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.146791 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="239d3b4c-fa06-4588-9740-10787420ad7f" containerName="extract-content" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.146948 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="239d3b4c-fa06-4588-9740-10787420ad7f" containerName="registry-server" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.147641 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564490-w6t7b" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.149255 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.154464 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.154516 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564490-w6t7b"] Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.161811 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.170913 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564490-n55gg"] Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.172367 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564490-n55gg" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.176597 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.176774 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.215518 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564490-n55gg"] Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.273923 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3bc38aaf-a449-4cec-bc8e-1f354080906a-config-volume\") pod \"collect-profiles-29564490-n55gg\" (UID: \"3bc38aaf-a449-4cec-bc8e-1f354080906a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564490-n55gg" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.274105 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3bc38aaf-a449-4cec-bc8e-1f354080906a-secret-volume\") pod \"collect-profiles-29564490-n55gg\" (UID: \"3bc38aaf-a449-4cec-bc8e-1f354080906a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564490-n55gg" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.274527 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzj67\" (UniqueName: \"kubernetes.io/projected/3bc38aaf-a449-4cec-bc8e-1f354080906a-kube-api-access-pzj67\") pod \"collect-profiles-29564490-n55gg\" (UID: \"3bc38aaf-a449-4cec-bc8e-1f354080906a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564490-n55gg" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.274643 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc7xp\" (UniqueName: \"kubernetes.io/projected/f18ea393-c189-46c3-9755-33c306d9dd07-kube-api-access-lc7xp\") pod \"auto-csr-approver-29564490-w6t7b\" (UID: \"f18ea393-c189-46c3-9755-33c306d9dd07\") " pod="openshift-infra/auto-csr-approver-29564490-w6t7b" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.325944 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-frr-files/0.log" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.379784 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3bc38aaf-a449-4cec-bc8e-1f354080906a-secret-volume\") pod \"collect-profiles-29564490-n55gg\" (UID: \"3bc38aaf-a449-4cec-bc8e-1f354080906a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564490-n55gg" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.380111 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzj67\" (UniqueName: \"kubernetes.io/projected/3bc38aaf-a449-4cec-bc8e-1f354080906a-kube-api-access-pzj67\") pod \"collect-profiles-29564490-n55gg\" (UID: \"3bc38aaf-a449-4cec-bc8e-1f354080906a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564490-n55gg" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.380187 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc7xp\" (UniqueName: \"kubernetes.io/projected/f18ea393-c189-46c3-9755-33c306d9dd07-kube-api-access-lc7xp\") pod \"auto-csr-approver-29564490-w6t7b\" (UID: \"f18ea393-c189-46c3-9755-33c306d9dd07\") " pod="openshift-infra/auto-csr-approver-29564490-w6t7b" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.380269 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3bc38aaf-a449-4cec-bc8e-1f354080906a-config-volume\") pod \"collect-profiles-29564490-n55gg\" (UID: \"3bc38aaf-a449-4cec-bc8e-1f354080906a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564490-n55gg" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.381522 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3bc38aaf-a449-4cec-bc8e-1f354080906a-config-volume\") pod \"collect-profiles-29564490-n55gg\" (UID: \"3bc38aaf-a449-4cec-bc8e-1f354080906a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564490-n55gg" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.405334 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzj67\" (UniqueName: \"kubernetes.io/projected/3bc38aaf-a449-4cec-bc8e-1f354080906a-kube-api-access-pzj67\") pod \"collect-profiles-29564490-n55gg\" (UID: \"3bc38aaf-a449-4cec-bc8e-1f354080906a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564490-n55gg" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.414530 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3bc38aaf-a449-4cec-bc8e-1f354080906a-secret-volume\") pod \"collect-profiles-29564490-n55gg\" (UID: \"3bc38aaf-a449-4cec-bc8e-1f354080906a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29564490-n55gg" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.416032 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc7xp\" (UniqueName: \"kubernetes.io/projected/f18ea393-c189-46c3-9755-33c306d9dd07-kube-api-access-lc7xp\") pod \"auto-csr-approver-29564490-w6t7b\" (UID: \"f18ea393-c189-46c3-9755-33c306d9dd07\") " pod="openshift-infra/auto-csr-approver-29564490-w6t7b" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.464074 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564490-w6t7b" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.489149 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564490-n55gg" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.598073 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-frr-files/0.log" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.718881 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-reloader/0.log" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.750835 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-reloader/0.log" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.857142 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-metrics/0.log" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.963239 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-reloader/0.log" Mar 18 21:30:00 crc kubenswrapper[4950]: I0318 21:30:00.999388 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-frr-files/0.log" Mar 18 21:30:01 crc kubenswrapper[4950]: I0318 21:30:01.053100 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-metrics/0.log" Mar 18 21:30:01 crc kubenswrapper[4950]: I0318 21:30:01.103514 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564490-n55gg"] Mar 18 21:30:01 crc kubenswrapper[4950]: I0318 21:30:01.133756 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564490-w6t7b"] Mar 18 21:30:01 crc kubenswrapper[4950]: I0318 21:30:01.143635 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-metrics/0.log" Mar 18 21:30:01 crc kubenswrapper[4950]: W0318 21:30:01.146924 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf18ea393_c189_46c3_9755_33c306d9dd07.slice/crio-ae1b78c2c7f0f2e6784c12118266d18dd4efba075df11f63c94744d783b2bcb8 WatchSource:0}: Error finding container ae1b78c2c7f0f2e6784c12118266d18dd4efba075df11f63c94744d783b2bcb8: Status 404 returned error can't find the container with id ae1b78c2c7f0f2e6784c12118266d18dd4efba075df11f63c94744d783b2bcb8 Mar 18 21:30:01 crc kubenswrapper[4950]: I0318 21:30:01.150695 4950 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 21:30:01 crc kubenswrapper[4950]: I0318 21:30:01.321311 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-reloader/0.log" Mar 18 21:30:01 crc kubenswrapper[4950]: I0318 21:30:01.390762 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/controller/0.log" Mar 18 21:30:01 crc kubenswrapper[4950]: I0318 21:30:01.416497 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-metrics/0.log" Mar 18 21:30:01 crc kubenswrapper[4950]: I0318 21:30:01.456170 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/cp-frr-files/0.log" Mar 18 21:30:01 crc kubenswrapper[4950]: I0318 21:30:01.649034 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/frr-metrics/0.log" Mar 18 21:30:01 crc kubenswrapper[4950]: I0318 21:30:01.717498 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/kube-rbac-proxy/0.log" Mar 18 21:30:01 crc kubenswrapper[4950]: I0318 21:30:01.744235 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/kube-rbac-proxy-frr/0.log" Mar 18 21:30:01 crc kubenswrapper[4950]: I0318 21:30:01.809986 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564490-n55gg" event={"ID":"3bc38aaf-a449-4cec-bc8e-1f354080906a","Type":"ContainerStarted","Data":"c59f85d68f4c479358bd29a50ca7dc85bea1384ebde8ca48330aa9881ea9c1a0"} Mar 18 21:30:01 crc kubenswrapper[4950]: I0318 21:30:01.810183 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564490-n55gg" event={"ID":"3bc38aaf-a449-4cec-bc8e-1f354080906a","Type":"ContainerStarted","Data":"de5665a86d226e526283060a0ee9052475245902027dc0a201b64d24127e641b"} Mar 18 21:30:01 crc kubenswrapper[4950]: I0318 21:30:01.814931 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564490-w6t7b" event={"ID":"f18ea393-c189-46c3-9755-33c306d9dd07","Type":"ContainerStarted","Data":"ae1b78c2c7f0f2e6784c12118266d18dd4efba075df11f63c94744d783b2bcb8"} Mar 18 21:30:01 crc kubenswrapper[4950]: I0318 21:30:01.834421 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29564490-n55gg" podStartSLOduration=1.834390221 podStartE2EDuration="1.834390221s" podCreationTimestamp="2026-03-18 21:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-18 21:30:01.823714662 +0000 UTC m=+5015.064556530" watchObservedRunningTime="2026-03-18 21:30:01.834390221 +0000 UTC m=+5015.075232089" Mar 18 21:30:01 crc kubenswrapper[4950]: I0318 21:30:01.985370 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/reloader/0.log" Mar 18 21:30:02 crc kubenswrapper[4950]: I0318 21:30:02.036167 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-nqppx_b4c0afa4-f131-44df-b2f1-9908b3b578e6/frr-k8s-webhook-server/0.log" Mar 18 21:30:02 crc kubenswrapper[4950]: I0318 21:30:02.280037 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-648849f9d4-9zr5p_871672d1-30f1-4221-a434-2a8d60173f46/manager/0.log" Mar 18 21:30:02 crc kubenswrapper[4950]: I0318 21:30:02.653633 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-bf6cfb5cf-tllp8_210bdf50-0f58-4342-af2c-e0df41ac3e91/webhook-server/0.log" Mar 18 21:30:02 crc kubenswrapper[4950]: I0318 21:30:02.796336 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6k4zn_6f8ab895-b55a-43e6-a6fd-71f4ac42a50a/kube-rbac-proxy/0.log" Mar 18 21:30:02 crc kubenswrapper[4950]: I0318 21:30:02.837429 4950 generic.go:334] "Generic (PLEG): container finished" podID="3bc38aaf-a449-4cec-bc8e-1f354080906a" containerID="c59f85d68f4c479358bd29a50ca7dc85bea1384ebde8ca48330aa9881ea9c1a0" exitCode=0 Mar 18 21:30:02 crc kubenswrapper[4950]: I0318 21:30:02.837509 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564490-n55gg" event={"ID":"3bc38aaf-a449-4cec-bc8e-1f354080906a","Type":"ContainerDied","Data":"c59f85d68f4c479358bd29a50ca7dc85bea1384ebde8ca48330aa9881ea9c1a0"} Mar 18 21:30:02 crc kubenswrapper[4950]: I0318 21:30:02.849603 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564490-w6t7b" event={"ID":"f18ea393-c189-46c3-9755-33c306d9dd07","Type":"ContainerStarted","Data":"f9e920952e609eaa30adaf65f14fd02b48d36b875c45307c599ed594061c2439"} Mar 18 21:30:02 crc kubenswrapper[4950]: I0318 21:30:02.889466 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29564490-w6t7b" podStartSLOduration=1.632730228 podStartE2EDuration="2.889446449s" podCreationTimestamp="2026-03-18 21:30:00 +0000 UTC" firstStartedPulling="2026-03-18 21:30:01.150491456 +0000 UTC m=+5014.391333324" lastFinishedPulling="2026-03-18 21:30:02.407207677 +0000 UTC m=+5015.648049545" observedRunningTime="2026-03-18 21:30:02.87958412 +0000 UTC m=+5016.120425988" watchObservedRunningTime="2026-03-18 21:30:02.889446449 +0000 UTC m=+5016.130288317" Mar 18 21:30:03 crc kubenswrapper[4950]: I0318 21:30:03.242940 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mlhr4_d221ef79-335c-46db-a491-b74bc07300c4/frr/0.log" Mar 18 21:30:03 crc kubenswrapper[4950]: I0318 21:30:03.597954 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6k4zn_6f8ab895-b55a-43e6-a6fd-71f4ac42a50a/speaker/0.log" Mar 18 21:30:04 crc kubenswrapper[4950]: I0318 21:30:04.227596 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564490-n55gg" Mar 18 21:30:04 crc kubenswrapper[4950]: I0318 21:30:04.357868 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzj67\" (UniqueName: \"kubernetes.io/projected/3bc38aaf-a449-4cec-bc8e-1f354080906a-kube-api-access-pzj67\") pod \"3bc38aaf-a449-4cec-bc8e-1f354080906a\" (UID: \"3bc38aaf-a449-4cec-bc8e-1f354080906a\") " Mar 18 21:30:04 crc kubenswrapper[4950]: I0318 21:30:04.357977 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3bc38aaf-a449-4cec-bc8e-1f354080906a-secret-volume\") pod \"3bc38aaf-a449-4cec-bc8e-1f354080906a\" (UID: \"3bc38aaf-a449-4cec-bc8e-1f354080906a\") " Mar 18 21:30:04 crc kubenswrapper[4950]: I0318 21:30:04.358102 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3bc38aaf-a449-4cec-bc8e-1f354080906a-config-volume\") pod \"3bc38aaf-a449-4cec-bc8e-1f354080906a\" (UID: \"3bc38aaf-a449-4cec-bc8e-1f354080906a\") " Mar 18 21:30:04 crc kubenswrapper[4950]: I0318 21:30:04.359048 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3bc38aaf-a449-4cec-bc8e-1f354080906a-config-volume" (OuterVolumeSpecName: "config-volume") pod "3bc38aaf-a449-4cec-bc8e-1f354080906a" (UID: "3bc38aaf-a449-4cec-bc8e-1f354080906a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 18 21:30:04 crc kubenswrapper[4950]: I0318 21:30:04.460684 4950 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3bc38aaf-a449-4cec-bc8e-1f354080906a-config-volume\") on node \"crc\" DevicePath \"\"" Mar 18 21:30:04 crc kubenswrapper[4950]: I0318 21:30:04.548812 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bc38aaf-a449-4cec-bc8e-1f354080906a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3bc38aaf-a449-4cec-bc8e-1f354080906a" (UID: "3bc38aaf-a449-4cec-bc8e-1f354080906a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 18 21:30:04 crc kubenswrapper[4950]: I0318 21:30:04.551225 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bc38aaf-a449-4cec-bc8e-1f354080906a-kube-api-access-pzj67" (OuterVolumeSpecName: "kube-api-access-pzj67") pod "3bc38aaf-a449-4cec-bc8e-1f354080906a" (UID: "3bc38aaf-a449-4cec-bc8e-1f354080906a"). InnerVolumeSpecName "kube-api-access-pzj67". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:30:04 crc kubenswrapper[4950]: I0318 21:30:04.563161 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzj67\" (UniqueName: \"kubernetes.io/projected/3bc38aaf-a449-4cec-bc8e-1f354080906a-kube-api-access-pzj67\") on node \"crc\" DevicePath \"\"" Mar 18 21:30:04 crc kubenswrapper[4950]: I0318 21:30:04.563189 4950 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3bc38aaf-a449-4cec-bc8e-1f354080906a-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 18 21:30:04 crc kubenswrapper[4950]: I0318 21:30:04.871132 4950 generic.go:334] "Generic (PLEG): container finished" podID="f18ea393-c189-46c3-9755-33c306d9dd07" containerID="f9e920952e609eaa30adaf65f14fd02b48d36b875c45307c599ed594061c2439" exitCode=0 Mar 18 21:30:04 crc kubenswrapper[4950]: I0318 21:30:04.871217 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564490-w6t7b" event={"ID":"f18ea393-c189-46c3-9755-33c306d9dd07","Type":"ContainerDied","Data":"f9e920952e609eaa30adaf65f14fd02b48d36b875c45307c599ed594061c2439"} Mar 18 21:30:04 crc kubenswrapper[4950]: I0318 21:30:04.873331 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29564490-n55gg" event={"ID":"3bc38aaf-a449-4cec-bc8e-1f354080906a","Type":"ContainerDied","Data":"de5665a86d226e526283060a0ee9052475245902027dc0a201b64d24127e641b"} Mar 18 21:30:04 crc kubenswrapper[4950]: I0318 21:30:04.873369 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de5665a86d226e526283060a0ee9052475245902027dc0a201b64d24127e641b" Mar 18 21:30:04 crc kubenswrapper[4950]: I0318 21:30:04.873509 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29564490-n55gg" Mar 18 21:30:04 crc kubenswrapper[4950]: I0318 21:30:04.904283 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564445-wtmkn"] Mar 18 21:30:04 crc kubenswrapper[4950]: I0318 21:30:04.911845 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29564445-wtmkn"] Mar 18 21:30:05 crc kubenswrapper[4950]: I0318 21:30:05.493073 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a66e031-b5dd-49e5-91e9-35ea7071c207" path="/var/lib/kubelet/pods/1a66e031-b5dd-49e5-91e9-35ea7071c207/volumes" Mar 18 21:30:07 crc kubenswrapper[4950]: I0318 21:30:07.030719 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564490-w6t7b" Mar 18 21:30:07 crc kubenswrapper[4950]: I0318 21:30:07.116861 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lc7xp\" (UniqueName: \"kubernetes.io/projected/f18ea393-c189-46c3-9755-33c306d9dd07-kube-api-access-lc7xp\") pod \"f18ea393-c189-46c3-9755-33c306d9dd07\" (UID: \"f18ea393-c189-46c3-9755-33c306d9dd07\") " Mar 18 21:30:07 crc kubenswrapper[4950]: I0318 21:30:07.534343 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f18ea393-c189-46c3-9755-33c306d9dd07-kube-api-access-lc7xp" (OuterVolumeSpecName: "kube-api-access-lc7xp") pod "f18ea393-c189-46c3-9755-33c306d9dd07" (UID: "f18ea393-c189-46c3-9755-33c306d9dd07"). InnerVolumeSpecName "kube-api-access-lc7xp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:30:07 crc kubenswrapper[4950]: I0318 21:30:07.630263 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lc7xp\" (UniqueName: \"kubernetes.io/projected/f18ea393-c189-46c3-9755-33c306d9dd07-kube-api-access-lc7xp\") on node \"crc\" DevicePath \"\"" Mar 18 21:30:07 crc kubenswrapper[4950]: I0318 21:30:07.922390 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564490-w6t7b" event={"ID":"f18ea393-c189-46c3-9755-33c306d9dd07","Type":"ContainerDied","Data":"ae1b78c2c7f0f2e6784c12118266d18dd4efba075df11f63c94744d783b2bcb8"} Mar 18 21:30:07 crc kubenswrapper[4950]: I0318 21:30:07.922516 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae1b78c2c7f0f2e6784c12118266d18dd4efba075df11f63c94744d783b2bcb8" Mar 18 21:30:07 crc kubenswrapper[4950]: I0318 21:30:07.922677 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564490-w6t7b" Mar 18 21:30:08 crc kubenswrapper[4950]: I0318 21:30:08.087128 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564484-6vdc2"] Mar 18 21:30:08 crc kubenswrapper[4950]: I0318 21:30:08.094747 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564484-6vdc2"] Mar 18 21:30:09 crc kubenswrapper[4950]: I0318 21:30:09.497657 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="603b946e-91e5-49e1-b6e6-1c983bce0741" path="/var/lib/kubelet/pods/603b946e-91e5-49e1-b6e6-1c983bce0741/volumes" Mar 18 21:30:19 crc kubenswrapper[4950]: I0318 21:30:19.684381 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k_534d37ef-a537-4ea3-832e-4609580d9f47/util/0.log" Mar 18 21:30:19 crc kubenswrapper[4950]: I0318 21:30:19.882655 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k_534d37ef-a537-4ea3-832e-4609580d9f47/util/0.log" Mar 18 21:30:19 crc kubenswrapper[4950]: I0318 21:30:19.918151 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k_534d37ef-a537-4ea3-832e-4609580d9f47/pull/0.log" Mar 18 21:30:19 crc kubenswrapper[4950]: I0318 21:30:19.940328 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k_534d37ef-a537-4ea3-832e-4609580d9f47/pull/0.log" Mar 18 21:30:20 crc kubenswrapper[4950]: I0318 21:30:20.103972 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k_534d37ef-a537-4ea3-832e-4609580d9f47/util/0.log" Mar 18 21:30:20 crc kubenswrapper[4950]: I0318 21:30:20.121880 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k_534d37ef-a537-4ea3-832e-4609580d9f47/pull/0.log" Mar 18 21:30:20 crc kubenswrapper[4950]: I0318 21:30:20.155590 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874hfk2k_534d37ef-a537-4ea3-832e-4609580d9f47/extract/0.log" Mar 18 21:30:20 crc kubenswrapper[4950]: I0318 21:30:20.320855 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk_c955c23a-4fa9-4745-bb00-05a54ae897f4/util/0.log" Mar 18 21:30:20 crc kubenswrapper[4950]: I0318 21:30:20.479095 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk_c955c23a-4fa9-4745-bb00-05a54ae897f4/util/0.log" Mar 18 21:30:20 crc kubenswrapper[4950]: I0318 21:30:20.487191 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk_c955c23a-4fa9-4745-bb00-05a54ae897f4/pull/0.log" Mar 18 21:30:20 crc kubenswrapper[4950]: I0318 21:30:20.518353 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk_c955c23a-4fa9-4745-bb00-05a54ae897f4/pull/0.log" Mar 18 21:30:20 crc kubenswrapper[4950]: I0318 21:30:20.692669 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk_c955c23a-4fa9-4745-bb00-05a54ae897f4/util/0.log" Mar 18 21:30:20 crc kubenswrapper[4950]: I0318 21:30:20.701072 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk_c955c23a-4fa9-4745-bb00-05a54ae897f4/pull/0.log" Mar 18 21:30:20 crc kubenswrapper[4950]: I0318 21:30:20.710188 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1hbdnk_c955c23a-4fa9-4745-bb00-05a54ae897f4/extract/0.log" Mar 18 21:30:20 crc kubenswrapper[4950]: I0318 21:30:20.886498 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xv6kx_df5f1fd5-ec77-45be-84de-d0d8a1568a73/extract-utilities/0.log" Mar 18 21:30:21 crc kubenswrapper[4950]: I0318 21:30:21.131029 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xv6kx_df5f1fd5-ec77-45be-84de-d0d8a1568a73/extract-content/0.log" Mar 18 21:30:21 crc kubenswrapper[4950]: I0318 21:30:21.175361 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xv6kx_df5f1fd5-ec77-45be-84de-d0d8a1568a73/extract-utilities/0.log" Mar 18 21:30:21 crc kubenswrapper[4950]: I0318 21:30:21.194567 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xv6kx_df5f1fd5-ec77-45be-84de-d0d8a1568a73/extract-content/0.log" Mar 18 21:30:21 crc kubenswrapper[4950]: I0318 21:30:21.342093 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xv6kx_df5f1fd5-ec77-45be-84de-d0d8a1568a73/extract-content/0.log" Mar 18 21:30:21 crc kubenswrapper[4950]: I0318 21:30:21.354113 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xv6kx_df5f1fd5-ec77-45be-84de-d0d8a1568a73/extract-utilities/0.log" Mar 18 21:30:21 crc kubenswrapper[4950]: I0318 21:30:21.585846 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zljc4_40ff6ac2-e78c-4501-9fdd-8d612fd6d29d/extract-utilities/0.log" Mar 18 21:30:21 crc kubenswrapper[4950]: I0318 21:30:21.883375 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zljc4_40ff6ac2-e78c-4501-9fdd-8d612fd6d29d/extract-utilities/0.log" Mar 18 21:30:21 crc kubenswrapper[4950]: I0318 21:30:21.943748 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zljc4_40ff6ac2-e78c-4501-9fdd-8d612fd6d29d/extract-content/0.log" Mar 18 21:30:21 crc kubenswrapper[4950]: I0318 21:30:21.979061 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zljc4_40ff6ac2-e78c-4501-9fdd-8d612fd6d29d/extract-content/0.log" Mar 18 21:30:22 crc kubenswrapper[4950]: I0318 21:30:22.062552 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xv6kx_df5f1fd5-ec77-45be-84de-d0d8a1568a73/registry-server/0.log" Mar 18 21:30:22 crc kubenswrapper[4950]: I0318 21:30:22.148629 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zljc4_40ff6ac2-e78c-4501-9fdd-8d612fd6d29d/extract-utilities/0.log" Mar 18 21:30:22 crc kubenswrapper[4950]: I0318 21:30:22.154653 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zljc4_40ff6ac2-e78c-4501-9fdd-8d612fd6d29d/extract-content/0.log" Mar 18 21:30:22 crc kubenswrapper[4950]: I0318 21:30:22.560245 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-c6tfb_51b0360c-046e-4360-b240-fdfd8d857a4b/marketplace-operator/0.log" Mar 18 21:30:22 crc kubenswrapper[4950]: I0318 21:30:22.600888 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-shgnx_19ec358c-0c71-4b43-b13f-688f2103d895/extract-utilities/0.log" Mar 18 21:30:22 crc kubenswrapper[4950]: I0318 21:30:22.915402 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zljc4_40ff6ac2-e78c-4501-9fdd-8d612fd6d29d/registry-server/0.log" Mar 18 21:30:22 crc kubenswrapper[4950]: I0318 21:30:22.973160 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-shgnx_19ec358c-0c71-4b43-b13f-688f2103d895/extract-utilities/0.log" Mar 18 21:30:23 crc kubenswrapper[4950]: I0318 21:30:23.048885 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-shgnx_19ec358c-0c71-4b43-b13f-688f2103d895/extract-content/0.log" Mar 18 21:30:23 crc kubenswrapper[4950]: I0318 21:30:23.049907 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-shgnx_19ec358c-0c71-4b43-b13f-688f2103d895/extract-content/0.log" Mar 18 21:30:23 crc kubenswrapper[4950]: I0318 21:30:23.188480 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-shgnx_19ec358c-0c71-4b43-b13f-688f2103d895/extract-content/0.log" Mar 18 21:30:23 crc kubenswrapper[4950]: I0318 21:30:23.193858 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-shgnx_19ec358c-0c71-4b43-b13f-688f2103d895/extract-utilities/0.log" Mar 18 21:30:23 crc kubenswrapper[4950]: I0318 21:30:23.386739 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-shgnx_19ec358c-0c71-4b43-b13f-688f2103d895/registry-server/0.log" Mar 18 21:30:23 crc kubenswrapper[4950]: I0318 21:30:23.436833 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-42xf9_d72524a5-b7b7-416f-b9ba-9527f98e787c/extract-utilities/0.log" Mar 18 21:30:23 crc kubenswrapper[4950]: I0318 21:30:23.660211 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-42xf9_d72524a5-b7b7-416f-b9ba-9527f98e787c/extract-content/0.log" Mar 18 21:30:23 crc kubenswrapper[4950]: I0318 21:30:23.669628 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-42xf9_d72524a5-b7b7-416f-b9ba-9527f98e787c/extract-utilities/0.log" Mar 18 21:30:23 crc kubenswrapper[4950]: I0318 21:30:23.695031 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-42xf9_d72524a5-b7b7-416f-b9ba-9527f98e787c/extract-content/0.log" Mar 18 21:30:23 crc kubenswrapper[4950]: I0318 21:30:23.889025 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-42xf9_d72524a5-b7b7-416f-b9ba-9527f98e787c/extract-utilities/0.log" Mar 18 21:30:24 crc kubenswrapper[4950]: I0318 21:30:24.013256 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-42xf9_d72524a5-b7b7-416f-b9ba-9527f98e787c/extract-content/0.log" Mar 18 21:30:24 crc kubenswrapper[4950]: I0318 21:30:24.141050 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-42xf9_d72524a5-b7b7-416f-b9ba-9527f98e787c/registry-server/0.log" Mar 18 21:30:36 crc kubenswrapper[4950]: I0318 21:30:36.362929 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-95m2h"] Mar 18 21:30:36 crc kubenswrapper[4950]: E0318 21:30:36.364779 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bc38aaf-a449-4cec-bc8e-1f354080906a" containerName="collect-profiles" Mar 18 21:30:36 crc kubenswrapper[4950]: I0318 21:30:36.364857 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bc38aaf-a449-4cec-bc8e-1f354080906a" containerName="collect-profiles" Mar 18 21:30:36 crc kubenswrapper[4950]: E0318 21:30:36.364943 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f18ea393-c189-46c3-9755-33c306d9dd07" containerName="oc" Mar 18 21:30:36 crc kubenswrapper[4950]: I0318 21:30:36.365001 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="f18ea393-c189-46c3-9755-33c306d9dd07" containerName="oc" Mar 18 21:30:36 crc kubenswrapper[4950]: I0318 21:30:36.365232 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="f18ea393-c189-46c3-9755-33c306d9dd07" containerName="oc" Mar 18 21:30:36 crc kubenswrapper[4950]: I0318 21:30:36.368252 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bc38aaf-a449-4cec-bc8e-1f354080906a" containerName="collect-profiles" Mar 18 21:30:36 crc kubenswrapper[4950]: I0318 21:30:36.369795 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-95m2h" Mar 18 21:30:36 crc kubenswrapper[4950]: I0318 21:30:36.390318 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-95m2h"] Mar 18 21:30:36 crc kubenswrapper[4950]: I0318 21:30:36.424797 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ab6c74f-668b-4941-97ba-a0af7df9a661-catalog-content\") pod \"redhat-operators-95m2h\" (UID: \"4ab6c74f-668b-4941-97ba-a0af7df9a661\") " pod="openshift-marketplace/redhat-operators-95m2h" Mar 18 21:30:36 crc kubenswrapper[4950]: I0318 21:30:36.424854 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ab6c74f-668b-4941-97ba-a0af7df9a661-utilities\") pod \"redhat-operators-95m2h\" (UID: \"4ab6c74f-668b-4941-97ba-a0af7df9a661\") " pod="openshift-marketplace/redhat-operators-95m2h" Mar 18 21:30:36 crc kubenswrapper[4950]: I0318 21:30:36.424904 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb8tb\" (UniqueName: \"kubernetes.io/projected/4ab6c74f-668b-4941-97ba-a0af7df9a661-kube-api-access-pb8tb\") pod \"redhat-operators-95m2h\" (UID: \"4ab6c74f-668b-4941-97ba-a0af7df9a661\") " pod="openshift-marketplace/redhat-operators-95m2h" Mar 18 21:30:36 crc kubenswrapper[4950]: I0318 21:30:36.527174 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ab6c74f-668b-4941-97ba-a0af7df9a661-catalog-content\") pod \"redhat-operators-95m2h\" (UID: \"4ab6c74f-668b-4941-97ba-a0af7df9a661\") " pod="openshift-marketplace/redhat-operators-95m2h" Mar 18 21:30:36 crc kubenswrapper[4950]: I0318 21:30:36.527457 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ab6c74f-668b-4941-97ba-a0af7df9a661-utilities\") pod \"redhat-operators-95m2h\" (UID: \"4ab6c74f-668b-4941-97ba-a0af7df9a661\") " pod="openshift-marketplace/redhat-operators-95m2h" Mar 18 21:30:36 crc kubenswrapper[4950]: I0318 21:30:36.527605 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb8tb\" (UniqueName: \"kubernetes.io/projected/4ab6c74f-668b-4941-97ba-a0af7df9a661-kube-api-access-pb8tb\") pod \"redhat-operators-95m2h\" (UID: \"4ab6c74f-668b-4941-97ba-a0af7df9a661\") " pod="openshift-marketplace/redhat-operators-95m2h" Mar 18 21:30:36 crc kubenswrapper[4950]: I0318 21:30:36.527893 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ab6c74f-668b-4941-97ba-a0af7df9a661-catalog-content\") pod \"redhat-operators-95m2h\" (UID: \"4ab6c74f-668b-4941-97ba-a0af7df9a661\") " pod="openshift-marketplace/redhat-operators-95m2h" Mar 18 21:30:36 crc kubenswrapper[4950]: I0318 21:30:36.527904 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ab6c74f-668b-4941-97ba-a0af7df9a661-utilities\") pod \"redhat-operators-95m2h\" (UID: \"4ab6c74f-668b-4941-97ba-a0af7df9a661\") " pod="openshift-marketplace/redhat-operators-95m2h" Mar 18 21:30:36 crc kubenswrapper[4950]: I0318 21:30:36.735348 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb8tb\" (UniqueName: \"kubernetes.io/projected/4ab6c74f-668b-4941-97ba-a0af7df9a661-kube-api-access-pb8tb\") pod \"redhat-operators-95m2h\" (UID: \"4ab6c74f-668b-4941-97ba-a0af7df9a661\") " pod="openshift-marketplace/redhat-operators-95m2h" Mar 18 21:30:36 crc kubenswrapper[4950]: I0318 21:30:36.992028 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-95m2h" Mar 18 21:30:37 crc kubenswrapper[4950]: I0318 21:30:37.441951 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-95m2h"] Mar 18 21:30:38 crc kubenswrapper[4950]: I0318 21:30:38.234876 4950 generic.go:334] "Generic (PLEG): container finished" podID="4ab6c74f-668b-4941-97ba-a0af7df9a661" containerID="7da743c14d0b6643a51e24178844cdbfc9f6aa4338652b065bd154d5b79ac011" exitCode=0 Mar 18 21:30:38 crc kubenswrapper[4950]: I0318 21:30:38.234940 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-95m2h" event={"ID":"4ab6c74f-668b-4941-97ba-a0af7df9a661","Type":"ContainerDied","Data":"7da743c14d0b6643a51e24178844cdbfc9f6aa4338652b065bd154d5b79ac011"} Mar 18 21:30:38 crc kubenswrapper[4950]: I0318 21:30:38.235153 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-95m2h" event={"ID":"4ab6c74f-668b-4941-97ba-a0af7df9a661","Type":"ContainerStarted","Data":"725dfc71c4f5a31a3fd54598909fe982fc5228584218aa14a2f9fb27babe5e2a"} Mar 18 21:30:40 crc kubenswrapper[4950]: I0318 21:30:40.251635 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-95m2h" event={"ID":"4ab6c74f-668b-4941-97ba-a0af7df9a661","Type":"ContainerStarted","Data":"7d7e41ed109b3808d312fb6f84a9dd1fe445a2c6d19b8e91704391a2e0eedcb9"} Mar 18 21:30:45 crc kubenswrapper[4950]: I0318 21:30:45.291024 4950 generic.go:334] "Generic (PLEG): container finished" podID="4ab6c74f-668b-4941-97ba-a0af7df9a661" containerID="7d7e41ed109b3808d312fb6f84a9dd1fe445a2c6d19b8e91704391a2e0eedcb9" exitCode=0 Mar 18 21:30:45 crc kubenswrapper[4950]: I0318 21:30:45.291079 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-95m2h" event={"ID":"4ab6c74f-668b-4941-97ba-a0af7df9a661","Type":"ContainerDied","Data":"7d7e41ed109b3808d312fb6f84a9dd1fe445a2c6d19b8e91704391a2e0eedcb9"} Mar 18 21:30:46 crc kubenswrapper[4950]: I0318 21:30:46.303056 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-95m2h" event={"ID":"4ab6c74f-668b-4941-97ba-a0af7df9a661","Type":"ContainerStarted","Data":"184e28cbc78dabc034cb338c28b62e4a93b6722e794daff7d97c6f837262f670"} Mar 18 21:30:46 crc kubenswrapper[4950]: I0318 21:30:46.326323 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-95m2h" podStartSLOduration=2.603632829 podStartE2EDuration="10.326300128s" podCreationTimestamp="2026-03-18 21:30:36 +0000 UTC" firstStartedPulling="2026-03-18 21:30:38.236630545 +0000 UTC m=+5051.477472413" lastFinishedPulling="2026-03-18 21:30:45.959297844 +0000 UTC m=+5059.200139712" observedRunningTime="2026-03-18 21:30:46.318659127 +0000 UTC m=+5059.559500995" watchObservedRunningTime="2026-03-18 21:30:46.326300128 +0000 UTC m=+5059.567142016" Mar 18 21:30:46 crc kubenswrapper[4950]: I0318 21:30:46.992887 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-95m2h" Mar 18 21:30:46 crc kubenswrapper[4950]: I0318 21:30:46.992940 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-95m2h" Mar 18 21:30:48 crc kubenswrapper[4950]: I0318 21:30:48.488583 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-95m2h" podUID="4ab6c74f-668b-4941-97ba-a0af7df9a661" containerName="registry-server" probeResult="failure" output=< Mar 18 21:30:48 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 21:30:48 crc kubenswrapper[4950]: > Mar 18 21:30:58 crc kubenswrapper[4950]: I0318 21:30:58.040144 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-95m2h" podUID="4ab6c74f-668b-4941-97ba-a0af7df9a661" containerName="registry-server" probeResult="failure" output=< Mar 18 21:30:58 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 21:30:58 crc kubenswrapper[4950]: > Mar 18 21:31:02 crc kubenswrapper[4950]: I0318 21:31:02.500035 4950 scope.go:117] "RemoveContainer" containerID="db2c55f91c2c35b9c837ceca97531e7241f2e3130c5bc62b34d7f2b9a6b1b649" Mar 18 21:31:02 crc kubenswrapper[4950]: I0318 21:31:02.527870 4950 scope.go:117] "RemoveContainer" containerID="dad741bf8d907bf950db1e7da6e030c01bea436686f80d2875f1db450f84a185" Mar 18 21:31:08 crc kubenswrapper[4950]: I0318 21:31:08.041405 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-95m2h" podUID="4ab6c74f-668b-4941-97ba-a0af7df9a661" containerName="registry-server" probeResult="failure" output=< Mar 18 21:31:08 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 21:31:08 crc kubenswrapper[4950]: > Mar 18 21:31:18 crc kubenswrapper[4950]: I0318 21:31:18.059537 4950 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-95m2h" podUID="4ab6c74f-668b-4941-97ba-a0af7df9a661" containerName="registry-server" probeResult="failure" output=< Mar 18 21:31:18 crc kubenswrapper[4950]: timeout: failed to connect service ":50051" within 1s Mar 18 21:31:18 crc kubenswrapper[4950]: > Mar 18 21:31:27 crc kubenswrapper[4950]: I0318 21:31:27.066690 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-95m2h" Mar 18 21:31:27 crc kubenswrapper[4950]: I0318 21:31:27.143612 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-95m2h" Mar 18 21:31:27 crc kubenswrapper[4950]: I0318 21:31:27.313856 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-95m2h"] Mar 18 21:31:28 crc kubenswrapper[4950]: I0318 21:31:28.757862 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-95m2h" podUID="4ab6c74f-668b-4941-97ba-a0af7df9a661" containerName="registry-server" containerID="cri-o://184e28cbc78dabc034cb338c28b62e4a93b6722e794daff7d97c6f837262f670" gracePeriod=2 Mar 18 21:31:29 crc kubenswrapper[4950]: I0318 21:31:29.475137 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-95m2h" Mar 18 21:31:29 crc kubenswrapper[4950]: I0318 21:31:29.494837 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pb8tb\" (UniqueName: \"kubernetes.io/projected/4ab6c74f-668b-4941-97ba-a0af7df9a661-kube-api-access-pb8tb\") pod \"4ab6c74f-668b-4941-97ba-a0af7df9a661\" (UID: \"4ab6c74f-668b-4941-97ba-a0af7df9a661\") " Mar 18 21:31:29 crc kubenswrapper[4950]: I0318 21:31:29.495661 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ab6c74f-668b-4941-97ba-a0af7df9a661-utilities\") pod \"4ab6c74f-668b-4941-97ba-a0af7df9a661\" (UID: \"4ab6c74f-668b-4941-97ba-a0af7df9a661\") " Mar 18 21:31:29 crc kubenswrapper[4950]: I0318 21:31:29.495784 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ab6c74f-668b-4941-97ba-a0af7df9a661-catalog-content\") pod \"4ab6c74f-668b-4941-97ba-a0af7df9a661\" (UID: \"4ab6c74f-668b-4941-97ba-a0af7df9a661\") " Mar 18 21:31:29 crc kubenswrapper[4950]: I0318 21:31:29.497627 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ab6c74f-668b-4941-97ba-a0af7df9a661-utilities" (OuterVolumeSpecName: "utilities") pod "4ab6c74f-668b-4941-97ba-a0af7df9a661" (UID: "4ab6c74f-668b-4941-97ba-a0af7df9a661"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:31:29 crc kubenswrapper[4950]: I0318 21:31:29.511136 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ab6c74f-668b-4941-97ba-a0af7df9a661-kube-api-access-pb8tb" (OuterVolumeSpecName: "kube-api-access-pb8tb") pod "4ab6c74f-668b-4941-97ba-a0af7df9a661" (UID: "4ab6c74f-668b-4941-97ba-a0af7df9a661"). InnerVolumeSpecName "kube-api-access-pb8tb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:31:29 crc kubenswrapper[4950]: I0318 21:31:29.598682 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pb8tb\" (UniqueName: \"kubernetes.io/projected/4ab6c74f-668b-4941-97ba-a0af7df9a661-kube-api-access-pb8tb\") on node \"crc\" DevicePath \"\"" Mar 18 21:31:29 crc kubenswrapper[4950]: I0318 21:31:29.599058 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ab6c74f-668b-4941-97ba-a0af7df9a661-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 21:31:29 crc kubenswrapper[4950]: I0318 21:31:29.649830 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ab6c74f-668b-4941-97ba-a0af7df9a661-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4ab6c74f-668b-4941-97ba-a0af7df9a661" (UID: "4ab6c74f-668b-4941-97ba-a0af7df9a661"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:31:29 crc kubenswrapper[4950]: I0318 21:31:29.700532 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ab6c74f-668b-4941-97ba-a0af7df9a661-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 21:31:29 crc kubenswrapper[4950]: I0318 21:31:29.771950 4950 generic.go:334] "Generic (PLEG): container finished" podID="4ab6c74f-668b-4941-97ba-a0af7df9a661" containerID="184e28cbc78dabc034cb338c28b62e4a93b6722e794daff7d97c6f837262f670" exitCode=0 Mar 18 21:31:29 crc kubenswrapper[4950]: I0318 21:31:29.772008 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-95m2h" event={"ID":"4ab6c74f-668b-4941-97ba-a0af7df9a661","Type":"ContainerDied","Data":"184e28cbc78dabc034cb338c28b62e4a93b6722e794daff7d97c6f837262f670"} Mar 18 21:31:29 crc kubenswrapper[4950]: I0318 21:31:29.772047 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-95m2h" event={"ID":"4ab6c74f-668b-4941-97ba-a0af7df9a661","Type":"ContainerDied","Data":"725dfc71c4f5a31a3fd54598909fe982fc5228584218aa14a2f9fb27babe5e2a"} Mar 18 21:31:29 crc kubenswrapper[4950]: I0318 21:31:29.772082 4950 scope.go:117] "RemoveContainer" containerID="184e28cbc78dabc034cb338c28b62e4a93b6722e794daff7d97c6f837262f670" Mar 18 21:31:29 crc kubenswrapper[4950]: I0318 21:31:29.772293 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-95m2h" Mar 18 21:31:29 crc kubenswrapper[4950]: I0318 21:31:29.814315 4950 scope.go:117] "RemoveContainer" containerID="7d7e41ed109b3808d312fb6f84a9dd1fe445a2c6d19b8e91704391a2e0eedcb9" Mar 18 21:31:29 crc kubenswrapper[4950]: I0318 21:31:29.820760 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-95m2h"] Mar 18 21:31:29 crc kubenswrapper[4950]: I0318 21:31:29.828979 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-95m2h"] Mar 18 21:31:29 crc kubenswrapper[4950]: I0318 21:31:29.948202 4950 scope.go:117] "RemoveContainer" containerID="7da743c14d0b6643a51e24178844cdbfc9f6aa4338652b065bd154d5b79ac011" Mar 18 21:31:30 crc kubenswrapper[4950]: I0318 21:31:30.037638 4950 scope.go:117] "RemoveContainer" containerID="184e28cbc78dabc034cb338c28b62e4a93b6722e794daff7d97c6f837262f670" Mar 18 21:31:30 crc kubenswrapper[4950]: E0318 21:31:30.040451 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"184e28cbc78dabc034cb338c28b62e4a93b6722e794daff7d97c6f837262f670\": container with ID starting with 184e28cbc78dabc034cb338c28b62e4a93b6722e794daff7d97c6f837262f670 not found: ID does not exist" containerID="184e28cbc78dabc034cb338c28b62e4a93b6722e794daff7d97c6f837262f670" Mar 18 21:31:30 crc kubenswrapper[4950]: I0318 21:31:30.040561 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"184e28cbc78dabc034cb338c28b62e4a93b6722e794daff7d97c6f837262f670"} err="failed to get container status \"184e28cbc78dabc034cb338c28b62e4a93b6722e794daff7d97c6f837262f670\": rpc error: code = NotFound desc = could not find container \"184e28cbc78dabc034cb338c28b62e4a93b6722e794daff7d97c6f837262f670\": container with ID starting with 184e28cbc78dabc034cb338c28b62e4a93b6722e794daff7d97c6f837262f670 not found: ID does not exist" Mar 18 21:31:30 crc kubenswrapper[4950]: I0318 21:31:30.040693 4950 scope.go:117] "RemoveContainer" containerID="7d7e41ed109b3808d312fb6f84a9dd1fe445a2c6d19b8e91704391a2e0eedcb9" Mar 18 21:31:30 crc kubenswrapper[4950]: E0318 21:31:30.041915 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d7e41ed109b3808d312fb6f84a9dd1fe445a2c6d19b8e91704391a2e0eedcb9\": container with ID starting with 7d7e41ed109b3808d312fb6f84a9dd1fe445a2c6d19b8e91704391a2e0eedcb9 not found: ID does not exist" containerID="7d7e41ed109b3808d312fb6f84a9dd1fe445a2c6d19b8e91704391a2e0eedcb9" Mar 18 21:31:30 crc kubenswrapper[4950]: I0318 21:31:30.041961 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d7e41ed109b3808d312fb6f84a9dd1fe445a2c6d19b8e91704391a2e0eedcb9"} err="failed to get container status \"7d7e41ed109b3808d312fb6f84a9dd1fe445a2c6d19b8e91704391a2e0eedcb9\": rpc error: code = NotFound desc = could not find container \"7d7e41ed109b3808d312fb6f84a9dd1fe445a2c6d19b8e91704391a2e0eedcb9\": container with ID starting with 7d7e41ed109b3808d312fb6f84a9dd1fe445a2c6d19b8e91704391a2e0eedcb9 not found: ID does not exist" Mar 18 21:31:30 crc kubenswrapper[4950]: I0318 21:31:30.042021 4950 scope.go:117] "RemoveContainer" containerID="7da743c14d0b6643a51e24178844cdbfc9f6aa4338652b065bd154d5b79ac011" Mar 18 21:31:30 crc kubenswrapper[4950]: E0318 21:31:30.042909 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7da743c14d0b6643a51e24178844cdbfc9f6aa4338652b065bd154d5b79ac011\": container with ID starting with 7da743c14d0b6643a51e24178844cdbfc9f6aa4338652b065bd154d5b79ac011 not found: ID does not exist" containerID="7da743c14d0b6643a51e24178844cdbfc9f6aa4338652b065bd154d5b79ac011" Mar 18 21:31:30 crc kubenswrapper[4950]: I0318 21:31:30.043004 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7da743c14d0b6643a51e24178844cdbfc9f6aa4338652b065bd154d5b79ac011"} err="failed to get container status \"7da743c14d0b6643a51e24178844cdbfc9f6aa4338652b065bd154d5b79ac011\": rpc error: code = NotFound desc = could not find container \"7da743c14d0b6643a51e24178844cdbfc9f6aa4338652b065bd154d5b79ac011\": container with ID starting with 7da743c14d0b6643a51e24178844cdbfc9f6aa4338652b065bd154d5b79ac011 not found: ID does not exist" Mar 18 21:31:31 crc kubenswrapper[4950]: I0318 21:31:31.491995 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ab6c74f-668b-4941-97ba-a0af7df9a661" path="/var/lib/kubelet/pods/4ab6c74f-668b-4941-97ba-a0af7df9a661/volumes" Mar 18 21:31:33 crc kubenswrapper[4950]: I0318 21:31:33.837130 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 21:31:33 crc kubenswrapper[4950]: I0318 21:31:33.837874 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 21:32:00 crc kubenswrapper[4950]: I0318 21:32:00.193218 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564492-xhm57"] Mar 18 21:32:00 crc kubenswrapper[4950]: E0318 21:32:00.195535 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab6c74f-668b-4941-97ba-a0af7df9a661" containerName="extract-utilities" Mar 18 21:32:00 crc kubenswrapper[4950]: I0318 21:32:00.195590 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab6c74f-668b-4941-97ba-a0af7df9a661" containerName="extract-utilities" Mar 18 21:32:00 crc kubenswrapper[4950]: E0318 21:32:00.195636 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab6c74f-668b-4941-97ba-a0af7df9a661" containerName="extract-content" Mar 18 21:32:00 crc kubenswrapper[4950]: I0318 21:32:00.195655 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab6c74f-668b-4941-97ba-a0af7df9a661" containerName="extract-content" Mar 18 21:32:00 crc kubenswrapper[4950]: E0318 21:32:00.195725 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab6c74f-668b-4941-97ba-a0af7df9a661" containerName="registry-server" Mar 18 21:32:00 crc kubenswrapper[4950]: I0318 21:32:00.195744 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab6c74f-668b-4941-97ba-a0af7df9a661" containerName="registry-server" Mar 18 21:32:00 crc kubenswrapper[4950]: I0318 21:32:00.196223 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab6c74f-668b-4941-97ba-a0af7df9a661" containerName="registry-server" Mar 18 21:32:00 crc kubenswrapper[4950]: I0318 21:32:00.197858 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564492-xhm57" Mar 18 21:32:00 crc kubenswrapper[4950]: I0318 21:32:00.204972 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564492-xhm57"] Mar 18 21:32:00 crc kubenswrapper[4950]: I0318 21:32:00.232375 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 21:32:00 crc kubenswrapper[4950]: I0318 21:32:00.232611 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 21:32:00 crc kubenswrapper[4950]: I0318 21:32:00.233129 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 21:32:00 crc kubenswrapper[4950]: I0318 21:32:00.234970 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nd957\" (UniqueName: \"kubernetes.io/projected/894f5ea5-a48c-4db5-a739-c736f09a748c-kube-api-access-nd957\") pod \"auto-csr-approver-29564492-xhm57\" (UID: \"894f5ea5-a48c-4db5-a739-c736f09a748c\") " pod="openshift-infra/auto-csr-approver-29564492-xhm57" Mar 18 21:32:00 crc kubenswrapper[4950]: I0318 21:32:00.336055 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nd957\" (UniqueName: \"kubernetes.io/projected/894f5ea5-a48c-4db5-a739-c736f09a748c-kube-api-access-nd957\") pod \"auto-csr-approver-29564492-xhm57\" (UID: \"894f5ea5-a48c-4db5-a739-c736f09a748c\") " pod="openshift-infra/auto-csr-approver-29564492-xhm57" Mar 18 21:32:00 crc kubenswrapper[4950]: I0318 21:32:00.359454 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nd957\" (UniqueName: \"kubernetes.io/projected/894f5ea5-a48c-4db5-a739-c736f09a748c-kube-api-access-nd957\") pod \"auto-csr-approver-29564492-xhm57\" (UID: \"894f5ea5-a48c-4db5-a739-c736f09a748c\") " pod="openshift-infra/auto-csr-approver-29564492-xhm57" Mar 18 21:32:00 crc kubenswrapper[4950]: I0318 21:32:00.546397 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564492-xhm57" Mar 18 21:32:01 crc kubenswrapper[4950]: I0318 21:32:01.000352 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564492-xhm57"] Mar 18 21:32:01 crc kubenswrapper[4950]: I0318 21:32:01.129244 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564492-xhm57" event={"ID":"894f5ea5-a48c-4db5-a739-c736f09a748c","Type":"ContainerStarted","Data":"c7a849145415e30c3e7955113cb394716067b7dd577b45c39e61ef5c37e0a0a9"} Mar 18 21:32:03 crc kubenswrapper[4950]: I0318 21:32:03.839910 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 21:32:03 crc kubenswrapper[4950]: I0318 21:32:03.840431 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 21:32:08 crc kubenswrapper[4950]: I0318 21:32:08.222729 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564492-xhm57" event={"ID":"894f5ea5-a48c-4db5-a739-c736f09a748c","Type":"ContainerStarted","Data":"0c0d212a095a6333ac9cbfe9881fde3f99eb3fc0d406f5e7343cac16784ce3de"} Mar 18 21:32:08 crc kubenswrapper[4950]: I0318 21:32:08.242536 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29564492-xhm57" podStartSLOduration=2.217110872 podStartE2EDuration="8.242519451s" podCreationTimestamp="2026-03-18 21:32:00 +0000 UTC" firstStartedPulling="2026-03-18 21:32:01.004238501 +0000 UTC m=+5134.245080369" lastFinishedPulling="2026-03-18 21:32:07.02964708 +0000 UTC m=+5140.270488948" observedRunningTime="2026-03-18 21:32:08.236172885 +0000 UTC m=+5141.477014753" watchObservedRunningTime="2026-03-18 21:32:08.242519451 +0000 UTC m=+5141.483361319" Mar 18 21:32:09 crc kubenswrapper[4950]: I0318 21:32:09.230910 4950 generic.go:334] "Generic (PLEG): container finished" podID="894f5ea5-a48c-4db5-a739-c736f09a748c" containerID="0c0d212a095a6333ac9cbfe9881fde3f99eb3fc0d406f5e7343cac16784ce3de" exitCode=0 Mar 18 21:32:09 crc kubenswrapper[4950]: I0318 21:32:09.230983 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564492-xhm57" event={"ID":"894f5ea5-a48c-4db5-a739-c736f09a748c","Type":"ContainerDied","Data":"0c0d212a095a6333ac9cbfe9881fde3f99eb3fc0d406f5e7343cac16784ce3de"} Mar 18 21:32:10 crc kubenswrapper[4950]: I0318 21:32:10.666016 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564492-xhm57" Mar 18 21:32:10 crc kubenswrapper[4950]: I0318 21:32:10.705174 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nd957\" (UniqueName: \"kubernetes.io/projected/894f5ea5-a48c-4db5-a739-c736f09a748c-kube-api-access-nd957\") pod \"894f5ea5-a48c-4db5-a739-c736f09a748c\" (UID: \"894f5ea5-a48c-4db5-a739-c736f09a748c\") " Mar 18 21:32:10 crc kubenswrapper[4950]: I0318 21:32:10.712340 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/894f5ea5-a48c-4db5-a739-c736f09a748c-kube-api-access-nd957" (OuterVolumeSpecName: "kube-api-access-nd957") pod "894f5ea5-a48c-4db5-a739-c736f09a748c" (UID: "894f5ea5-a48c-4db5-a739-c736f09a748c"). InnerVolumeSpecName "kube-api-access-nd957". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:32:10 crc kubenswrapper[4950]: I0318 21:32:10.807765 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nd957\" (UniqueName: \"kubernetes.io/projected/894f5ea5-a48c-4db5-a739-c736f09a748c-kube-api-access-nd957\") on node \"crc\" DevicePath \"\"" Mar 18 21:32:11 crc kubenswrapper[4950]: I0318 21:32:11.250977 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564492-xhm57" event={"ID":"894f5ea5-a48c-4db5-a739-c736f09a748c","Type":"ContainerDied","Data":"c7a849145415e30c3e7955113cb394716067b7dd577b45c39e61ef5c37e0a0a9"} Mar 18 21:32:11 crc kubenswrapper[4950]: I0318 21:32:11.251017 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7a849145415e30c3e7955113cb394716067b7dd577b45c39e61ef5c37e0a0a9" Mar 18 21:32:11 crc kubenswrapper[4950]: I0318 21:32:11.251039 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564492-xhm57" Mar 18 21:32:11 crc kubenswrapper[4950]: I0318 21:32:11.737308 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564486-4qpb9"] Mar 18 21:32:11 crc kubenswrapper[4950]: I0318 21:32:11.745552 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564486-4qpb9"] Mar 18 21:32:13 crc kubenswrapper[4950]: I0318 21:32:13.488252 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be5f2e02-ab24-4e5f-8a15-eb438e75c46c" path="/var/lib/kubelet/pods/be5f2e02-ab24-4e5f-8a15-eb438e75c46c/volumes" Mar 18 21:32:33 crc kubenswrapper[4950]: I0318 21:32:33.836046 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 21:32:33 crc kubenswrapper[4950]: I0318 21:32:33.836733 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 21:32:33 crc kubenswrapper[4950]: I0318 21:32:33.836778 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 21:32:33 crc kubenswrapper[4950]: I0318 21:32:33.837277 4950 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8a3de554c0ec05e3ab570b15a3ddb9444d28e1d8a8dd09556e19d1262de8dcfc"} pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 21:32:33 crc kubenswrapper[4950]: I0318 21:32:33.837320 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" containerID="cri-o://8a3de554c0ec05e3ab570b15a3ddb9444d28e1d8a8dd09556e19d1262de8dcfc" gracePeriod=600 Mar 18 21:32:34 crc kubenswrapper[4950]: I0318 21:32:34.459681 4950 generic.go:334] "Generic (PLEG): container finished" podID="4048b439-3266-46e7-9de0-22377efacc46" containerID="8a3de554c0ec05e3ab570b15a3ddb9444d28e1d8a8dd09556e19d1262de8dcfc" exitCode=0 Mar 18 21:32:34 crc kubenswrapper[4950]: I0318 21:32:34.459757 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerDied","Data":"8a3de554c0ec05e3ab570b15a3ddb9444d28e1d8a8dd09556e19d1262de8dcfc"} Mar 18 21:32:34 crc kubenswrapper[4950]: I0318 21:32:34.460435 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerStarted","Data":"6ea174cb4f877354d40698f8225ba9da25bb0ae0a816d0571cda56c7af7db12c"} Mar 18 21:32:34 crc kubenswrapper[4950]: I0318 21:32:34.460466 4950 scope.go:117] "RemoveContainer" containerID="cb7a5dccba6a129c79df5f36ab3e44e2bf20f230cbbcfe5b407f9e2a040a997d" Mar 18 21:32:57 crc kubenswrapper[4950]: I0318 21:32:57.719853 4950 generic.go:334] "Generic (PLEG): container finished" podID="b71b3828-70c5-4ec1-9b72-b761fc24e58a" containerID="24f78ca40020d2e14b0e130e8d00374b0bfe298ae379b47d22f5b37d6a4f3dd0" exitCode=0 Mar 18 21:32:57 crc kubenswrapper[4950]: I0318 21:32:57.719949 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xtwkj/must-gather-c9z75" event={"ID":"b71b3828-70c5-4ec1-9b72-b761fc24e58a","Type":"ContainerDied","Data":"24f78ca40020d2e14b0e130e8d00374b0bfe298ae379b47d22f5b37d6a4f3dd0"} Mar 18 21:32:57 crc kubenswrapper[4950]: I0318 21:32:57.722029 4950 scope.go:117] "RemoveContainer" containerID="24f78ca40020d2e14b0e130e8d00374b0bfe298ae379b47d22f5b37d6a4f3dd0" Mar 18 21:32:57 crc kubenswrapper[4950]: I0318 21:32:57.801991 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xtwkj_must-gather-c9z75_b71b3828-70c5-4ec1-9b72-b761fc24e58a/gather/0.log" Mar 18 21:33:02 crc kubenswrapper[4950]: I0318 21:33:02.716625 4950 scope.go:117] "RemoveContainer" containerID="2272ce00289e64e90ff5e828d8a949259a4efa3bd66f8986c377527f8309eca8" Mar 18 21:33:11 crc kubenswrapper[4950]: I0318 21:33:11.524212 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xtwkj/must-gather-c9z75"] Mar 18 21:33:11 crc kubenswrapper[4950]: I0318 21:33:11.526015 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-xtwkj/must-gather-c9z75" podUID="b71b3828-70c5-4ec1-9b72-b761fc24e58a" containerName="copy" containerID="cri-o://ba4fe7ecf4b2cb9c3ed96352b8e023b399cf69cf24921e59f0bb7e1b527d256d" gracePeriod=2 Mar 18 21:33:11 crc kubenswrapper[4950]: I0318 21:33:11.542205 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xtwkj/must-gather-c9z75"] Mar 18 21:33:11 crc kubenswrapper[4950]: I0318 21:33:11.921731 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xtwkj_must-gather-c9z75_b71b3828-70c5-4ec1-9b72-b761fc24e58a/copy/0.log" Mar 18 21:33:11 crc kubenswrapper[4950]: I0318 21:33:11.922304 4950 generic.go:334] "Generic (PLEG): container finished" podID="b71b3828-70c5-4ec1-9b72-b761fc24e58a" containerID="ba4fe7ecf4b2cb9c3ed96352b8e023b399cf69cf24921e59f0bb7e1b527d256d" exitCode=143 Mar 18 21:33:12 crc kubenswrapper[4950]: I0318 21:33:12.072578 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xtwkj_must-gather-c9z75_b71b3828-70c5-4ec1-9b72-b761fc24e58a/copy/0.log" Mar 18 21:33:12 crc kubenswrapper[4950]: I0318 21:33:12.072884 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xtwkj/must-gather-c9z75" Mar 18 21:33:12 crc kubenswrapper[4950]: I0318 21:33:12.205671 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfkgz\" (UniqueName: \"kubernetes.io/projected/b71b3828-70c5-4ec1-9b72-b761fc24e58a-kube-api-access-gfkgz\") pod \"b71b3828-70c5-4ec1-9b72-b761fc24e58a\" (UID: \"b71b3828-70c5-4ec1-9b72-b761fc24e58a\") " Mar 18 21:33:12 crc kubenswrapper[4950]: I0318 21:33:12.205938 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b71b3828-70c5-4ec1-9b72-b761fc24e58a-must-gather-output\") pod \"b71b3828-70c5-4ec1-9b72-b761fc24e58a\" (UID: \"b71b3828-70c5-4ec1-9b72-b761fc24e58a\") " Mar 18 21:33:12 crc kubenswrapper[4950]: I0318 21:33:12.212072 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b71b3828-70c5-4ec1-9b72-b761fc24e58a-kube-api-access-gfkgz" (OuterVolumeSpecName: "kube-api-access-gfkgz") pod "b71b3828-70c5-4ec1-9b72-b761fc24e58a" (UID: "b71b3828-70c5-4ec1-9b72-b761fc24e58a"). InnerVolumeSpecName "kube-api-access-gfkgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:33:12 crc kubenswrapper[4950]: I0318 21:33:12.307600 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfkgz\" (UniqueName: \"kubernetes.io/projected/b71b3828-70c5-4ec1-9b72-b761fc24e58a-kube-api-access-gfkgz\") on node \"crc\" DevicePath \"\"" Mar 18 21:33:12 crc kubenswrapper[4950]: I0318 21:33:12.395626 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b71b3828-70c5-4ec1-9b72-b761fc24e58a-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "b71b3828-70c5-4ec1-9b72-b761fc24e58a" (UID: "b71b3828-70c5-4ec1-9b72-b761fc24e58a"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:33:12 crc kubenswrapper[4950]: I0318 21:33:12.408858 4950 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b71b3828-70c5-4ec1-9b72-b761fc24e58a-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 18 21:33:12 crc kubenswrapper[4950]: I0318 21:33:12.931567 4950 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xtwkj_must-gather-c9z75_b71b3828-70c5-4ec1-9b72-b761fc24e58a/copy/0.log" Mar 18 21:33:12 crc kubenswrapper[4950]: I0318 21:33:12.932014 4950 scope.go:117] "RemoveContainer" containerID="ba4fe7ecf4b2cb9c3ed96352b8e023b399cf69cf24921e59f0bb7e1b527d256d" Mar 18 21:33:12 crc kubenswrapper[4950]: I0318 21:33:12.932078 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xtwkj/must-gather-c9z75" Mar 18 21:33:12 crc kubenswrapper[4950]: I0318 21:33:12.951770 4950 scope.go:117] "RemoveContainer" containerID="24f78ca40020d2e14b0e130e8d00374b0bfe298ae379b47d22f5b37d6a4f3dd0" Mar 18 21:33:13 crc kubenswrapper[4950]: I0318 21:33:13.491473 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b71b3828-70c5-4ec1-9b72-b761fc24e58a" path="/var/lib/kubelet/pods/b71b3828-70c5-4ec1-9b72-b761fc24e58a/volumes" Mar 18 21:34:00 crc kubenswrapper[4950]: I0318 21:34:00.150019 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564494-lrwv5"] Mar 18 21:34:00 crc kubenswrapper[4950]: E0318 21:34:00.150909 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="894f5ea5-a48c-4db5-a739-c736f09a748c" containerName="oc" Mar 18 21:34:00 crc kubenswrapper[4950]: I0318 21:34:00.150921 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="894f5ea5-a48c-4db5-a739-c736f09a748c" containerName="oc" Mar 18 21:34:00 crc kubenswrapper[4950]: E0318 21:34:00.150936 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b71b3828-70c5-4ec1-9b72-b761fc24e58a" containerName="gather" Mar 18 21:34:00 crc kubenswrapper[4950]: I0318 21:34:00.150942 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="b71b3828-70c5-4ec1-9b72-b761fc24e58a" containerName="gather" Mar 18 21:34:00 crc kubenswrapper[4950]: E0318 21:34:00.150951 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b71b3828-70c5-4ec1-9b72-b761fc24e58a" containerName="copy" Mar 18 21:34:00 crc kubenswrapper[4950]: I0318 21:34:00.150957 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="b71b3828-70c5-4ec1-9b72-b761fc24e58a" containerName="copy" Mar 18 21:34:00 crc kubenswrapper[4950]: I0318 21:34:00.151130 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="b71b3828-70c5-4ec1-9b72-b761fc24e58a" containerName="gather" Mar 18 21:34:00 crc kubenswrapper[4950]: I0318 21:34:00.151138 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="894f5ea5-a48c-4db5-a739-c736f09a748c" containerName="oc" Mar 18 21:34:00 crc kubenswrapper[4950]: I0318 21:34:00.151147 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="b71b3828-70c5-4ec1-9b72-b761fc24e58a" containerName="copy" Mar 18 21:34:00 crc kubenswrapper[4950]: I0318 21:34:00.151783 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564494-lrwv5" Mar 18 21:34:00 crc kubenswrapper[4950]: I0318 21:34:00.155017 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 21:34:00 crc kubenswrapper[4950]: I0318 21:34:00.156448 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 21:34:00 crc kubenswrapper[4950]: I0318 21:34:00.160743 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564494-lrwv5"] Mar 18 21:34:00 crc kubenswrapper[4950]: I0318 21:34:00.162050 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cb48\" (UniqueName: \"kubernetes.io/projected/27138219-10be-4484-a5bf-20ed9e0977a4-kube-api-access-7cb48\") pod \"auto-csr-approver-29564494-lrwv5\" (UID: \"27138219-10be-4484-a5bf-20ed9e0977a4\") " pod="openshift-infra/auto-csr-approver-29564494-lrwv5" Mar 18 21:34:00 crc kubenswrapper[4950]: I0318 21:34:00.163047 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 21:34:00 crc kubenswrapper[4950]: I0318 21:34:00.264197 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cb48\" (UniqueName: \"kubernetes.io/projected/27138219-10be-4484-a5bf-20ed9e0977a4-kube-api-access-7cb48\") pod \"auto-csr-approver-29564494-lrwv5\" (UID: \"27138219-10be-4484-a5bf-20ed9e0977a4\") " pod="openshift-infra/auto-csr-approver-29564494-lrwv5" Mar 18 21:34:00 crc kubenswrapper[4950]: I0318 21:34:00.283395 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cb48\" (UniqueName: \"kubernetes.io/projected/27138219-10be-4484-a5bf-20ed9e0977a4-kube-api-access-7cb48\") pod \"auto-csr-approver-29564494-lrwv5\" (UID: \"27138219-10be-4484-a5bf-20ed9e0977a4\") " pod="openshift-infra/auto-csr-approver-29564494-lrwv5" Mar 18 21:34:00 crc kubenswrapper[4950]: I0318 21:34:00.473486 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564494-lrwv5" Mar 18 21:34:00 crc kubenswrapper[4950]: I0318 21:34:00.924186 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564494-lrwv5"] Mar 18 21:34:01 crc kubenswrapper[4950]: I0318 21:34:01.402909 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564494-lrwv5" event={"ID":"27138219-10be-4484-a5bf-20ed9e0977a4","Type":"ContainerStarted","Data":"62a8117650af9ae3277fd553ff1a42dc0339dccd65fb3cbbe90a48495d81ad31"} Mar 18 21:34:02 crc kubenswrapper[4950]: I0318 21:34:02.415311 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564494-lrwv5" event={"ID":"27138219-10be-4484-a5bf-20ed9e0977a4","Type":"ContainerStarted","Data":"01b012327a5e1394520fea3a0bd7f5e3db819fc341e8a777707272a6cb170a7e"} Mar 18 21:34:02 crc kubenswrapper[4950]: I0318 21:34:02.442886 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29564494-lrwv5" podStartSLOduration=1.454498299 podStartE2EDuration="2.44285978s" podCreationTimestamp="2026-03-18 21:34:00 +0000 UTC" firstStartedPulling="2026-03-18 21:34:00.931361995 +0000 UTC m=+5254.172203863" lastFinishedPulling="2026-03-18 21:34:01.919723476 +0000 UTC m=+5255.160565344" observedRunningTime="2026-03-18 21:34:02.428009471 +0000 UTC m=+5255.668851389" watchObservedRunningTime="2026-03-18 21:34:02.44285978 +0000 UTC m=+5255.683701668" Mar 18 21:34:03 crc kubenswrapper[4950]: I0318 21:34:03.429880 4950 generic.go:334] "Generic (PLEG): container finished" podID="27138219-10be-4484-a5bf-20ed9e0977a4" containerID="01b012327a5e1394520fea3a0bd7f5e3db819fc341e8a777707272a6cb170a7e" exitCode=0 Mar 18 21:34:03 crc kubenswrapper[4950]: I0318 21:34:03.432363 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564494-lrwv5" event={"ID":"27138219-10be-4484-a5bf-20ed9e0977a4","Type":"ContainerDied","Data":"01b012327a5e1394520fea3a0bd7f5e3db819fc341e8a777707272a6cb170a7e"} Mar 18 21:34:05 crc kubenswrapper[4950]: I0318 21:34:05.276862 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564494-lrwv5" Mar 18 21:34:05 crc kubenswrapper[4950]: I0318 21:34:05.471171 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cb48\" (UniqueName: \"kubernetes.io/projected/27138219-10be-4484-a5bf-20ed9e0977a4-kube-api-access-7cb48\") pod \"27138219-10be-4484-a5bf-20ed9e0977a4\" (UID: \"27138219-10be-4484-a5bf-20ed9e0977a4\") " Mar 18 21:34:05 crc kubenswrapper[4950]: I0318 21:34:05.507070 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564494-lrwv5" Mar 18 21:34:05 crc kubenswrapper[4950]: I0318 21:34:05.529865 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27138219-10be-4484-a5bf-20ed9e0977a4-kube-api-access-7cb48" (OuterVolumeSpecName: "kube-api-access-7cb48") pod "27138219-10be-4484-a5bf-20ed9e0977a4" (UID: "27138219-10be-4484-a5bf-20ed9e0977a4"). InnerVolumeSpecName "kube-api-access-7cb48". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:34:05 crc kubenswrapper[4950]: I0318 21:34:05.544954 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564494-lrwv5" event={"ID":"27138219-10be-4484-a5bf-20ed9e0977a4","Type":"ContainerDied","Data":"62a8117650af9ae3277fd553ff1a42dc0339dccd65fb3cbbe90a48495d81ad31"} Mar 18 21:34:05 crc kubenswrapper[4950]: I0318 21:34:05.545015 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62a8117650af9ae3277fd553ff1a42dc0339dccd65fb3cbbe90a48495d81ad31" Mar 18 21:34:05 crc kubenswrapper[4950]: I0318 21:34:05.558106 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564488-t7mz2"] Mar 18 21:34:05 crc kubenswrapper[4950]: I0318 21:34:05.568172 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564488-t7mz2"] Mar 18 21:34:05 crc kubenswrapper[4950]: I0318 21:34:05.573503 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cb48\" (UniqueName: \"kubernetes.io/projected/27138219-10be-4484-a5bf-20ed9e0977a4-kube-api-access-7cb48\") on node \"crc\" DevicePath \"\"" Mar 18 21:34:07 crc kubenswrapper[4950]: I0318 21:34:07.501153 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="053e72dd-7249-4665-848b-e1c98eedabb6" path="/var/lib/kubelet/pods/053e72dd-7249-4665-848b-e1c98eedabb6/volumes" Mar 18 21:34:48 crc kubenswrapper[4950]: I0318 21:34:48.920277 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-42rg6"] Mar 18 21:34:48 crc kubenswrapper[4950]: E0318 21:34:48.921234 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27138219-10be-4484-a5bf-20ed9e0977a4" containerName="oc" Mar 18 21:34:48 crc kubenswrapper[4950]: I0318 21:34:48.921248 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="27138219-10be-4484-a5bf-20ed9e0977a4" containerName="oc" Mar 18 21:34:48 crc kubenswrapper[4950]: I0318 21:34:48.921471 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="27138219-10be-4484-a5bf-20ed9e0977a4" containerName="oc" Mar 18 21:34:48 crc kubenswrapper[4950]: I0318 21:34:48.923005 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-42rg6" Mar 18 21:34:49 crc kubenswrapper[4950]: I0318 21:34:48.960014 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-42rg6"] Mar 18 21:34:49 crc kubenswrapper[4950]: I0318 21:34:49.028782 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e6cdb80-21e2-4fdf-a64b-86328e236349-utilities\") pod \"certified-operators-42rg6\" (UID: \"8e6cdb80-21e2-4fdf-a64b-86328e236349\") " pod="openshift-marketplace/certified-operators-42rg6" Mar 18 21:34:49 crc kubenswrapper[4950]: I0318 21:34:49.028843 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlfqs\" (UniqueName: \"kubernetes.io/projected/8e6cdb80-21e2-4fdf-a64b-86328e236349-kube-api-access-tlfqs\") pod \"certified-operators-42rg6\" (UID: \"8e6cdb80-21e2-4fdf-a64b-86328e236349\") " pod="openshift-marketplace/certified-operators-42rg6" Mar 18 21:34:49 crc kubenswrapper[4950]: I0318 21:34:49.028894 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e6cdb80-21e2-4fdf-a64b-86328e236349-catalog-content\") pod \"certified-operators-42rg6\" (UID: \"8e6cdb80-21e2-4fdf-a64b-86328e236349\") " pod="openshift-marketplace/certified-operators-42rg6" Mar 18 21:34:49 crc kubenswrapper[4950]: I0318 21:34:49.130396 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e6cdb80-21e2-4fdf-a64b-86328e236349-utilities\") pod \"certified-operators-42rg6\" (UID: \"8e6cdb80-21e2-4fdf-a64b-86328e236349\") " pod="openshift-marketplace/certified-operators-42rg6" Mar 18 21:34:49 crc kubenswrapper[4950]: I0318 21:34:49.130470 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlfqs\" (UniqueName: \"kubernetes.io/projected/8e6cdb80-21e2-4fdf-a64b-86328e236349-kube-api-access-tlfqs\") pod \"certified-operators-42rg6\" (UID: \"8e6cdb80-21e2-4fdf-a64b-86328e236349\") " pod="openshift-marketplace/certified-operators-42rg6" Mar 18 21:34:49 crc kubenswrapper[4950]: I0318 21:34:49.130536 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e6cdb80-21e2-4fdf-a64b-86328e236349-catalog-content\") pod \"certified-operators-42rg6\" (UID: \"8e6cdb80-21e2-4fdf-a64b-86328e236349\") " pod="openshift-marketplace/certified-operators-42rg6" Mar 18 21:34:49 crc kubenswrapper[4950]: I0318 21:34:49.131063 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e6cdb80-21e2-4fdf-a64b-86328e236349-catalog-content\") pod \"certified-operators-42rg6\" (UID: \"8e6cdb80-21e2-4fdf-a64b-86328e236349\") " pod="openshift-marketplace/certified-operators-42rg6" Mar 18 21:34:49 crc kubenswrapper[4950]: I0318 21:34:49.131322 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e6cdb80-21e2-4fdf-a64b-86328e236349-utilities\") pod \"certified-operators-42rg6\" (UID: \"8e6cdb80-21e2-4fdf-a64b-86328e236349\") " pod="openshift-marketplace/certified-operators-42rg6" Mar 18 21:34:49 crc kubenswrapper[4950]: I0318 21:34:49.149477 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlfqs\" (UniqueName: \"kubernetes.io/projected/8e6cdb80-21e2-4fdf-a64b-86328e236349-kube-api-access-tlfqs\") pod \"certified-operators-42rg6\" (UID: \"8e6cdb80-21e2-4fdf-a64b-86328e236349\") " pod="openshift-marketplace/certified-operators-42rg6" Mar 18 21:34:49 crc kubenswrapper[4950]: I0318 21:34:49.258373 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-42rg6" Mar 18 21:34:49 crc kubenswrapper[4950]: I0318 21:34:49.688994 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-42rg6"] Mar 18 21:34:50 crc kubenswrapper[4950]: I0318 21:34:50.141504 4950 generic.go:334] "Generic (PLEG): container finished" podID="8e6cdb80-21e2-4fdf-a64b-86328e236349" containerID="31b721ac21bc6c3724d727af394a1b43e3fb1c641c89c389c58035725f4b1443" exitCode=0 Mar 18 21:34:50 crc kubenswrapper[4950]: I0318 21:34:50.141564 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42rg6" event={"ID":"8e6cdb80-21e2-4fdf-a64b-86328e236349","Type":"ContainerDied","Data":"31b721ac21bc6c3724d727af394a1b43e3fb1c641c89c389c58035725f4b1443"} Mar 18 21:34:50 crc kubenswrapper[4950]: I0318 21:34:50.141619 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42rg6" event={"ID":"8e6cdb80-21e2-4fdf-a64b-86328e236349","Type":"ContainerStarted","Data":"cfa854a733f035ddc2853337a962b80a1fa8dcc5a688da63b13d92f5d3aec65a"} Mar 18 21:34:51 crc kubenswrapper[4950]: I0318 21:34:51.153117 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42rg6" event={"ID":"8e6cdb80-21e2-4fdf-a64b-86328e236349","Type":"ContainerStarted","Data":"493b5f61b72d68e015d3efd25662409aefef9e9666da0eb17beec8f193e946f2"} Mar 18 21:34:53 crc kubenswrapper[4950]: I0318 21:34:53.176965 4950 generic.go:334] "Generic (PLEG): container finished" podID="8e6cdb80-21e2-4fdf-a64b-86328e236349" containerID="493b5f61b72d68e015d3efd25662409aefef9e9666da0eb17beec8f193e946f2" exitCode=0 Mar 18 21:34:53 crc kubenswrapper[4950]: I0318 21:34:53.177381 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42rg6" event={"ID":"8e6cdb80-21e2-4fdf-a64b-86328e236349","Type":"ContainerDied","Data":"493b5f61b72d68e015d3efd25662409aefef9e9666da0eb17beec8f193e946f2"} Mar 18 21:34:54 crc kubenswrapper[4950]: I0318 21:34:54.186729 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42rg6" event={"ID":"8e6cdb80-21e2-4fdf-a64b-86328e236349","Type":"ContainerStarted","Data":"34496330b7741881f5e20551c57404be79d0c5d6717f80c0f5f407c195b32fa4"} Mar 18 21:34:54 crc kubenswrapper[4950]: I0318 21:34:54.209523 4950 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-42rg6" podStartSLOduration=2.793271059 podStartE2EDuration="6.209499089s" podCreationTimestamp="2026-03-18 21:34:48 +0000 UTC" firstStartedPulling="2026-03-18 21:34:50.143397755 +0000 UTC m=+5303.384239613" lastFinishedPulling="2026-03-18 21:34:53.559625765 +0000 UTC m=+5306.800467643" observedRunningTime="2026-03-18 21:34:54.203973154 +0000 UTC m=+5307.444815032" watchObservedRunningTime="2026-03-18 21:34:54.209499089 +0000 UTC m=+5307.450340957" Mar 18 21:34:59 crc kubenswrapper[4950]: I0318 21:34:59.260368 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-42rg6" Mar 18 21:34:59 crc kubenswrapper[4950]: I0318 21:34:59.261019 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-42rg6" Mar 18 21:34:59 crc kubenswrapper[4950]: I0318 21:34:59.326550 4950 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-42rg6" Mar 18 21:35:00 crc kubenswrapper[4950]: I0318 21:35:00.323573 4950 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-42rg6" Mar 18 21:35:00 crc kubenswrapper[4950]: I0318 21:35:00.407717 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-42rg6"] Mar 18 21:35:02 crc kubenswrapper[4950]: I0318 21:35:02.261969 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-42rg6" podUID="8e6cdb80-21e2-4fdf-a64b-86328e236349" containerName="registry-server" containerID="cri-o://34496330b7741881f5e20551c57404be79d0c5d6717f80c0f5f407c195b32fa4" gracePeriod=2 Mar 18 21:35:02 crc kubenswrapper[4950]: I0318 21:35:02.705504 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-42rg6" Mar 18 21:35:02 crc kubenswrapper[4950]: I0318 21:35:02.833937 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e6cdb80-21e2-4fdf-a64b-86328e236349-utilities\") pod \"8e6cdb80-21e2-4fdf-a64b-86328e236349\" (UID: \"8e6cdb80-21e2-4fdf-a64b-86328e236349\") " Mar 18 21:35:02 crc kubenswrapper[4950]: I0318 21:35:02.834370 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlfqs\" (UniqueName: \"kubernetes.io/projected/8e6cdb80-21e2-4fdf-a64b-86328e236349-kube-api-access-tlfqs\") pod \"8e6cdb80-21e2-4fdf-a64b-86328e236349\" (UID: \"8e6cdb80-21e2-4fdf-a64b-86328e236349\") " Mar 18 21:35:02 crc kubenswrapper[4950]: I0318 21:35:02.834457 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e6cdb80-21e2-4fdf-a64b-86328e236349-catalog-content\") pod \"8e6cdb80-21e2-4fdf-a64b-86328e236349\" (UID: \"8e6cdb80-21e2-4fdf-a64b-86328e236349\") " Mar 18 21:35:02 crc kubenswrapper[4950]: I0318 21:35:02.835038 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e6cdb80-21e2-4fdf-a64b-86328e236349-utilities" (OuterVolumeSpecName: "utilities") pod "8e6cdb80-21e2-4fdf-a64b-86328e236349" (UID: "8e6cdb80-21e2-4fdf-a64b-86328e236349"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:35:02 crc kubenswrapper[4950]: I0318 21:35:02.846616 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e6cdb80-21e2-4fdf-a64b-86328e236349-kube-api-access-tlfqs" (OuterVolumeSpecName: "kube-api-access-tlfqs") pod "8e6cdb80-21e2-4fdf-a64b-86328e236349" (UID: "8e6cdb80-21e2-4fdf-a64b-86328e236349"). InnerVolumeSpecName "kube-api-access-tlfqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:35:02 crc kubenswrapper[4950]: I0318 21:35:02.896724 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e6cdb80-21e2-4fdf-a64b-86328e236349-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e6cdb80-21e2-4fdf-a64b-86328e236349" (UID: "8e6cdb80-21e2-4fdf-a64b-86328e236349"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 18 21:35:02 crc kubenswrapper[4950]: I0318 21:35:02.916278 4950 scope.go:117] "RemoveContainer" containerID="2272e55f06f7d860bacc8aed3a0983a63354f4c1b71f8ba8c387cfa7397aae53" Mar 18 21:35:02 crc kubenswrapper[4950]: I0318 21:35:02.936799 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlfqs\" (UniqueName: \"kubernetes.io/projected/8e6cdb80-21e2-4fdf-a64b-86328e236349-kube-api-access-tlfqs\") on node \"crc\" DevicePath \"\"" Mar 18 21:35:02 crc kubenswrapper[4950]: I0318 21:35:02.936835 4950 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e6cdb80-21e2-4fdf-a64b-86328e236349-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 18 21:35:02 crc kubenswrapper[4950]: I0318 21:35:02.936845 4950 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e6cdb80-21e2-4fdf-a64b-86328e236349-utilities\") on node \"crc\" DevicePath \"\"" Mar 18 21:35:03 crc kubenswrapper[4950]: I0318 21:35:03.272123 4950 generic.go:334] "Generic (PLEG): container finished" podID="8e6cdb80-21e2-4fdf-a64b-86328e236349" containerID="34496330b7741881f5e20551c57404be79d0c5d6717f80c0f5f407c195b32fa4" exitCode=0 Mar 18 21:35:03 crc kubenswrapper[4950]: I0318 21:35:03.272175 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42rg6" event={"ID":"8e6cdb80-21e2-4fdf-a64b-86328e236349","Type":"ContainerDied","Data":"34496330b7741881f5e20551c57404be79d0c5d6717f80c0f5f407c195b32fa4"} Mar 18 21:35:03 crc kubenswrapper[4950]: I0318 21:35:03.272184 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-42rg6" Mar 18 21:35:03 crc kubenswrapper[4950]: I0318 21:35:03.272213 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42rg6" event={"ID":"8e6cdb80-21e2-4fdf-a64b-86328e236349","Type":"ContainerDied","Data":"cfa854a733f035ddc2853337a962b80a1fa8dcc5a688da63b13d92f5d3aec65a"} Mar 18 21:35:03 crc kubenswrapper[4950]: I0318 21:35:03.272262 4950 scope.go:117] "RemoveContainer" containerID="34496330b7741881f5e20551c57404be79d0c5d6717f80c0f5f407c195b32fa4" Mar 18 21:35:03 crc kubenswrapper[4950]: I0318 21:35:03.303550 4950 scope.go:117] "RemoveContainer" containerID="493b5f61b72d68e015d3efd25662409aefef9e9666da0eb17beec8f193e946f2" Mar 18 21:35:03 crc kubenswrapper[4950]: I0318 21:35:03.309942 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-42rg6"] Mar 18 21:35:03 crc kubenswrapper[4950]: I0318 21:35:03.318863 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-42rg6"] Mar 18 21:35:03 crc kubenswrapper[4950]: I0318 21:35:03.329636 4950 scope.go:117] "RemoveContainer" containerID="31b721ac21bc6c3724d727af394a1b43e3fb1c641c89c389c58035725f4b1443" Mar 18 21:35:03 crc kubenswrapper[4950]: I0318 21:35:03.346207 4950 scope.go:117] "RemoveContainer" containerID="34496330b7741881f5e20551c57404be79d0c5d6717f80c0f5f407c195b32fa4" Mar 18 21:35:03 crc kubenswrapper[4950]: E0318 21:35:03.346798 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34496330b7741881f5e20551c57404be79d0c5d6717f80c0f5f407c195b32fa4\": container with ID starting with 34496330b7741881f5e20551c57404be79d0c5d6717f80c0f5f407c195b32fa4 not found: ID does not exist" containerID="34496330b7741881f5e20551c57404be79d0c5d6717f80c0f5f407c195b32fa4" Mar 18 21:35:03 crc kubenswrapper[4950]: I0318 21:35:03.346830 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34496330b7741881f5e20551c57404be79d0c5d6717f80c0f5f407c195b32fa4"} err="failed to get container status \"34496330b7741881f5e20551c57404be79d0c5d6717f80c0f5f407c195b32fa4\": rpc error: code = NotFound desc = could not find container \"34496330b7741881f5e20551c57404be79d0c5d6717f80c0f5f407c195b32fa4\": container with ID starting with 34496330b7741881f5e20551c57404be79d0c5d6717f80c0f5f407c195b32fa4 not found: ID does not exist" Mar 18 21:35:03 crc kubenswrapper[4950]: I0318 21:35:03.346851 4950 scope.go:117] "RemoveContainer" containerID="493b5f61b72d68e015d3efd25662409aefef9e9666da0eb17beec8f193e946f2" Mar 18 21:35:03 crc kubenswrapper[4950]: E0318 21:35:03.347197 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"493b5f61b72d68e015d3efd25662409aefef9e9666da0eb17beec8f193e946f2\": container with ID starting with 493b5f61b72d68e015d3efd25662409aefef9e9666da0eb17beec8f193e946f2 not found: ID does not exist" containerID="493b5f61b72d68e015d3efd25662409aefef9e9666da0eb17beec8f193e946f2" Mar 18 21:35:03 crc kubenswrapper[4950]: I0318 21:35:03.347259 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"493b5f61b72d68e015d3efd25662409aefef9e9666da0eb17beec8f193e946f2"} err="failed to get container status \"493b5f61b72d68e015d3efd25662409aefef9e9666da0eb17beec8f193e946f2\": rpc error: code = NotFound desc = could not find container \"493b5f61b72d68e015d3efd25662409aefef9e9666da0eb17beec8f193e946f2\": container with ID starting with 493b5f61b72d68e015d3efd25662409aefef9e9666da0eb17beec8f193e946f2 not found: ID does not exist" Mar 18 21:35:03 crc kubenswrapper[4950]: I0318 21:35:03.347283 4950 scope.go:117] "RemoveContainer" containerID="31b721ac21bc6c3724d727af394a1b43e3fb1c641c89c389c58035725f4b1443" Mar 18 21:35:03 crc kubenswrapper[4950]: E0318 21:35:03.347561 4950 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31b721ac21bc6c3724d727af394a1b43e3fb1c641c89c389c58035725f4b1443\": container with ID starting with 31b721ac21bc6c3724d727af394a1b43e3fb1c641c89c389c58035725f4b1443 not found: ID does not exist" containerID="31b721ac21bc6c3724d727af394a1b43e3fb1c641c89c389c58035725f4b1443" Mar 18 21:35:03 crc kubenswrapper[4950]: I0318 21:35:03.347609 4950 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31b721ac21bc6c3724d727af394a1b43e3fb1c641c89c389c58035725f4b1443"} err="failed to get container status \"31b721ac21bc6c3724d727af394a1b43e3fb1c641c89c389c58035725f4b1443\": rpc error: code = NotFound desc = could not find container \"31b721ac21bc6c3724d727af394a1b43e3fb1c641c89c389c58035725f4b1443\": container with ID starting with 31b721ac21bc6c3724d727af394a1b43e3fb1c641c89c389c58035725f4b1443 not found: ID does not exist" Mar 18 21:35:03 crc kubenswrapper[4950]: I0318 21:35:03.490393 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e6cdb80-21e2-4fdf-a64b-86328e236349" path="/var/lib/kubelet/pods/8e6cdb80-21e2-4fdf-a64b-86328e236349/volumes" Mar 18 21:35:03 crc kubenswrapper[4950]: I0318 21:35:03.836572 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 21:35:03 crc kubenswrapper[4950]: I0318 21:35:03.836898 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 21:35:33 crc kubenswrapper[4950]: I0318 21:35:33.836064 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 21:35:33 crc kubenswrapper[4950]: I0318 21:35:33.836823 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 21:36:00 crc kubenswrapper[4950]: I0318 21:36:00.182220 4950 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29564496-d7pqm"] Mar 18 21:36:00 crc kubenswrapper[4950]: E0318 21:36:00.184938 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e6cdb80-21e2-4fdf-a64b-86328e236349" containerName="registry-server" Mar 18 21:36:00 crc kubenswrapper[4950]: I0318 21:36:00.185103 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e6cdb80-21e2-4fdf-a64b-86328e236349" containerName="registry-server" Mar 18 21:36:00 crc kubenswrapper[4950]: E0318 21:36:00.185255 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e6cdb80-21e2-4fdf-a64b-86328e236349" containerName="extract-utilities" Mar 18 21:36:00 crc kubenswrapper[4950]: I0318 21:36:00.185376 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e6cdb80-21e2-4fdf-a64b-86328e236349" containerName="extract-utilities" Mar 18 21:36:00 crc kubenswrapper[4950]: E0318 21:36:00.185565 4950 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e6cdb80-21e2-4fdf-a64b-86328e236349" containerName="extract-content" Mar 18 21:36:00 crc kubenswrapper[4950]: I0318 21:36:00.185732 4950 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e6cdb80-21e2-4fdf-a64b-86328e236349" containerName="extract-content" Mar 18 21:36:00 crc kubenswrapper[4950]: I0318 21:36:00.186261 4950 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e6cdb80-21e2-4fdf-a64b-86328e236349" containerName="registry-server" Mar 18 21:36:00 crc kubenswrapper[4950]: I0318 21:36:00.187565 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564496-d7pqm" Mar 18 21:36:00 crc kubenswrapper[4950]: I0318 21:36:00.190375 4950 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-2lkk5" Mar 18 21:36:00 crc kubenswrapper[4950]: I0318 21:36:00.190866 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 18 21:36:00 crc kubenswrapper[4950]: I0318 21:36:00.193615 4950 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 18 21:36:00 crc kubenswrapper[4950]: I0318 21:36:00.203015 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564496-d7pqm"] Mar 18 21:36:00 crc kubenswrapper[4950]: I0318 21:36:00.275220 4950 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9dq6\" (UniqueName: \"kubernetes.io/projected/a3baef8a-fed5-4c75-8335-2e653b524cff-kube-api-access-w9dq6\") pod \"auto-csr-approver-29564496-d7pqm\" (UID: \"a3baef8a-fed5-4c75-8335-2e653b524cff\") " pod="openshift-infra/auto-csr-approver-29564496-d7pqm" Mar 18 21:36:00 crc kubenswrapper[4950]: I0318 21:36:00.377442 4950 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9dq6\" (UniqueName: \"kubernetes.io/projected/a3baef8a-fed5-4c75-8335-2e653b524cff-kube-api-access-w9dq6\") pod \"auto-csr-approver-29564496-d7pqm\" (UID: \"a3baef8a-fed5-4c75-8335-2e653b524cff\") " pod="openshift-infra/auto-csr-approver-29564496-d7pqm" Mar 18 21:36:00 crc kubenswrapper[4950]: I0318 21:36:00.401796 4950 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9dq6\" (UniqueName: \"kubernetes.io/projected/a3baef8a-fed5-4c75-8335-2e653b524cff-kube-api-access-w9dq6\") pod \"auto-csr-approver-29564496-d7pqm\" (UID: \"a3baef8a-fed5-4c75-8335-2e653b524cff\") " pod="openshift-infra/auto-csr-approver-29564496-d7pqm" Mar 18 21:36:00 crc kubenswrapper[4950]: I0318 21:36:00.522600 4950 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564496-d7pqm" Mar 18 21:36:01 crc kubenswrapper[4950]: I0318 21:36:01.007821 4950 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29564496-d7pqm"] Mar 18 21:36:01 crc kubenswrapper[4950]: W0318 21:36:01.017017 4950 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3baef8a_fed5_4c75_8335_2e653b524cff.slice/crio-e9bd235ef7e4bf07baa668dfe9cb1975ca2a35caff02651ca56a984018f94d34 WatchSource:0}: Error finding container e9bd235ef7e4bf07baa668dfe9cb1975ca2a35caff02651ca56a984018f94d34: Status 404 returned error can't find the container with id e9bd235ef7e4bf07baa668dfe9cb1975ca2a35caff02651ca56a984018f94d34 Mar 18 21:36:01 crc kubenswrapper[4950]: I0318 21:36:01.021665 4950 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 18 21:36:02 crc kubenswrapper[4950]: I0318 21:36:02.034163 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564496-d7pqm" event={"ID":"a3baef8a-fed5-4c75-8335-2e653b524cff","Type":"ContainerStarted","Data":"e9bd235ef7e4bf07baa668dfe9cb1975ca2a35caff02651ca56a984018f94d34"} Mar 18 21:36:03 crc kubenswrapper[4950]: I0318 21:36:03.836388 4950 patch_prober.go:28] interesting pod/machine-config-daemon-k2n6r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 18 21:36:03 crc kubenswrapper[4950]: I0318 21:36:03.836944 4950 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 18 21:36:03 crc kubenswrapper[4950]: I0318 21:36:03.836990 4950 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" Mar 18 21:36:03 crc kubenswrapper[4950]: I0318 21:36:03.837668 4950 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6ea174cb4f877354d40698f8225ba9da25bb0ae0a816d0571cda56c7af7db12c"} pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 18 21:36:03 crc kubenswrapper[4950]: I0318 21:36:03.837721 4950 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" containerName="machine-config-daemon" containerID="cri-o://6ea174cb4f877354d40698f8225ba9da25bb0ae0a816d0571cda56c7af7db12c" gracePeriod=600 Mar 18 21:36:04 crc kubenswrapper[4950]: I0318 21:36:04.053387 4950 generic.go:334] "Generic (PLEG): container finished" podID="a3baef8a-fed5-4c75-8335-2e653b524cff" containerID="52e3045fb1c5d5ad9a3a47b6214b4eaabcc3577f747847d9c1666b7e7c334ae1" exitCode=0 Mar 18 21:36:04 crc kubenswrapper[4950]: I0318 21:36:04.053510 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564496-d7pqm" event={"ID":"a3baef8a-fed5-4c75-8335-2e653b524cff","Type":"ContainerDied","Data":"52e3045fb1c5d5ad9a3a47b6214b4eaabcc3577f747847d9c1666b7e7c334ae1"} Mar 18 21:36:04 crc kubenswrapper[4950]: E0318 21:36:04.179262 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:36:05 crc kubenswrapper[4950]: I0318 21:36:05.069802 4950 generic.go:334] "Generic (PLEG): container finished" podID="4048b439-3266-46e7-9de0-22377efacc46" containerID="6ea174cb4f877354d40698f8225ba9da25bb0ae0a816d0571cda56c7af7db12c" exitCode=0 Mar 18 21:36:05 crc kubenswrapper[4950]: I0318 21:36:05.069884 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" event={"ID":"4048b439-3266-46e7-9de0-22377efacc46","Type":"ContainerDied","Data":"6ea174cb4f877354d40698f8225ba9da25bb0ae0a816d0571cda56c7af7db12c"} Mar 18 21:36:05 crc kubenswrapper[4950]: I0318 21:36:05.069963 4950 scope.go:117] "RemoveContainer" containerID="8a3de554c0ec05e3ab570b15a3ddb9444d28e1d8a8dd09556e19d1262de8dcfc" Mar 18 21:36:05 crc kubenswrapper[4950]: I0318 21:36:05.071737 4950 scope.go:117] "RemoveContainer" containerID="6ea174cb4f877354d40698f8225ba9da25bb0ae0a816d0571cda56c7af7db12c" Mar 18 21:36:05 crc kubenswrapper[4950]: E0318 21:36:05.072634 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:36:05 crc kubenswrapper[4950]: I0318 21:36:05.494544 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564496-d7pqm" Mar 18 21:36:05 crc kubenswrapper[4950]: I0318 21:36:05.582242 4950 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9dq6\" (UniqueName: \"kubernetes.io/projected/a3baef8a-fed5-4c75-8335-2e653b524cff-kube-api-access-w9dq6\") pod \"a3baef8a-fed5-4c75-8335-2e653b524cff\" (UID: \"a3baef8a-fed5-4c75-8335-2e653b524cff\") " Mar 18 21:36:05 crc kubenswrapper[4950]: I0318 21:36:05.590012 4950 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3baef8a-fed5-4c75-8335-2e653b524cff-kube-api-access-w9dq6" (OuterVolumeSpecName: "kube-api-access-w9dq6") pod "a3baef8a-fed5-4c75-8335-2e653b524cff" (UID: "a3baef8a-fed5-4c75-8335-2e653b524cff"). InnerVolumeSpecName "kube-api-access-w9dq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 18 21:36:05 crc kubenswrapper[4950]: I0318 21:36:05.684824 4950 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9dq6\" (UniqueName: \"kubernetes.io/projected/a3baef8a-fed5-4c75-8335-2e653b524cff-kube-api-access-w9dq6\") on node \"crc\" DevicePath \"\"" Mar 18 21:36:06 crc kubenswrapper[4950]: I0318 21:36:06.080084 4950 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29564496-d7pqm" event={"ID":"a3baef8a-fed5-4c75-8335-2e653b524cff","Type":"ContainerDied","Data":"e9bd235ef7e4bf07baa668dfe9cb1975ca2a35caff02651ca56a984018f94d34"} Mar 18 21:36:06 crc kubenswrapper[4950]: I0318 21:36:06.080120 4950 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9bd235ef7e4bf07baa668dfe9cb1975ca2a35caff02651ca56a984018f94d34" Mar 18 21:36:06 crc kubenswrapper[4950]: I0318 21:36:06.080174 4950 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29564496-d7pqm" Mar 18 21:36:06 crc kubenswrapper[4950]: I0318 21:36:06.585682 4950 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29564490-w6t7b"] Mar 18 21:36:06 crc kubenswrapper[4950]: I0318 21:36:06.594563 4950 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29564490-w6t7b"] Mar 18 21:36:07 crc kubenswrapper[4950]: I0318 21:36:07.499493 4950 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f18ea393-c189-46c3-9755-33c306d9dd07" path="/var/lib/kubelet/pods/f18ea393-c189-46c3-9755-33c306d9dd07/volumes" Mar 18 21:36:19 crc kubenswrapper[4950]: I0318 21:36:19.482991 4950 scope.go:117] "RemoveContainer" containerID="6ea174cb4f877354d40698f8225ba9da25bb0ae0a816d0571cda56c7af7db12c" Mar 18 21:36:19 crc kubenswrapper[4950]: E0318 21:36:19.484157 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:36:31 crc kubenswrapper[4950]: I0318 21:36:31.479758 4950 scope.go:117] "RemoveContainer" containerID="6ea174cb4f877354d40698f8225ba9da25bb0ae0a816d0571cda56c7af7db12c" Mar 18 21:36:31 crc kubenswrapper[4950]: E0318 21:36:31.480502 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" Mar 18 21:36:46 crc kubenswrapper[4950]: I0318 21:36:46.479867 4950 scope.go:117] "RemoveContainer" containerID="6ea174cb4f877354d40698f8225ba9da25bb0ae0a816d0571cda56c7af7db12c" Mar 18 21:36:46 crc kubenswrapper[4950]: E0318 21:36:46.480513 4950 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k2n6r_openshift-machine-config-operator(4048b439-3266-46e7-9de0-22377efacc46)\"" pod="openshift-machine-config-operator/machine-config-daemon-k2n6r" podUID="4048b439-3266-46e7-9de0-22377efacc46" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515156615401024451 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015156615402017367 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015156602243016511 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015156602243015461 5ustar corecore